From 9092883d58758100716a7d423e712ec769abd862 Mon Sep 17 00:00:00 2001 From: slaren Date: Wed, 13 Mar 2024 18:09:56 +0100 Subject: [PATCH] llama : better n_batch and n_ubatch comment --- llama.h | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llama.h b/llama.h index 5e1d74a09..2d16cc9b9 100644 --- a/llama.h +++ b/llama.h @@ -234,8 +234,8 @@ extern "C" { struct llama_context_params { uint32_t seed; // RNG seed, -1 for random uint32_t n_ctx; // text context, 0 = from model - uint32_t n_batch; // prompt processing maximum batch size (ignored if n_ubatch is not zero) - uint32_t n_ubatch; // prompt processing maximum batch size + uint32_t n_batch; // logical maximum batch size that can be submitted to llama_decode + uint32_t n_ubatch; // physical maximum batch size uint32_t n_seq_max; // max number of sequences (i.e. distinct states for recurrent models) uint32_t n_threads; // number of threads to use for generation uint32_t n_threads_batch; // number of threads to use for batch processing