mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-02-07 06:53:33 +00:00
Make llama.com n_predict be -1 by default
This commit is contained in:
parent
ca19ecf49c
commit
bbfe4fbd11
1 changed files with 1 additions and 1 deletions
2
third_party/ggml/common.h
vendored
2
third_party/ggml/common.h
vendored
|
@ -22,7 +22,7 @@ struct gpt_params {
|
|||
int32_t seed = -1; // RNG seed
|
||||
int32_t verbose = 0; // Logging verbosity
|
||||
int32_t n_threads = std::min(1, (int)(_getcpucount() * 0.75));
|
||||
int32_t n_predict = 128; // new tokens to predict
|
||||
int32_t n_predict = -1; // new tokens to predict
|
||||
int32_t n_parts = -1; // amount of model parts (-1 = determine from model dimensions)
|
||||
int32_t n_ctx = 512; // context size
|
||||
int32_t n_batch = 64; // batch size for prompt processing (must be >=32 to use BLAS)
|
||||
|
|
Loading…
Reference in a new issue