Update llama.cpp
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
parent
2f8e6078b0
commit
7b996502e7
1 changed files with 2 additions and 2 deletions
|
@ -8371,12 +8371,12 @@ void llama_sample_top_k(struct llama_context * ctx, llama_token_data_array * can
|
|||
// return;
|
||||
// }
|
||||
|
||||
const int64_t t_start_sample_us = ggml_time_us();
|
||||
|
||||
if (k <= 0) {
|
||||
k = candidates->size;
|
||||
}
|
||||
|
||||
const int64_t t_start_sample_us = ggml_time_us();
|
||||
|
||||
k = std::max(k, (int) min_keep);
|
||||
k = std::min(k, (int) candidates->size);
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue