Update llama.cpp
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
parent
2f8e6078b0
commit
7b996502e7
1 changed files with 2 additions and 2 deletions
|
@ -8371,12 +8371,12 @@ void llama_sample_top_k(struct llama_context * ctx, llama_token_data_array * can
|
||||||
// return;
|
// return;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
|
const int64_t t_start_sample_us = ggml_time_us();
|
||||||
|
|
||||||
if (k <= 0) {
|
if (k <= 0) {
|
||||||
k = candidates->size;
|
k = candidates->size;
|
||||||
}
|
}
|
||||||
|
|
||||||
const int64_t t_start_sample_us = ggml_time_us();
|
|
||||||
|
|
||||||
k = std::max(k, (int) min_keep);
|
k = std::max(k, (int) min_keep);
|
||||||
k = std::min(k, (int) candidates->size);
|
k = std::min(k, (int) candidates->size);
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue