From bc88fece87460e0121942a314aa4de71b97bad32 Mon Sep 17 00:00:00 2001 From: "Wang Haoran(Robin)" Date: Mon, 26 Jun 2023 18:11:27 -0700 Subject: [PATCH] server: fix llama_sample_top_k order --- examples/server/server.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/server/server.cpp b/examples/server/server.cpp index c0ed4b016..e4ddbe986 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -378,10 +378,10 @@ struct llama_server_context { } else { // Temperature sampling size_t min_keep = std::max(1, n_probs); + llama_sample_top_k(ctx, &candidates_p, top_k, min_keep); llama_sample_tail_free(ctx, &candidates_p, tfs_z, min_keep); llama_sample_typical(ctx, &candidates_p, typical_p, min_keep); llama_sample_top_p(ctx, &candidates_p, top_p, min_keep); - llama_sample_top_k(ctx, &candidates_p, top_k, min_keep); llama_sample_temperature(ctx, &candidates_p, temp); result.tok = llama_sample_token(ctx, &candidates_p); }