diff --git a/examples/perplexity/perplexity.cpp b/examples/perplexity/perplexity.cpp index 7ed661ec5..534b0ae3f 100644 --- a/examples/perplexity/perplexity.cpp +++ b/examples/perplexity/perplexity.cpp @@ -66,7 +66,7 @@ void perplexity(llama_context * ctx, const gpt_params & params) { tokens[batch_start] = llama_token_bos(); } - if (llama_eval(ctx, tokens.data() + batch_start, batch_size, j * n_batch, params.n_threads, params.n_threads)) { + if (llama_eval(ctx, tokens.data() + batch_start, batch_size, j * n_batch, params.n_threads, params.pp_threads)) { fprintf(stderr, "%s : failed to eval\n", __func__); return; } @@ -337,8 +337,8 @@ int main(int argc, char ** argv) { // print system information { fprintf(stderr, "\n"); - fprintf(stderr, "system_info: n_threads = %d / %d | %s\n", - params.n_threads, std::thread::hardware_concurrency(), llama_print_system_info()); + fprintf(stderr, "system_info: pp_threads = %d / %d | %s\n", + params.pp_threads, std::thread::hardware_concurrency(), llama_print_system_info()); } if (params.hellaswag) {