From 3a8e8b7a0f74deaadfe2af677b8ce107d950451b Mon Sep 17 00:00:00 2001 From: Jed Fox Date: Fri, 24 Mar 2023 17:28:34 -0400 Subject: [PATCH] Fix typo --- llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama.cpp b/llama.cpp index 07482eaf2..fc114b164 100644 --- a/llama.cpp +++ b/llama.cpp @@ -1638,7 +1638,7 @@ struct llama_context * llama_init_from_file( ggml_type memory_type = params.f16_kv ? GGML_TYPE_F16 : GGML_TYPE_F32; if (!llama_model_load(path_model, *ctx, params.n_ctx, params.n_parts, memory_type, - params.vocab_only, , params.progress_callback, + params.vocab_only, params.progress_callback, params.progress_callback_user_data)) { fprintf(stderr, "%s: failed to load model\n", __func__); llama_free(ctx);