From 948ce2cf9c5ec93c784367833e4abd97aabf7148 Mon Sep 17 00:00:00 2001 From: Didzis Gosko Date: Mon, 19 Jun 2023 11:48:43 +0300 Subject: [PATCH] Missing model memory release --- llama.cpp | 1 + 1 file changed, 1 insertion(+) diff --git a/llama.cpp b/llama.cpp index de89ec316..96cb59362 100644 --- a/llama.cpp +++ b/llama.cpp @@ -2543,6 +2543,7 @@ struct llama_model * llama_load_model_from_file( if (!llama_model_load(path_model, *model, model->vocab, params.n_ctx, params.n_batch, params.n_gpu_layers, params.main_gpu, params.tensor_split, memory_type, params.use_mmap, params.use_mlock, params.vocab_only, params.progress_callback, params.progress_callback_user_data)) { + delete model; fprintf(stderr, "%s: failed to load model\n", __func__); return nullptr; }