From 8263fd7bdb247f2c3ff21debb50b22bd9b030339 Mon Sep 17 00:00:00 2001 From: askmyteapot <62238146+askmyteapot@users.noreply.github.com> Date: Thu, 24 Aug 2023 00:15:48 +1000 Subject: [PATCH] Update llama_v3.cpp (#393) Fixing C2065 compiler error. Missed '3' on 3 separate identifiers (kB > kB3, MB > MB3) --- otherarch/llama_v3.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/otherarch/llama_v3.cpp b/otherarch/llama_v3.cpp index 4429fe628..47bcc4879 100644 --- a/otherarch/llama_v3.cpp +++ b/otherarch/llama_v3.cpp @@ -1315,8 +1315,8 @@ static void llama_v3_model_load_internal( ggml_cuda_set_scratch_size(vram_scratch); if (n_gpu_layers > 0) { LLAMA_V3_LOG_INFO("%s: allocating batch_size x (%zd kB + n_ctx x %zd B) = %zd MB VRAM for the scratch buffer\n", - __func__, vram_scratch_base / kB, vram_scratch_per_context, - (vram_scratch + MB - 1) / MB); // round up + __func__, vram_scratch_base / kB3, vram_scratch_per_context, + (vram_scratch + MB3 - 1) / MB3); // round up } } #endif // GGML_USE_CUBLAS