From a48dddab86c4ab6e50e095daf3dd8e50e405573a Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Sat, 6 May 2023 11:48:22 +0800 Subject: [PATCH] slightly bump the RAM up to support chinese alpaca --- llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama.cpp b/llama.cpp index da5db76d1..1c9a12c5b 100644 --- a/llama.cpp +++ b/llama.cpp @@ -87,7 +87,7 @@ static const std::map & MEM_REQ_KV_SELF() static const std::map & MEM_REQ_EVAL() { static std::map _MEM_REQ_EVAL = { - { MODEL_7B, 768ull * MB }, + { MODEL_7B, 800ull * MB }, { MODEL_13B, 1024ull * MB }, { MODEL_30B, 1280ull * MB }, { MODEL_65B, 1536ull * MB },