From 27bc11da0f6349eff044c5861bc30d23267281ef Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sun, 13 Oct 2024 13:57:19 +0300 Subject: [PATCH] llama.vim : update server command [no ci] --- examples/llama.vim | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/llama.vim b/examples/llama.vim index 2818b754e..130af3a26 100644 --- a/examples/llama.vim +++ b/examples/llama.vim @@ -17,7 +17,7 @@ " " start the llama.cpp server with a FIM-compatible model. for example: " -" $ llama-server -m {model.gguf} --port 8012 -ngl 99 -fa --ubatch-size 1024 --batch-size 2048 +" $ llama-server -m {model.gguf} --port 8012 -ngl 99 -fa --ubatch-size 1024 --batch-size 2048 --cache-reuse 512 " " --batch-size [512, model max context] "