From 7744aa6a9cbcdccb059fc29b8dd3c83b64d3d303 Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Tue, 24 Oct 2023 15:37:47 +0800 Subject: [PATCH] updated colab --- colab.ipynb | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/colab.ipynb b/colab.ipynb index 760cf41f3..f91d6a2cd 100644 --- a/colab.ipynb +++ b/colab.ipynb @@ -42,7 +42,7 @@ "source": [ "#@title v-- Enter your model below and then click this to start Koboldcpp\r\n", "\r\n", - "Model = \"https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter-GGUF/resolve/main/LLaMA2-13B-Tiefighter.Q4_K_S.gguf\" #@param [\"\"]{allow-input: true}\r\n", + "Model = \"https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter-GGUF/resolve/main/LLaMA2-13B-Tiefighter.Q4_K_M.gguf\" #@param [\"\"]{allow-input: true}\r\n", "Layers = 43 #@param [43]{allow-input: true}\r\n", "\r\n", "%cd /content\r\n", @@ -58,10 +58,11 @@ "!wget $Model -O model.ggml\r\n", "!wget -c https://github.com/cloudflare/cloudflared/releases/latest/download/cloudflared-linux-amd64\r\n", "!chmod +x cloudflared-linux-amd64\r\n", + "!rm nohup.out\r\n", "!nohup ./cloudflared-linux-amd64 tunnel --url http://localhost:5001 &\r\n", "!sleep 8\r\n", "!cat nohup.out\r\n", - "!python koboldcpp.py model.ggml --usecublas 0 mmq --multiuser --gpulayers $Layers --hordeconfig concedo 1 1 --onready \"echo Connect to the link below && cat nohup.out | grep trycloudflare.com\"\r\n" + "!python koboldcpp.py model.ggml --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize 4096 --hordeconfig concedo 1 1 --onready \"echo Connect to the link below && cat nohup.out | grep trycloudflare.com\"\r\n" ] } ],