From 47ea33ab59f3fcca34a7009c02052d810fd97ba8 Mon Sep 17 00:00:00 2001 From: LostRuins <39025047+LostRuins@users.noreply.github.com> Date: Thu, 23 Mar 2023 16:02:19 +0800 Subject: [PATCH] Update README.md --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 0514bacae..83efc7eb7 100644 --- a/README.md +++ b/README.md @@ -7,9 +7,10 @@ What does it mean? You get llama.cpp with a fancy UI, persistent stories, editin ![Preview](preview.png) ## Usage +- [Download the latest release here](https://github.com/LostRuins/llamacpp-for-kobold/releases/latest) or clone the repo. - Windows binaries are provided in the form of **llamacpp.dll** but if you feel worried go ahead and rebuild it yourself. - Weights are not included, you can use the llama.cpp quantize.exe to generate them from your official weight files (or download them from...places). -- To run, simply clone the repo and run `llama_for_kobold.py [ggml_quant_model.bin] [port]`, and then connect with Kobold or Kobold Lite. +- To run, execute the script providing the model as a parameter `llama_for_kobold.py [ggml_quant_model.bin] [port]`, and then connect with Kobold or Kobold Lite. - By default, you can connect to http://localhost:5001 (you can also use https://lite.koboldai.net/?local=1&port=5001). ## Considerations