From 3614956bc7024317c22dff06fe2f89cae5f66c05 Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Tue, 18 Apr 2023 21:39:05 +0800 Subject: [PATCH] update readme --- README.md | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/README.md b/README.md index 00aa7b355..100e5670a 100644 --- a/README.md +++ b/README.md @@ -8,6 +8,7 @@ What does it mean? You get llama.cpp with a fancy UI, persistent stories, editin # Highlights - Now has experimental CLBlast support. +- Now supports RWKV models WITHOUT pytorch or tokenizers! Yep, just GGML! ## Usage - [Download the latest release here](https://github.com/LostRuins/koboldcpp/releases/latest) or clone the repo. @@ -50,3 +51,9 @@ What does it mean? You get llama.cpp with a fancy UI, persistent stories, editin ## Notes - Generation delay scales linearly with original prompt length. If OpenBLAS is enabled then prompt ingestion becomes about 2-3x faster. This is automatic on windows, but will require linking on OSX and Linux. - I have heard of someone claiming a false AV positive report. The exe is a simple pyinstaller bundle that includes the necessary python scripts and dlls to run. If this still concerns you, you might wish to rebuild everything from source code using the makefile, and you can rebuild the exe yourself with pyinstaller by using `make_pyinstaller.bat` +- Supported GGML models: + - LLAMA (All versions including ggml, ggmf, ggjt, gpt4all) + - GPT-2 (All versions, including legacy f16, newer format + quanitzed, cerebras) + - GPT-J (All versions including legacy f16, newer format + quantized, pyg.cpp, new pygmalion, janeway etc.), + - RWKV (f16 GGMF format) + - Basically every single current and historical GGML format that has ever existed should be supported, except for bloomz.cpp due to lack of demand. \ No newline at end of file