diff --git a/README.md b/README.md index d1d24a507..851bc6887 100644 --- a/README.md +++ b/README.md @@ -1,10 +1,8 @@ # llama-for-kobold -A hacky little script from Concedo that exposes llama.cpp function bindings, allowing it to be used via a simulated Kobold API endpoint. +A self contained distributable from Concedo that exposes llama.cpp function bindings, allowing it to be used via a simulated Kobold API endpoint. -It's not very usable as there is a fundamental flaw with llama.cpp, which causes generation delay to scale linearly with original prompt length. Nobody knows why or really cares much, so I'm just going to publish whatever I have at this point. - -If you care, **please contribute to [this discussion](https://github.com/ggerganov/llama.cpp/discussions/229)** which, if resolved, will actually make this viable. +![Preview](preview.png) ## Considerations - Don't want to use pybind11 due to dependencies on MSVCC @@ -22,4 +20,7 @@ If you care, **please contribute to [this discussion](https://github.com/ggergan ## License - The original GGML library and llama.cpp by ggerganov are licensed under the MIT License - However, Kobold Lite is licensed under the AGPL v3.0 License -- The provided python ctypes bindings in llamacpp.dll are also under the AGPL v3.0 License \ No newline at end of file +- The provided python ctypes bindings in llamacpp.dll are also under the AGPL v3.0 License + +## Notes +- There is a fundamental flaw with llama.cpp, which causes generation delay to scale linearly with original prompt length. If you care, **please contribute to [this discussion](https://github.com/ggerganov/llama.cpp/discussions/229)** which, if resolved, will actually make this viable. \ No newline at end of file diff --git a/preview.png b/preview.png new file mode 100644 index 000000000..808970715 Binary files /dev/null and b/preview.png differ