Merge ee1c6a4d89
into 19f65187cb
This commit is contained in:
commit
8da041c7b7
1 changed files with 17 additions and 0 deletions
|
@ -4,6 +4,11 @@ You can also use the [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-
|
||||||
|
|
||||||
Note: It is synced from llama.cpp `main` every 6 hours.
|
Note: It is synced from llama.cpp `main` every 6 hours.
|
||||||
|
|
||||||
|
Using llama-quantize, needs cmake to create excutables. Install cmake as per your operating system.
|
||||||
|
https://cmake.org/download/
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
Example usage:
|
Example usage:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
|
@ -17,6 +22,18 @@ ls ./models
|
||||||
ls ./models
|
ls ./models
|
||||||
<folder containing weights and tokenizer json>
|
<folder containing weights and tokenizer json>
|
||||||
|
|
||||||
|
#clone git repository llama.cpp
|
||||||
|
git clone https://github.com/ggerganov/llama.cpp
|
||||||
|
cd llama.cpp
|
||||||
|
|
||||||
|
# create build directory and run cmake
|
||||||
|
mkdir build && cd build && cmake ../examples/quantize
|
||||||
|
|
||||||
|
# build files are created in root directory(llama.cpp)
|
||||||
|
# run make in root directory to create executables
|
||||||
|
cd .. && make
|
||||||
|
|
||||||
|
|
||||||
# install Python dependencies
|
# install Python dependencies
|
||||||
python3 -m pip install -r requirements.txt
|
python3 -m pip install -r requirements.txt
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue