From ee1c6a4d89229304f78d0272df25911c8cc4e008 Mon Sep 17 00:00:00 2001 From: Amit Kumar Jha Date: Wed, 17 Jul 2024 17:20:07 +0400 Subject: [PATCH] Update README.md to include steps to run cmake --- examples/quantize/README.md | 17 +++++++++++++++++ 1 file changed, 17 insertions(+) diff --git a/examples/quantize/README.md b/examples/quantize/README.md index 553c2701b..f55186169 100644 --- a/examples/quantize/README.md +++ b/examples/quantize/README.md @@ -4,6 +4,11 @@ You can also use the [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf- Note: It is synced from llama.cpp `main` every 6 hours. +Using llama-quantize, needs cmake to create excutables. Install cmake as per your operating system. +https://cmake.org/download/ + + + Example usage: ```bash @@ -17,6 +22,18 @@ ls ./models ls ./models +#clone git repository llama.cpp +git clone https://github.com/ggerganov/llama.cpp +cd llama.cpp + +# create build directory and run cmake +mkdir build && cd build && cmake ../examples/quantize + +# build files are created in root directory(llama.cpp) +# run make in root directory to create executables +cd .. && make + + # install Python dependencies python3 -m pip install -r requirements.txt