Fix inference example lacks required parameters

Signed-off-by: Aisuko <urakiny@gmail.com>
This commit is contained in:
Aisuko 2024-08-16 10:43:20 +10:00
parent 5fd89a70ea
commit b8c85df705

View file

@ -34,7 +34,7 @@ Run the quantized model:
```bash ```bash
# start inference on a gguf model # start inference on a gguf model
./llama-cli -m ./models/mymodel/ggml-model-Q4_K_M.gguf -n 128 ./llama-cli -m ./models/mymodel/ggml-model-Q4_K_M.gguf -cnv -p "You are a helpful assistant"
``` ```
When running the larger models, make sure you have enough disk space to store all the intermediate files. When running the larger models, make sure you have enough disk space to store all the intermediate files.