fix: readme
This commit is contained in:
parent
48cd819e64
commit
6fcdb07773
1 changed files with 1 additions and 1 deletions
|
@ -48,7 +48,7 @@ python convert-hf-to-gguf.py models/llama-7b/ --awq-path awq_cache/llama-7b-w4-g
|
||||||
|
|
||||||
## Quantize
|
## Quantize
|
||||||
```bash
|
```bash
|
||||||
# For now I only benchmark and confirm the results on q4_0, q4_1 and q2_k types.
|
# We only benchmark and confirm the results on q4_0, q4_1, and q2_k types.
|
||||||
./quantize models/llama_7b_fp16.gguf models/llama_7b_q4_0.gguf q4_0
|
./quantize models/llama_7b_fp16.gguf models/llama_7b_q4_0.gguf q4_0
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue