fix: readme
This commit is contained in:
parent
6fcdb07773
commit
b00e2d90f9
1 changed files with 2 additions and 2 deletions
|
@ -43,7 +43,7 @@ Example for llama model
|
||||||
# For llama7b and llama2 models
|
# For llama7b and llama2 models
|
||||||
python convert.py models/llama-7b/ --awq-path awq_cache/llama-7b-w4-g128.pt --outfile models/llama_7b_fp16.gguf
|
python convert.py models/llama-7b/ --awq-path awq_cache/llama-7b-w4-g128.pt --outfile models/llama_7b_fp16.gguf
|
||||||
# For mistral and mpt models
|
# For mistral and mpt models
|
||||||
python convert-hf-to-gguf.py models/llama-7b/ --awq-path awq_cache/llama-7b-w4-g128.pt --outfile models/llama_7b_fp16.gguf
|
python convert-hf-to-gguf.py models/mpt-7b/ --awq-path awq_cache/llama-7b-w4-g128.pt --outfile models/mpt_7b_fp16.gguf
|
||||||
```
|
```
|
||||||
|
|
||||||
## Quantize
|
## Quantize
|
||||||
|
@ -57,7 +57,7 @@ python convert-hf-to-gguf.py models/llama-7b/ --awq-path awq_cache/llama-7b-w4-g
|
||||||
# For llama and llama2, and mistral models.
|
# For llama and llama2, and mistral models.
|
||||||
./build/bin/main -m models/llama_7b_q4_0.gguf -n 128 --prompt "Once upon a time"
|
./build/bin/main -m models/llama_7b_q4_0.gguf -n 128 --prompt "Once upon a time"
|
||||||
# For mpt models.
|
# For mpt models.
|
||||||
./build/bin/main -m models/llama_7b_q4_0.gguf -awq -n 128 --prompt "Once upon a time"
|
./build/bin/main -m models/mpt_7b_q4_0.gguf -awq -n 128 --prompt "Once upon a time"
|
||||||
```
|
```
|
||||||
|
|
||||||
## Benchmark
|
## Benchmark
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue