Update README.md
This commit is contained in:
parent
2797754843
commit
f75125615a
1 changed files with 2 additions and 1 deletions
|
@ -1,9 +1,10 @@
|
||||||
llama.cpp modification to run Falcon (work in progress)
|
llama.cpp modification to run Falcon (work in progress)
|
||||||
|
|
||||||
Status:
|
Status/Bugs:
|
||||||
* Quantization works except for Q_K_ types
|
* Quantization works except for Q_K_ types
|
||||||
* CUDA not yet functional
|
* CUDA not yet functional
|
||||||
* python conversion script is very basic (produces ggml v0)
|
* python conversion script is very basic (produces ggml v0)
|
||||||
|
* On linux Q5_1 7B user reports a batch token ingestion context memory issue, with -b 1 it's gone. Not reproduced on Windows
|
||||||
|
|
||||||
It appears the Q5 Falcon 40B inference time on CPU is as fast as the A100 fp16 inference time at 2 tk/second
|
It appears the Q5 Falcon 40B inference time on CPU is as fast as the A100 fp16 inference time at 2 tk/second
|
||||||
CPU inference examples:
|
CPU inference examples:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue