Update README.md
This commit is contained in:
parent
abc77a7496
commit
6e137abe56
1 changed files with 2 additions and 1 deletions
|
@ -2,7 +2,8 @@ llama.cpp modification to run Falcon (work in progress)
|
||||||
|
|
||||||
Status:
|
Status:
|
||||||
* Quantization works except for Q_K_ types
|
* Quantization works except for Q_K_ types
|
||||||
* CUDA not yet functional
|
* CUDA not yet functional
|
||||||
|
* context size calculation not proper (cuda as well as cpu)
|
||||||
|
|
||||||
|
|
||||||
It appears the Q5 Falcon 40B inference time on CPU is as fast as the A100 fp16 inference time at 2 tk/second
|
It appears the Q5 Falcon 40B inference time on CPU is as fast as the A100 fp16 inference time at 2 tk/second
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue