2023-04-27 21:31:20 +00:00
|
|
|
DESCRIPTION
|
|
|
|
|
|
|
|
ggml is a machine learning library useful for LLM inference on CPUs
|
|
|
|
|
|
|
|
LICENSE
|
|
|
|
|
|
|
|
MIT
|
|
|
|
|
|
|
|
ORIGIN
|
|
|
|
|
|
|
|
https://github.com/ggerganov/llama.cpp
|
2023-06-03 20:48:52 +00:00
|
|
|
d8bd0013e8768aaa3dc9cfc1ff01499419d5348e
|
2023-04-27 21:31:20 +00:00
|
|
|
|
|
|
|
LOCAL CHANGES
|
|
|
|
|
2023-06-03 20:48:52 +00:00
|
|
|
- Maintaining support for deprecated file formats
|
2023-04-28 23:15:26 +00:00
|
|
|
- Make it possible for loaded prompts to be cached to disk
|
2023-04-28 09:54:11 +00:00
|
|
|
- Introduce -v and --verbose flags
|
2023-04-28 23:15:26 +00:00
|
|
|
- Reduce batch size from 512 to 32
|
2023-04-29 07:48:14 +00:00
|
|
|
- Allow --n_keep to specify a substring of prompt
|
2023-04-28 09:54:11 +00:00
|
|
|
- Don't print stats / diagnostics unless -v is passed
|
|
|
|
- Reduce --top_p default from 0.95 to 0.70
|
|
|
|
- Change --reverse-prompt to no longer imply --interactive
|
|
|
|
- Permit --reverse-prompt specifying custom EOS if non-interactive
|
2023-04-27 21:31:20 +00:00
|
|
|
- Refactor headers per cosmo convention
|
|
|
|
- Remove C++ exceptions; use Die() function instead
|
2023-05-16 15:07:23 +00:00
|
|
|
- Removed division from matrix multiplication.
|
|
|
|
- Let quantizer convert between ggmt formats
|