llama : add gemma
model (#5631)
There are couple things in this architecture: 1. Shared input and output embedding parameters. 2. Key length and value length are not derived from `n_embd`. More information about the models can be found at https://ai.google.dev/gemma. GGUFs can be downloaded from https://huggingface.co/google.
This commit is contained in:
parent
88c46cbdac
commit
580111d42b
3 changed files with 186 additions and 0 deletions
|
@ -107,6 +107,7 @@ Typically finetunes of the base models below are supported as well.
|
|||
- [x] [Orion 14B](https://github.com/ggerganov/llama.cpp/pull/5118)
|
||||
- [x] [InternLM2](https://huggingface.co/models?search=internlm2)
|
||||
- [x] [CodeShell](https://github.com/WisdomShell/codeshell)
|
||||
- [x] [Gemma](https://ai.google.dev/gemma)
|
||||
|
||||
**Multimodal models:**
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue