llama : add gemma model (#5631)

There are couple things in this architecture:

1. Shared input and output embedding parameters.
2. Key length and value length are not derived from `n_embd`.

More information about the models can be found at
https://ai.google.dev/gemma. GGUFs can be downloaded from
https://huggingface.co/google.
This commit is contained in:
postmasters 2024-02-21 05:08:22 -08:00 committed by GitHub
parent 88c46cbdac
commit 580111d42b
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
3 changed files with 186 additions and 0 deletions

View file

@ -107,6 +107,7 @@ Typically finetunes of the base models below are supported as well.
- [x] [Orion 14B](https://github.com/ggerganov/llama.cpp/pull/5118)
- [x] [InternLM2](https://huggingface.co/models?search=internlm2)
- [x] [CodeShell](https://github.com/WisdomShell/codeshell)
- [x] [Gemma](https://ai.google.dev/gemma)
**Multimodal models:**