llama : add n_expert and n_expert_used to hparams + change quants

This commit is contained in:
Georgi Gerganov 2023-12-10 13:57:54 +02:00
parent d1259b7b35
commit e640cbe055
No known key found for this signature in database
GPG key ID: 449E073F9DC10735
6 changed files with 111 additions and 54 deletions

View file

@ -339,6 +339,12 @@ class GGUFWriter:
def add_clamp_kqv(self, value: float) -> None:
self.add_float32(Keys.Attention.CLAMP_KQV.format(arch=self.arch), value)
def add_expert_count(self, count: int) -> None:
self.add_uint32(Keys.LLM.EXPERT_COUNT.format(arch=self.arch), count)
def add_expert_used_count(self, count: int) -> None:
self.add_uint32(Keys.LLM.EXPERT_USED_COUNT.format(arch=self.arch), count)
def add_layer_norm_eps(self, value: float) -> None:
self.add_float32(Keys.Attention.LAYERNORM_EPS.format(arch=self.arch), value)