From d4a952e099cbea482c6d71ff40bb09b0ea3be820 Mon Sep 17 00:00:00 2001 From: Anas Ahouzi <112881240+aahouzi@users.noreply.github.com> Date: Sat, 24 Feb 2024 12:46:35 -0800 Subject: [PATCH] Fix hard coded layer_norm_eps --- convert-hf-to-gguf.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index 58e78d588..712e55f74 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -1077,7 +1077,7 @@ class StableLMModel(Model): self.gguf_writer.add_rope_dimension_count(int(hparams["partial_rotary_factor"] * (hparams["hidden_size"] // hparams["num_attention_heads"]))) self.gguf_writer.add_head_count(hparams["num_attention_heads"]) self.gguf_writer.add_parallel_residual(hparams["use_parallel_residual"] if "use_parallel_residual" in hparams else True) - self.gguf_writer.add_layer_norm_eps(1e-5) + self.gguf_writer.add_layer_norm_eps(hparams["layer_norm_eps"]) class MixtralModel(Model):