diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index b51d68307..2e518e71d 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -2262,6 +2262,11 @@ class GemmaModel(Model): tensor_map = gguf.get_tensor_name_map(self.model_arch, block_count) for name, data_torch in self.get_tensors(): + # lm_head is not used in llama.cpp, while autoawq will include this tensor in model + # To prevent errors, skip loading lm_head.weight. + if "lm_head.weight" in name: + continue + old_dtype = data_torch.dtype # convert any unsupported data types to float32