diff --git a/convert_lora_to_gguf.py b/convert_lora_to_gguf.py index 868ffe184..c6af4bf06 100755 --- a/convert_lora_to_gguf.py +++ b/convert_lora_to_gguf.py @@ -359,7 +359,6 @@ if __name__ == '__main__': def set_gguf_parameters(self): self.gguf_writer.add_float32(gguf.Keys.Adapter.LORA_ALPHA, self.lora_alpha) - super().set_gguf_parameters() def generate_extra_tensors(self) -> Iterable[tuple[str, Tensor]]: # Never add extra tensors (e.g. rope_freqs) for LoRA adapters