From b2662907005628d25dc1d14253daccbb851ba195 Mon Sep 17 00:00:00 2001 From: kalineid Date: Mon, 14 Oct 2024 17:31:48 +0800 Subject: [PATCH] [llama.cpp] update convert_hf_to_gguf.py --- convert_hf_to_gguf.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index 45a3cb707..f44c4ca6d 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -331,10 +331,10 @@ class Model: data_torch = torch.from_numpy(data.reshape(old_shape)) if self.ftype == gguf.LlamaFileType.MOSTLY_F16: data_torch = data_torch.to(torch.float16) - else: - return self.modify_tensors(data_torch, name, bid) - return [(self.map_tensor_name(name), data_torch)] + return [(self.map_tensor_name(name), data_torch)] + + return self.modify_tensors(data_torch, name, bid) def tensor_force_quant(self, name: str, new_name: str, bid: int | None, n_dims: int) -> gguf.GGMLQuantizationType | bool: del name, new_name, bid, n_dims # unused