From c17546fffab812c7dbb1daa420f237d43328394b Mon Sep 17 00:00:00 2001 From: HimariO Date: Sat, 21 Sep 2024 17:01:34 +0800 Subject: [PATCH] Barebone Qwen2VL LLM convertor --- convert_hf_to_gguf.py | 26 ++++++++++++++++++++++++++ 1 file changed, 26 insertions(+) diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index b931049d1..0604ccc60 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -1976,6 +1976,32 @@ class Qwen2Model(Model): self._set_vocab_gpt2() +@Model.register("Qwen2VLForConditionalGeneration") +class Qwen2VLModel(Model): + model_arch = gguf.MODEL_ARCH.QWEN2 + + def set_vocab(self): + try: + self._set_vocab_sentencepiece() + except FileNotFoundError: + self._set_vocab_gpt2() + + # def map_tensor_name(self, name: str, try_suffixes: Sequence[str] = (".weight", ".bias")) -> str: + # new_name = self.tensor_map.get_name(key=name, try_suffixes=try_suffixes) + # if name.startswith("visual."): + # breakpoint() + # return "" + # if new_name is None: + # raise ValueError(f"Can not map tensor {name!r}") + # return new_name + + def get_tensors(self) -> Iterator[tuple[str, Tensor]]: + for name, data in super().get_tensors(): + if name.startswith("visual."): + continue + yield name, data + + @Model.register("Qwen2MoeForCausalLM") class Qwen2MoeModel(Model): model_arch = gguf.MODEL_ARCH.QWEN2MOE