convert : update instructions and warning message

This commit is contained in:
Georgi Gerganov 2024-04-30 10:31:27 +03:00
parent 5a91d63d8b
commit dc94e5c03f
No known key found for this signature in database
GPG key ID: 449E073F9DC10735
2 changed files with 16 additions and 8 deletions

View file

@ -156,15 +156,19 @@ src_func += " print(f\"chkhsh: {chkhsh}\")\n"
src_func += "\n" src_func += "\n"
src_func += " res = None\n" src_func += " res = None\n"
src_func += "\n" src_func += "\n"
src_func += " # NOTE: if you get an error here, you need to add the model to the if-elif chain below\n" src_func += " # NOTE: if you get an error here, you need to update the convert-hf-to-gguf-update.py script\n"
src_func += " # don't do this manually - use the convert-hf-to-gguf-update.py script!\n" src_func += " # or pull the latest version of the model from Huggingface\n"
src_func += " # don't edit the hashes manually!\n"
src_func += f"{src_ifs}\n" src_func += f"{src_ifs}\n"
src_func += " if res is None:\n" src_func += " if res is None:\n"
src_func += " print(\"\\n\")\n" src_func += " print(\"\\n\")\n"
src_func += " print(\"**************************************************************************************\")\n" src_func += " print(\"**************************************************************************************\")\n"
src_func += " print(\"** WARNING: The BPE pre-tokenizer was not recognized!\")\n" src_func += " print(\"** WARNING: The BPE pre-tokenizer was not recognized!\")\n"
src_func += " print(\"** This means that it was not added yet or you are using an older version.\")\n" src_func += " print(\"** There are 2 possible reasons for this:\")\n"
src_func += " print(\"** Check convert-hf-to-gguf-update.py and update it accordingly.\")\n" src_func += " print(\"** - the model has not been added to convert-hf-to-gguf-update.py yet\")\n"
src_func += " print(\"** - the pre-tokenization config has changed upstream\")\n"
src_func += " print(\"** Check your model files and convert-hf-to-gguf-update.py and update them accordingly.\")\n"
src_func += " print(\"** ref: https://github.com/ggerganov/llama.cpp/pull/6920\")\n"
src_func += " print(\"**\")\n" src_func += " print(\"**\")\n"
src_func += " print(f\"** chkhsh: {chkhsh}\")\n" src_func += " print(f\"** chkhsh: {chkhsh}\")\n"
src_func += " print(\"**************************************************************************************\")\n" src_func += " print(\"**************************************************************************************\")\n"

View file

@ -279,8 +279,9 @@ class Model(ABC):
res = None res = None
# NOTE: if you get an error here, you need to add the model to the if-elif chain below # NOTE: if you get an error here, you need to update the convert-hf-to-gguf-update.py script
# don't do this manually - use the convert-hf-to-gguf-update.py script! # or pull the latest version of the model from Huggingface
# don't edit the hashes manually!
if chkhsh == "0ef9807a4087ebef797fc749390439009c3b9eda9ad1a097abbe738f486c01e5": if chkhsh == "0ef9807a4087ebef797fc749390439009c3b9eda9ad1a097abbe738f486c01e5":
# ref: https://huggingface.co/meta-llama/Meta-Llama-3-8B # ref: https://huggingface.co/meta-llama/Meta-Llama-3-8B
res = "llama-bpe" res = "llama-bpe"
@ -310,8 +311,11 @@ class Model(ABC):
print("\n") print("\n")
print("**************************************************************************************") print("**************************************************************************************")
print("** WARNING: The BPE pre-tokenizer was not recognized!") print("** WARNING: The BPE pre-tokenizer was not recognized!")
print("** This means that it was not added yet or you are using an older version.") print("** There are 2 possible reasons for this:")
print("** Check convert-hf-to-gguf-update.py and update it accordingly.") print("** - the model has not been added to convert-hf-to-gguf-update.py yet")
print("** - the pre-tokenization config has changed upstream")
print("** Check your model files and convert-hf-to-gguf-update.py and update them accordingly.")
print("** ref: https://github.com/ggerganov/llama.cpp/pull/6920")
print("**") print("**")
print(f"** chkhsh: {chkhsh}") print(f"** chkhsh: {chkhsh}")
print("**************************************************************************************") print("**************************************************************************************")