diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index e15a526fc..b2852e8db 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -13,7 +13,6 @@ from pathlib import Path from typing import TYPE_CHECKING, Any, ContextManager, Iterator, cast, Optional import numpy as np -import pdb import torch if TYPE_CHECKING: diff --git a/llama.cpp b/llama.cpp index 4cf76c159..17c0ef72e 100644 --- a/llama.cpp +++ b/llama.cpp @@ -2452,22 +2452,22 @@ static std::string llama_model_ftype_name(llama_ftype ftype) { static const char * llama_model_type_name(e_model type) { switch (type) { - case MODEL_1B: return "1B"; - case MODEL_3B: return "3B"; - case MODEL_7B: return "7B"; - case MODEL_8B: return "8B"; - case MODEL_13B: return "13B"; - case MODEL_15B: return "15B"; - case MODEL_30B: return "30B"; - case MODEL_34B: return "34B"; - case MODEL_40B: return "40B"; - case MODEL_65B: return "65B"; - case MODEL_70B: return "70B"; - case MODEL_SMALL: return "0.1B"; + case MODEL_1B: return "1B"; + case MODEL_3B: return "3B"; + case MODEL_7B: return "7B"; + case MODEL_8B: return "8B"; + case MODEL_13B: return "13B"; + case MODEL_15B: return "15B"; + case MODEL_30B: return "30B"; + case MODEL_34B: return "34B"; + case MODEL_40B: return "40B"; + case MODEL_65B: return "65B"; + case MODEL_70B: return "70B"; + case MODEL_SMALL: return "0.1B"; case MODEL_MEDIUM: return "0.4B"; - case MODEL_LARGE: return "0.8B"; - case MODEL_XL: return "1.5B"; - default: return "?B"; + case MODEL_LARGE: return "0.8B"; + case MODEL_XL: return "1.5B"; + default: return "?B"; } }