Update convert-llama-h5-to-gguf.py
This commit is contained in:
parent
f3de876a12
commit
da4900e835
1 changed files with 0 additions and 5 deletions
|
@ -104,11 +104,6 @@ if Path(dir_model + "/tokenizer.model").is_file():
|
||||||
print("Adding sentencepiece tokenizer vocab.")
|
print("Adding sentencepiece tokenizer vocab.")
|
||||||
tokenizer = SentencePieceProcessor(dir_model + "/tokenizer.model")
|
tokenizer = SentencePieceProcessor(dir_model + "/tokenizer.model")
|
||||||
|
|
||||||
# output vocab_size followed by all piece/score pairs
|
|
||||||
outbytes: bytes
|
|
||||||
outbytes = b""
|
|
||||||
outbytes += struct.pack("I", tokenizer.vocab_size())
|
|
||||||
|
|
||||||
for i in range(tokenizer.vocab_size()):
|
for i in range(tokenizer.vocab_size()):
|
||||||
text: bytes
|
text: bytes
|
||||||
if tokenizer.is_unknown(i):
|
if tokenizer.is_unknown(i):
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue