convert : add "tokenizer.ggml.pre" GGUF KV (wip)
This commit is contained in:
parent
e3f6dc7409
commit
9b4d63ae53
6 changed files with 58 additions and 7 deletions
|
@ -99,6 +99,7 @@ def main():
|
|||
|
||||
tokens, scores, toktypes = _get_sentencepiece_tokenizer_info(args.model_dir)
|
||||
gguf_writer.add_tokenizer_model('llama')
|
||||
gguf_writer.add_tokenizer_pre('default')
|
||||
gguf_writer.add_token_list(tokens)
|
||||
gguf_writer.add_token_scores(scores)
|
||||
gguf_writer.add_token_types(toktypes)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue