llama: fixed n_vocab for no_vocab
models
This commit is contained in:
parent
23e0d70bac
commit
a5e87bf438
1 changed files with 6 additions and 1 deletions
|
@ -6061,8 +6061,13 @@ static void llm_load_vocab(
|
||||||
vocab.special_mask_id = -1;
|
vocab.special_mask_id = -1;
|
||||||
vocab.linefeed_id = -1;
|
vocab.linefeed_id = -1;
|
||||||
|
|
||||||
|
// read vocab size from metadata
|
||||||
|
ml.get_key(LLM_KV_VOCAB_SIZE, vocab.n_vocab);
|
||||||
|
|
||||||
return;
|
return;
|
||||||
} else if (tokenizer_model == "llama") {
|
}
|
||||||
|
|
||||||
|
if (tokenizer_model == "llama") {
|
||||||
vocab.type = LLAMA_VOCAB_TYPE_SPM;
|
vocab.type = LLAMA_VOCAB_TYPE_SPM;
|
||||||
|
|
||||||
// default special tokens
|
// default special tokens
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue