Workaround for #3454
This commit is contained in:
parent
ff5a3f0c09
commit
ee8e2b2604
1 changed files with 6 additions and 2 deletions
|
@ -7713,7 +7713,9 @@ int llama_token_to_piece(const struct llama_model * model, llama_token token, ch
|
|||
buf[0] = llama_token_to_byte(model->vocab, token);
|
||||
return 1;
|
||||
} else {
|
||||
GGML_ASSERT(false);
|
||||
// TODO: for now we accept all unsupported token types,
|
||||
// suppressing them like CONTROL tokens.
|
||||
// GGML_ASSERT(false);
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
@ -7729,7 +7731,9 @@ int llama_token_to_piece(const struct llama_model * model, llama_token token, ch
|
|||
} else if (llama_is_control_token(model->vocab, token)) {
|
||||
;
|
||||
} else {
|
||||
GGML_ASSERT(false);
|
||||
// TODO: for now we accept all unsupported token types,
|
||||
// suppressing them like CONTROL tokens.
|
||||
// GGML_ASSERT(false);
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue