From eae597182cb61bbde0b26e7cec5999d28b9327fe Mon Sep 17 00:00:00 2001 From: slaren Date: Sun, 8 Sep 2024 12:41:51 +0200 Subject: [PATCH] llama : sanitize tokens in the upper bound (#9359) --- src/llama.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/llama.cpp b/src/llama.cpp index f590bcd3b..39e20440e 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -16077,7 +16077,7 @@ static int llama_decode_internal( } for (uint32_t i = 0; i < n_tokens_all; ++i) { - if (batch_all.token[i] < 0) { + if (batch_all.token[i] < 0 || (uint32_t)batch_all.token[i] >= lctx.model.vocab.n_vocab) { LLAMA_LOG_ERROR("%s: invalid token[%d] = %d", __func__, i, batch_all.token[i]); return -1; } @@ -16376,7 +16376,7 @@ static int llama_encode_internal( } for (uint32_t i = 0; i < n_tokens; ++i) { - if (batch.token[i] < 0) { + if (batch.token[i] < 0 || (uint32_t)batch.token[i] >= lctx.model.vocab.n_vocab) { LLAMA_LOG_ERROR("%s: invalid token[%d] = %d", __func__, i, batch.token[i]); return -1; }