From 0f8418079ccc0ed20257ebb2641c4af20c6606c1 Mon Sep 17 00:00:00 2001 From: marcus Date: Thu, 21 Dec 2023 13:51:19 -0800 Subject: [PATCH] formatting and missed `int` in `llama_token_to_piece` --- llama.cpp | 2 +- llama.h | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/llama.cpp b/llama.cpp index 3a186565b..4d5c46ee8 100644 --- a/llama.cpp +++ b/llama.cpp @@ -10266,7 +10266,7 @@ static std::string llama_decode_text(const std::string & text) { } // does not write null-terminator to buf -int32_t llama_token_to_piece(const struct llama_model * model, llama_token token, char * buf, int length) { +int32_t llama_token_to_piece(const struct llama_model * model, llama_token token, char * buf, int32_t length) { if (0 <= token && token < llama_n_vocab(model)) { switch (llama_vocab_get_type(model->vocab)) { case LLAMA_VOCAB_TYPE_SPM: { diff --git a/llama.h b/llama.h index fcd2e7e0a..979ec3948 100644 --- a/llama.h +++ b/llama.h @@ -649,7 +649,7 @@ extern "C" { const struct llama_model * model, llama_token token, char * buf, - int32_t length); + int32_t length); // // Grammar