From cfdc3494e3c22f8a8e34cfc88d37890d1c3e0ab7 Mon Sep 17 00:00:00 2001 From: crasm Date: Mon, 7 Aug 2023 19:03:28 -0400 Subject: [PATCH] Always print num tokens Also, add better context to --verbose-prompt output. --- examples/main/main.cpp | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/examples/main/main.cpp b/examples/main/main.cpp index 56ada7e69..43f6b3055 100644 --- a/examples/main/main.cpp +++ b/examples/main/main.cpp @@ -273,10 +273,11 @@ int main(int argc, char ** argv) { // determine newline token auto llama_token_newline = ::llama_tokenize(ctx, "\n", false); + fprintf(stderr, "\n%s: number of tokens in prompt = %zu\n", __func__, embd_inp.size()); + if (params.verbose_prompt) { - fprintf(stderr, "\n"); - fprintf(stderr, "%s: prompt: '%s'\n", __func__, params.prompt.c_str()); - fprintf(stderr, "%s: number of tokens in prompt = %zu\n", __func__, embd_inp.size()); + fprintf(stderr, "%s: full prompt: '%s'\n", __func__, params.prompt.c_str()); + fprintf(stderr, "%s: prompt tokens:\n", __func__); for (int i = 0; i < (int) embd_inp.size(); i++) { fprintf(stderr, "%6d -> '%s'\n", embd_inp[i], llama_token_to_str(ctx, embd_inp[i])); }