From 9f4505a0c6e05f4d1f67c321506e34d703ee0122 Mon Sep 17 00:00:00 2001 From: FSSRepo Date: Wed, 3 May 2023 14:25:14 -0600 Subject: [PATCH] fixed some bugs --- examples/server/server.cpp | 2 ++ 1 file changed, 2 insertions(+) diff --git a/examples/server/server.cpp b/examples/server/server.cpp index f78cb6b35..28747159c 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -682,6 +682,7 @@ int main(int argc, char ** argv) { [&llama](size_t offset, DataSink &sink) { int ignore = 0; + llama->tokens_completion = 0; while(!llama->is_antiprompt) { std::string result = llama->inference(); // ignore ### Human: and ### Assistant: @@ -699,6 +700,7 @@ int main(int argc, char ** argv) { } printf("\rProcessing: %i tokens processed.", llama->tokens_completion); } + sink.write("[DONE]", 6); sink.done(); // No more data printf("\rCompletion finished: %i tokens predicted.\n", llama->tokens_completion); return true; // return 'false' if you want to cancel the process.