diff --git a/llama.cpp b/llama.cpp index ca5a939dd..f6912c310 100644 --- a/llama.cpp +++ b/llama.cpp @@ -17628,7 +17628,7 @@ static int32_t llama_chat_apply_template_internal( } } // llama2 templates seem to not care about "add_generation_prompt" - } else if (tmpl == "phi3" || (tmpl.find("<|assistant|>") != std::string::npos && tmpl.find("<|end|>") != std::string::npos )) { + } else if (tmpl == "phi3" || (tmpl.find("<|assistant|>") != std::string::npos && tmpl.find("<|end|>") != std::string::npos)) { // Phi 3 for (auto message : chat) { std::string role(message->role);