Update src/llama.cpp
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com>
This commit is contained in:
parent
facec466e8
commit
1135b0f816
1 changed files with 1 additions and 0 deletions
|
@ -21698,6 +21698,7 @@ static int32_t llama_chat_apply_template_internal(
|
|||
ss << "[|assistant|]";
|
||||
}
|
||||
} else if (tmpl == "rwkv-world" || tmpl_contains("rwkv-world")) {
|
||||
// this template requires the model to have "\n\n" as EOT token
|
||||
for (auto message : chat) {
|
||||
std::string role(message->role);
|
||||
if (role == "user") {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue