Update src/llama.cpp
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com>
This commit is contained in:
parent
facec466e8
commit
1135b0f816
1 changed files with 1 additions and 0 deletions
|
@ -21698,6 +21698,7 @@ static int32_t llama_chat_apply_template_internal(
|
||||||
ss << "[|assistant|]";
|
ss << "[|assistant|]";
|
||||||
}
|
}
|
||||||
} else if (tmpl == "rwkv-world" || tmpl_contains("rwkv-world")) {
|
} else if (tmpl == "rwkv-world" || tmpl_contains("rwkv-world")) {
|
||||||
|
// this template requires the model to have "\n\n" as EOT token
|
||||||
for (auto message : chat) {
|
for (auto message : chat) {
|
||||||
std::string role(message->role);
|
std::string role(message->role);
|
||||||
if (role == "user") {
|
if (role == "user") {
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue