Add chat template for vicuna
This commit is contained in:
parent
0d24c6af89
commit
f6104b9b77
1 changed files with 18 additions and 1 deletions
19
llama.cpp
19
llama.cpp
|
@ -15711,7 +15711,24 @@ static int32_t llama_chat_apply_template_internal(
|
|||
ss << message->content << "<|end_of_turn|>";
|
||||
}
|
||||
if (add_ass) {
|
||||
ss << "GPT4 Correct Assistant: ";
|
||||
ss << "GPT4 Correct Assistant:";
|
||||
}
|
||||
} else if (tmpl == "vicuna" || tmpl.find("USER: ") != std::string::npos) {
|
||||
for (auto message : chat) {
|
||||
std::string role(message->role);
|
||||
if (role == "user") {
|
||||
ss << "USER: ";
|
||||
} else if (role == "assistant") {
|
||||
ss << "ASSISTANT: ";
|
||||
}
|
||||
ss << message->content << "\n";
|
||||
if (role == "system") {
|
||||
// Extra newline after system message
|
||||
ss << "\n";
|
||||
}
|
||||
}
|
||||
if (add_ass) {
|
||||
ss << "ASSISTANT:";
|
||||
}
|
||||
} else {
|
||||
// template not supported
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue