Fix for the regen issue.
This commit is contained in:
parent
42cf4d8433
commit
03ea8f013a
1 changed files with 4 additions and 0 deletions
|
@ -66,6 +66,10 @@ struct llama_server_context
|
||||||
bool loadPrompt() {
|
bool loadPrompt() {
|
||||||
params.prompt.insert(0, 1, ' '); // always add a first space
|
params.prompt.insert(0, 1, ' '); // always add a first space
|
||||||
std::vector<llama_token> prompt_tokens = ::llama_tokenize(ctx, params.prompt, true);
|
std::vector<llama_token> prompt_tokens = ::llama_tokenize(ctx, params.prompt, true);
|
||||||
|
if (prompt_tokens == last_prompt_tokens)
|
||||||
|
{
|
||||||
|
embd.clear();
|
||||||
|
}
|
||||||
// compare the evaluated prompt with the new prompt
|
// compare the evaluated prompt with the new prompt
|
||||||
for (n_past = 0; n_past < prompt_tokens.size() - 1 && n_past < processed_tokens.size(); n_past++) {
|
for (n_past = 0; n_past < prompt_tokens.size() - 1 && n_past < processed_tokens.size(); n_past++) {
|
||||||
if (prompt_tokens[n_past] != processed_tokens[n_past]) {
|
if (prompt_tokens[n_past] != processed_tokens[n_past]) {
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue