keep processed from the beginning
this means no limit to the input prompt, it will just get reset again as normal
This commit is contained in:
parent
df0e0d094c
commit
549291fe61
1 changed files with 1 additions and 5 deletions
|
@ -100,7 +100,7 @@ struct llama_server_context
|
||||||
// Reset context
|
// Reset context
|
||||||
const int n_left = n_past - params.n_keep;
|
const int n_left = n_past - params.n_keep;
|
||||||
n_past = std::max(1, params.n_keep);
|
n_past = std::max(1, params.n_keep);
|
||||||
processed_tokens.erase(processed_tokens.begin() + n_past, processed_tokens.end());
|
//processed_tokens.erase(processed_tokens.begin() + n_past, processed_tokens.end());
|
||||||
embd.insert(embd.begin(), last_n_tokens.begin() + params.n_ctx - n_left / 2 - embd.size(), last_n_tokens.end() - embd.size());
|
embd.insert(embd.begin(), last_n_tokens.begin() + params.n_ctx - n_left / 2 - embd.size(), last_n_tokens.end() - embd.size());
|
||||||
}
|
}
|
||||||
for (int i = 0; i < (int)embd.size(); i += params.n_batch)
|
for (int i = 0; i < (int)embd.size(); i += params.n_batch)
|
||||||
|
@ -499,10 +499,6 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
|
||||||
{
|
{
|
||||||
llama.params.n_keep = body["n_keep"].get<int>();
|
llama.params.n_keep = body["n_keep"].get<int>();
|
||||||
}
|
}
|
||||||
if (!body["as_loop"].is_null())
|
|
||||||
{
|
|
||||||
llama.as_loop = body["as_loop"].get<bool>();
|
|
||||||
}
|
|
||||||
if (!body["prompt"].is_null())
|
if (!body["prompt"].is_null())
|
||||||
{
|
{
|
||||||
llama.params.prompt = body["prompt"].get<std::string>();
|
llama.params.prompt = body["prompt"].get<std::string>();
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue