use mmq
This commit is contained in:
parent
3ef3cf8c3d
commit
c436f18cc4
1 changed files with 1 additions and 0 deletions
|
@ -25,6 +25,7 @@ struct chat {
|
||||||
lparams = llama_context_default_params();
|
lparams = llama_context_default_params();
|
||||||
lparams.n_ctx = 4096;
|
lparams.n_ctx = 4096;
|
||||||
lparams.n_gpu_layers = 99;
|
lparams.n_gpu_layers = 99;
|
||||||
|
lparams.mul_mat_q = true;
|
||||||
|
|
||||||
model = llama_load_model_from_file(model_file.c_str(), lparams);
|
model = llama_load_model_from_file(model_file.c_str(), lparams);
|
||||||
if (model == NULL) {
|
if (model == NULL) {
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue