llama : on Metal, by default offload the full model
ggml-ci
This commit is contained in:
parent
3cb1c1fb4e
commit
07a1b052e5
1 changed files with 2 additions and 1 deletions
|
@ -9069,7 +9069,8 @@ struct llama_model_params llama_model_default_params() {
|
||||||
};
|
};
|
||||||
|
|
||||||
#ifdef GGML_USE_METAL
|
#ifdef GGML_USE_METAL
|
||||||
result.n_gpu_layers = 1;
|
// note: we usually have plenty of VRAM, so by default offload all layers to the GPU
|
||||||
|
result.n_gpu_layers = 999;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
return result;
|
return result;
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue