llama : on Metal, by default offload the full model

ggml-ci
This commit is contained in:
Georgi Gerganov 2024-01-10 10:15:36 +02:00
parent 3cb1c1fb4e
commit 07a1b052e5
No known key found for this signature in database
GPG key ID: 449E073F9DC10735

View file

@ -9069,7 +9069,8 @@ struct llama_model_params llama_model_default_params() {
};
#ifdef GGML_USE_METAL
result.n_gpu_layers = 1;
// note: we usually have plenty of VRAM, so by default offload all layers to the GPU
result.n_gpu_layers = 999;
#endif
return result;