diff --git a/llama.cpp b/llama.cpp index ceb70025d..b2e2ca298 100644 --- a/llama.cpp +++ b/llama.cpp @@ -9069,7 +9069,8 @@ struct llama_model_params llama_model_default_params() { }; #ifdef GGML_USE_METAL - result.n_gpu_layers = 1; + // note: we usually have plenty of VRAM, so by default offload all layers to the GPU + result.n_gpu_layers = 999; #endif return result;