From 6e4080450ddd3c812c5a1e27f3b651e0801912e2 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Johannes=20G=C3=A4=C3=9Fler?= Date: Sat, 24 Aug 2024 19:10:12 +0200 Subject: [PATCH] remove metal check --- src/llama.cpp | 8 -------- 1 file changed, 8 deletions(-) diff --git a/src/llama.cpp b/src/llama.cpp index cb2fd5369..aeea54cff 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -17534,14 +17534,6 @@ struct llama_context * llama_new_context_with_model( params.flash_attn = false; } -#ifdef GGML_USE_METAL - if (params.flash_attn && model->hparams.attn_soft_cap) { - LLAMA_LOG_WARN("%s: flash_attn is not compatible with attn_soft_cap - forcing off\n", __func__); - params.flash_attn = false; - } -#endif // GGML_USE_METAL - - if (params.flash_attn && model->hparams.n_embd_head_k != model->hparams.n_embd_head_v) { LLAMA_LOG_WARN("%s: flash_attn requires n_embd_head_k == n_embd_head_v - forcing off\n", __func__); params.flash_attn = false;