From 0ee394ae1b9cde04955bcf2c419637b90af2c23a Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Wed, 30 Aug 2023 22:35:24 +0800 Subject: [PATCH] falcon disable offload only for clblast --- gpttype_adapter.cpp | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/gpttype_adapter.cpp b/gpttype_adapter.cpp index 83043a244..7d78d6910 100644 --- a/gpttype_adapter.cpp +++ b/gpttype_adapter.cpp @@ -606,11 +606,13 @@ ModelLoadResult gpttype_load_model(const load_model_inputs inputs, FileFormat in llama_ctx_params.use_mmap = inputs.use_mmap; llama_ctx_params.use_mlock = inputs.use_mlock; llama_ctx_params.n_gpu_layers = inputs.gpulayers; + #if defined(GGML_USE_CLBLAST) if(file_format==FileFormat::GGUF_FALCON && llama_ctx_params.n_gpu_layers>0) { - printf("\nGPU layer offload for GGUF FALCON is known to have issues, it has been set to 0.\n"); + printf("\nGPU layer offload for GGUF FALCON on OpenCL is known to have issues, it has been set to 0.\n"); llama_ctx_params.n_gpu_layers = 0; } + #endif llama_ctx_params.main_gpu = cu_parseinfo_maindevice; llama_ctx_params.rope_freq_base = rope_freq_base; llama_ctx_params.rope_freq_scale = rope_freq_scale;