falcon disable offload only for clblast
This commit is contained in:
parent
29757de61f
commit
0ee394ae1b
1 changed files with 3 additions and 1 deletions
|
@ -606,11 +606,13 @@ ModelLoadResult gpttype_load_model(const load_model_inputs inputs, FileFormat in
|
||||||
llama_ctx_params.use_mmap = inputs.use_mmap;
|
llama_ctx_params.use_mmap = inputs.use_mmap;
|
||||||
llama_ctx_params.use_mlock = inputs.use_mlock;
|
llama_ctx_params.use_mlock = inputs.use_mlock;
|
||||||
llama_ctx_params.n_gpu_layers = inputs.gpulayers;
|
llama_ctx_params.n_gpu_layers = inputs.gpulayers;
|
||||||
|
#if defined(GGML_USE_CLBLAST)
|
||||||
if(file_format==FileFormat::GGUF_FALCON && llama_ctx_params.n_gpu_layers>0)
|
if(file_format==FileFormat::GGUF_FALCON && llama_ctx_params.n_gpu_layers>0)
|
||||||
{
|
{
|
||||||
printf("\nGPU layer offload for GGUF FALCON is known to have issues, it has been set to 0.\n");
|
printf("\nGPU layer offload for GGUF FALCON on OpenCL is known to have issues, it has been set to 0.\n");
|
||||||
llama_ctx_params.n_gpu_layers = 0;
|
llama_ctx_params.n_gpu_layers = 0;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
llama_ctx_params.main_gpu = cu_parseinfo_maindevice;
|
llama_ctx_params.main_gpu = cu_parseinfo_maindevice;
|
||||||
llama_ctx_params.rope_freq_base = rope_freq_base;
|
llama_ctx_params.rope_freq_base = rope_freq_base;
|
||||||
llama_ctx_params.rope_freq_scale = rope_freq_scale;
|
llama_ctx_params.rope_freq_scale = rope_freq_scale;
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue