From 05b7f9be6f24104071344f1ff8c73273c84430c9 Mon Sep 17 00:00:00 2001 From: "Meng, Hengyu" Date: Thu, 25 Jan 2024 09:44:19 +0000 Subject: [PATCH] revert setting n_threads in sycl --- llama.cpp | 4 ---- 1 file changed, 4 deletions(-) diff --git a/llama.cpp b/llama.cpp index 5f27088f7..1593fb50d 100644 --- a/llama.cpp +++ b/llama.cpp @@ -6707,10 +6707,6 @@ static int llama_decode_internal( n_threads = 1; } -#ifdef GGML_USE_SYCL - n_threads = 1; -#endif - #ifdef GGML_USE_MPI const int64_t n_layer = hparams.n_layer; ggml_mpi_graph_compute_pre(lctx.ctx_mpi, gf, n_layer);