From 3aabd8a274817492861b5aa01507b21cae8ed2aa Mon Sep 17 00:00:00 2001 From: "Meng, Hengyu" Date: Wed, 24 Jan 2024 12:37:29 +0000 Subject: [PATCH] revert no mmq --- common/common.cpp | 9 +-------- 1 file changed, 1 insertion(+), 8 deletions(-) diff --git a/common/common.cpp b/common/common.cpp index 1bd2bc1ec..26351bdb1 100644 --- a/common/common.cpp +++ b/common/common.cpp @@ -631,7 +631,6 @@ bool gpt_params_parse_ex(int argc, char ** argv, gpt_params & params) { invalid_param = true; break; } -#if defined(GGML_USE_CUBLAS) || defined(GGML_USE_SYCL) std::string arg_next = argv[i]; // split string by , and / @@ -649,15 +648,9 @@ bool gpt_params_parse_ex(int argc, char ** argv, gpt_params & params) { params.tensor_split[i] = 0.0f; } } -#else +#ifndef GGML_USE_CUBLAS_SYCL fprintf(stderr, "warning: llama.cpp was compiled without cuBLAS/SYCL. Setting a tensor split has no effect.\n"); #endif // GGML_USE_CUBLAS_SYCL - } else if (arg == "--no-mul-mat-q" || arg == "-nommq") { -#if defined(GGML_USE_CUBLAS) || defined(GGML_USE_SYCL) - params.mul_mat_q = false; -#else - fprintf(stderr, "warning: llama.cpp was compiled without cuBLAS or SYCL. Disabling mul_mat_q kernels has no effect.\n"); -#endif // GGML_USE_CUBLAS || GGML_USE_SYCL } else if (arg == "--no-mmap") { params.use_mmap = false; } else if (arg == "--numa") {