diff --git a/llama.cpp b/llama.cpp index 46671dd2a..48e717d5b 100644 --- a/llama.cpp +++ b/llama.cpp @@ -12841,7 +12841,7 @@ struct llama_context * llama_new_context_with_model( ggml_backend_t backend = ggml_backend_sycl_init(model->main_gpu); if (backend == nullptr) { int main_gpu_id = ggml_backend_sycl_get_device_id(model->main_gpu); - LLAMA_LOG_ERROR("%s: failed to initialize SYCL%d (index %d)backend\n", __func__, main_gpu_id, model->main_gpu); + LLAMA_LOG_ERROR("%s: failed to initialize SYCL%d (index %d) backend\n", __func__, main_gpu_id, model->main_gpu); llama_free(ctx); return nullptr; }