rebase
This commit is contained in:
parent
776275cf9c
commit
8b6cb2c7f2
2 changed files with 3 additions and 4 deletions
|
@ -217,11 +217,11 @@ struct ggml_sycl_device_info {
|
||||||
bool sycl_visible_devices_existed = false;
|
bool sycl_visible_devices_existed = false;
|
||||||
|
|
||||||
struct sycl_device_info {
|
struct sycl_device_info {
|
||||||
int cc; // compute capability
|
int cc; // compute capability
|
||||||
// int nsm; // number of streaming multiprocessors
|
// int nsm; // number of streaming multiprocessors
|
||||||
// size_t smpb; // max. shared memory per block
|
// size_t smpb; // max. shared memory per block
|
||||||
bool vmm; // virtual memory support
|
bool vmm; // virtual memory support
|
||||||
size_t total_vram;
|
size_t total_vram;
|
||||||
};
|
};
|
||||||
|
|
||||||
sycl_device_info devices[GGML_SYCL_MAX_DEVICES] = {};
|
sycl_device_info devices[GGML_SYCL_MAX_DEVICES] = {};
|
||||||
|
|
|
@ -17620,7 +17620,6 @@ struct llama_context * llama_new_context_with_model(
|
||||||
#elif defined(GGML_USE_SYCL)
|
#elif defined(GGML_USE_SYCL)
|
||||||
// with split_mode LLAMA_SPLIT_MODE_NONE or LLAMA_SPLIT_MODE_ROW, only the main GPU backend is used
|
// with split_mode LLAMA_SPLIT_MODE_NONE or LLAMA_SPLIT_MODE_ROW, only the main GPU backend is used
|
||||||
if (model->split_mode == LLAMA_SPLIT_MODE_NONE || model->split_mode == LLAMA_SPLIT_MODE_ROW) {
|
if (model->split_mode == LLAMA_SPLIT_MODE_NONE || model->split_mode == LLAMA_SPLIT_MODE_ROW) {
|
||||||
ggml_sycl_set_single_device(model->main_gpu);
|
|
||||||
ggml_backend_t backend = ggml_backend_sycl_init(model->main_gpu);
|
ggml_backend_t backend = ggml_backend_sycl_init(model->main_gpu);
|
||||||
if (backend == nullptr) {
|
if (backend == nullptr) {
|
||||||
LLAMA_LOG_ERROR("%s: failed to initialize SYCL%d backend\n", __func__, model->main_gpu);
|
LLAMA_LOG_ERROR("%s: failed to initialize SYCL%d backend\n", __func__, model->main_gpu);
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue