implement llama_max_devices() for RPC

This commit is contained in:
Radoslav Gerganov 2024-04-30 14:34:09 +03:00
parent 95c16c263c
commit 654c1cc279
2 changed files with 5 additions and 1 deletions

View file

@ -37,6 +37,8 @@ enum rpc_cmd {
GRAPH_COMPUTE,
};
#define GGML_RPC_MAX_SERVERS 16
// backend API
GGML_API GGML_CALL ggml_backend_t ggml_backend_rpc_init(const std::string & endpoint);
GGML_API GGML_CALL bool ggml_backend_is_rpc(ggml_backend_t backend);

View file

@ -15480,7 +15480,9 @@ struct llama_model_quantize_params llama_model_quantize_default_params() {
}
size_t llama_max_devices(void) {
#if defined(GGML_USE_METAL)
#if defined(GGML_USE_RPC)
return GGML_RPC_MAX_SERVERS;
#elif defined(GGML_USE_METAL)
return 1;
#elif defined(GGML_USE_CUDA)
return GGML_CUDA_MAX_DEVICES;