implement llama_max_devices() for RPC
This commit is contained in:
parent
95c16c263c
commit
654c1cc279
2 changed files with 5 additions and 1 deletions
|
@ -37,6 +37,8 @@ enum rpc_cmd {
|
|||
GRAPH_COMPUTE,
|
||||
};
|
||||
|
||||
#define GGML_RPC_MAX_SERVERS 16
|
||||
|
||||
// backend API
|
||||
GGML_API GGML_CALL ggml_backend_t ggml_backend_rpc_init(const std::string & endpoint);
|
||||
GGML_API GGML_CALL bool ggml_backend_is_rpc(ggml_backend_t backend);
|
||||
|
|
|
@ -15480,7 +15480,9 @@ struct llama_model_quantize_params llama_model_quantize_default_params() {
|
|||
}
|
||||
|
||||
size_t llama_max_devices(void) {
|
||||
#if defined(GGML_USE_METAL)
|
||||
#if defined(GGML_USE_RPC)
|
||||
return GGML_RPC_MAX_SERVERS;
|
||||
#elif defined(GGML_USE_METAL)
|
||||
return 1;
|
||||
#elif defined(GGML_USE_CUDA)
|
||||
return GGML_CUDA_MAX_DEVICES;
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue