readme : trim trailing whitespace
This commit is contained in:
parent
7975f43eb1
commit
df54adabea
1 changed files with 3 additions and 3 deletions
|
@ -1,6 +1,6 @@
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
The `rpc-server` allows running a `ggml` backend on a remote host.
|
The `rpc-server` allows running a `ggml` backend on a remote host.
|
||||||
The RPC backend communicates with one or several instances of `rpc-server` and offloads computations to them.
|
The RPC backend communicates with one or several instances of `rpc-server` and offloads computations to them.
|
||||||
This can be used for distributed LLM inference with `llama.cpp` in the following way:
|
This can be used for distributed LLM inference with `llama.cpp` in the following way:
|
||||||
|
|
||||||
|
@ -37,9 +37,9 @@ cd build-rpc-cuda
|
||||||
cmake .. -DLLAMA_CUDA=ON -DLLAMA_RPC=ON
|
cmake .. -DLLAMA_CUDA=ON -DLLAMA_RPC=ON
|
||||||
make -j
|
make -j
|
||||||
```
|
```
|
||||||
|
|
||||||
Then, start the `rpc-server` with the backend:
|
Then, start the `rpc-server` with the backend:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ bin/rpc-server 0.0.0.0 50052
|
$ bin/rpc-server 0.0.0.0 50052
|
||||||
create_backend: using CUDA backend
|
create_backend: using CUDA backend
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue