From df54adabeade8d4b909cea09dddda4cf3d100c83 Mon Sep 17 00:00:00 2001 From: Radoslav Gerganov Date: Fri, 10 May 2024 14:35:24 +0300 Subject: [PATCH] readme : trim trailing whitespace --- examples/rpc/README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/examples/rpc/README.md b/examples/rpc/README.md index bd8587c1a..2c974d1a5 100644 --- a/examples/rpc/README.md +++ b/examples/rpc/README.md @@ -1,6 +1,6 @@ ## Overview -The `rpc-server` allows running a `ggml` backend on a remote host. +The `rpc-server` allows running a `ggml` backend on a remote host. The RPC backend communicates with one or several instances of `rpc-server` and offloads computations to them. This can be used for distributed LLM inference with `llama.cpp` in the following way: @@ -37,9 +37,9 @@ cd build-rpc-cuda cmake .. -DLLAMA_CUDA=ON -DLLAMA_RPC=ON make -j ``` - + Then, start the `rpc-server` with the backend: - + ```bash $ bin/rpc-server 0.0.0.0 50052 create_backend: using CUDA backend