diff --git a/examples/server/server.cpp b/examples/server/server.cpp index d6fb84cd9..7d7d508f0 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -750,8 +750,11 @@ int main(int argc, char **argv) Server svr; - svr.Get("/", [](const Request &, Response &res) - { res.set_content("

llama.cpp server works

", "text/html"); }); + svr.Get("/", [](const Request &req, Response &res) + { + fprintf(stderr, "request: GET / [remote_addr: %s]", req.remote_addr.c_str()); + res.set_content("

llama.cpp server works

", "text/html"); + }); svr.Post("/completion", [&llama](const Request &req, Response &res) { if (llama.params.embedding) { @@ -772,6 +775,8 @@ int main(int argc, char **argv) return; } + fprintf(stderr, "request: POST /completion [remote_addr: %s, stream: %s]", req.remote_addr.c_str(), llama.stream ? "true" : "false"); + if (!llama.loadPrompt()) { json data = {{"status", "error"}, {"reason", "Context too long."}}; res.set_content( @@ -885,6 +890,7 @@ int main(int argc, char **argv) svr.Post("/tokenize", [&llama](const Request &req, Response &res) { + fprintf(stderr, "request: POST /tokenize [remote_addr: %s]", req.remote_addr.c_str()); json body = json::parse(req.body); json data = { {"tokens", ::llama_tokenize(llama.ctx, body["content"].get(), false) } }; @@ -893,6 +899,7 @@ int main(int argc, char **argv) svr.Post("/embedding", [&llama](const Request &req, Response &res) { + fprintf(stderr, "request: POST /embedding [remote_addr: %s]", req.remote_addr.c_str()); if(!llama.params.embedding) { std::vector empty; json data = {