diff --git a/examples/server/server.cpp b/examples/server/server.cpp
index d6fb84cd9..7d7d508f0 100644
--- a/examples/server/server.cpp
+++ b/examples/server/server.cpp
@@ -750,8 +750,11 @@ int main(int argc, char **argv)
Server svr;
- svr.Get("/", [](const Request &, Response &res)
- { res.set_content("
llama.cpp server works
", "text/html"); });
+ svr.Get("/", [](const Request &req, Response &res)
+ {
+ fprintf(stderr, "request: GET / [remote_addr: %s]", req.remote_addr.c_str());
+ res.set_content("llama.cpp server works
", "text/html");
+ });
svr.Post("/completion", [&llama](const Request &req, Response &res) {
if (llama.params.embedding) {
@@ -772,6 +775,8 @@ int main(int argc, char **argv)
return;
}
+ fprintf(stderr, "request: POST /completion [remote_addr: %s, stream: %s]", req.remote_addr.c_str(), llama.stream ? "true" : "false");
+
if (!llama.loadPrompt()) {
json data = {{"status", "error"}, {"reason", "Context too long."}};
res.set_content(
@@ -885,6 +890,7 @@ int main(int argc, char **argv)
svr.Post("/tokenize", [&llama](const Request &req, Response &res)
{
+ fprintf(stderr, "request: POST /tokenize [remote_addr: %s]", req.remote_addr.c_str());
json body = json::parse(req.body);
json data = {
{"tokens", ::llama_tokenize(llama.ctx, body["content"].get(), false) } };
@@ -893,6 +899,7 @@ int main(int argc, char **argv)
svr.Post("/embedding", [&llama](const Request &req, Response &res)
{
+ fprintf(stderr, "request: POST /embedding [remote_addr: %s]", req.remote_addr.c_str());
if(!llama.params.embedding) {
std::vector empty;
json data = {