From 41f7a5004ac98aed6217f0f8549c902123011ce6 Mon Sep 17 00:00:00 2001 From: jwj7140 Date: Wed, 5 Jul 2023 01:02:05 +0900 Subject: [PATCH] fix bug & add truncation return --- examples/server/api_like_OAI.py | 13 ++++++++++++- 1 file changed, 12 insertions(+), 1 deletion(-) diff --git a/examples/server/api_like_OAI.py b/examples/server/api_like_OAI.py index 4567f5eed..e07324ba7 100755 --- a/examples/server/api_like_OAI.py +++ b/examples/server/api_like_OAI.py @@ -69,7 +69,10 @@ def make_postData(body, chat=False, stream=False): if(is_present(body, "mirostat_eta")): postData["mirostat_eta"] = body["mirostat_eta"] if(is_present(body, "seed")): postData["seed"] = body["seed"] if(is_present(body, "logit_bias")): postData["logit_bias"] = [[int(token), body["logit_bias"][token]] for token in body["logit_bias"].keys()] - postData["stop"] = [args.stop] + if (args.stop != ""): + postData["stop"] = [args.stop] + else: + postData["stop"] = [] if(is_present(body, "stop")): postData["stop"] += body["stop"] postData["n_keep"] = -1 postData["stream"] = stream @@ -81,6 +84,7 @@ def make_resData(data, chat=False, promptToken=[]): "id": "chatcmpl" if (chat) else "cmpl", "object": "chat.completion" if (chat) else "text_completion", "created": int(time.time()), + "truncated": data["truncated"], "model": "LLaMA_CPP", "usage": { "prompt_tokens": data["tokens_evaluated"], @@ -196,6 +200,7 @@ def completion(): if (not stream): data = requests.request("POST", urllib.parse.urljoin(args.llama_api, "/completion"), data=json.dumps(postData)) + print(data.json()) resData = make_resData(data.json(), chat=False, promptToken=promptToken) return jsonify(resData) else: @@ -210,5 +215,11 @@ def completion(): return Response(generate(), mimetype='text/event-stream') + + + + + + if __name__ == '__main__': app.run(args.host, port=args.port)