Some more code cleanup

This commit is contained in:
InconsolableCellist 2023-03-28 16:59:27 -06:00
parent 13addf2a78
commit 13b4c05d66

View file

@ -87,39 +87,42 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):
super().__init__(*args, **kwargs)
def do_GET(self):
if self.path=="/" or self.path.startswith('/?') or self.path.startswith('?'):
global maxctx, maxlen, friendlymodelname
if self.path in ["/", "/?"] or self.path.startswith('/?'):
if self.embedded_kailite is None:
self.send_response(200)
self.end_headers()
self.wfile.write(b'Embedded Kobold Lite is not found.<br>You will have to connect via the main KoboldAI client, or <a href=\'https://lite.koboldai.net?local=1&port='+str(self.port).encode()+b'\'>use this URL</a> to connect.')
response_body = (
b"Embedded Kobold Lite is not found.<br>You will have to connect via the main KoboldAI client, or "
b"<a href='https://lite.koboldai.net?local=1&port={0}'>use this URL</a> to connect.").format(self.port).encode()
else:
response_body = self.embedded_kailite
self.send_response(200)
self.send_header('Content-Length', str(len(response_body)))
self.end_headers()
self.wfile.write(self.embedded_kailite)
self.wfile.write(response_body)
return
if self.path.endswith('/api/v1/model/') or self.path.endswith('/api/latest/model/') or self.path.endswith('/api/v1/model') or self.path.endswith('/api/latest/model'):
self.path = self.path.rstrip('/')
if self.path.endswith(('/api/v1/model', '/api/latest/model')):
self.send_response(200)
self.end_headers()
global friendlymodelname
self.wfile.write(json.dumps({"result": friendlymodelname }).encode())
result = {'result': friendlymodelname }
self.wfile.write(json.dumps(result).encode())
return
if self.path.endswith('/api/v1/config/max_length/') or self.path.endswith('/api/latest/config/max_length/') or self.path.endswith('/api/v1/config/max_length') or self.path.endswith('/api/latest/config/max_length'):
if self.path.endswith(('/api/v1/config/max_length', '/api/latest/config/max_length')):
self.send_response(200)
self.end_headers()
global maxlen
self.wfile.write(json.dumps({"value":maxlen}).encode())
self.wfile.write(json.dumps({"value": maxlen}).encode())
return
if self.path.endswith('/api/v1/config/max_context_length/') or self.path.endswith('/api/latest/config/max_context_length/') or self.path.endswith('/api/v1/config/max_context_length') or self.path.endswith('/api/latest/config/max_context_length'):
if self.path.endswith(('/api/v1/config/max_context_length', '/api/latest/config/max_context_length')):
self.send_response(200)
self.end_headers()
global maxctx
self.wfile.write(json.dumps({"value":maxctx}).encode())
self.wfile.write(json.dumps({"value": maxctx}).encode())
return
if self.path.endswith('/api/v1/config/soft_prompt') or self.path.endswith('/api/v1/config/soft_prompt/') or self.path.endswith('/api/latest/config/soft_prompt') or self.path.endswith('/api/latest/config/soft_prompt/'):
if self.path.endswith(('/api/v1/config/soft_prompt', '/api/latest/config/soft_prompt')):
self.send_response(200)
self.end_headers()
self.wfile.write(json.dumps({"value":""}).encode())
@ -130,12 +133,14 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):
rp = 'Error: HTTP Server is running, but this endpoint does not exist. Please check the URL.'
self.wfile.write(rp.encode())
return
def do_POST(self):
global modelbusy
content_length = int(self.headers['Content-Length'])
body = self.rfile.read(content_length)
basic_api_flag = False
kai_api_flag = False
self.path = self.path.rstrip('/')
if modelbusy:
self.send_response(503)
@ -146,11 +151,10 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):
}}).encode())
return
basic_api_flag = False
kai_api_flag = False
if self.path.endswith('/request') or self.path.endswith('/request'):
if self.path.endswith('/request'):
basic_api_flag = True
if self.path.endswith('/api/v1/generate/') or self.path.endswith('/api/latest/generate/') or self.path.endswith('/api/v1/generate') or self.path.endswith('/api/latest/generate'):
if self.path.endswith(('/api/v1/generate', '/api/latest/generate')):
kai_api_flag = True
if basic_api_flag or kai_api_flag:
@ -170,7 +174,6 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):
fullprompt = genparams.get('text', "")
newprompt = fullprompt
recvtxt = ""
if kai_api_flag:
recvtxt = generate(
@ -207,7 +210,6 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):
self.wfile.write(json.dumps(res).encode())
modelbusy = False
return
self.send_response(404)
self.end_headers()