diff --git a/koboldcpp.py b/koboldcpp.py index d65a22285..c7a07cc93 100755 --- a/koboldcpp.py +++ b/koboldcpp.py @@ -394,7 +394,7 @@ maxhordelen = 256 modelbusy = threading.Lock() requestsinqueue = 0 defaultport = 5001 -KcppVersion = "1.53" +KcppVersion = "1.53.1" showdebug = True showsamplerwarning = True showmaxctxwarning = True @@ -766,7 +766,7 @@ Enter Prompt:
elif self.path.endswith('/api/extra/generate/check'): pendtxtStr = "" - if requestsinqueue==0 and totalgens>0 and (not args.multiuser or currentusergenkey==""): + if requestsinqueue==0 and totalgens>0 and currentusergenkey=="": pendtxt = handle.get_pending_output() pendtxtStr = ctypes.string_at(pendtxt).decode("UTF-8","ignore") response_body = (json.dumps({"results": [{"text": pendtxtStr}]}).encode()) @@ -861,7 +861,7 @@ Enter Prompt:
multiuserkey = "" if totalgens>0: - if (multiuserkey=="" and multiuserkey==currentusergenkey) or (multiuserkey=="" and requestsinqueue==0 and not args.multiuser) or (multiuserkey!="" and multiuserkey==currentusergenkey): #avoid leaking prompts in multiuser + if (multiuserkey=="" and multiuserkey==currentusergenkey and requestsinqueue==0) or (multiuserkey!="" and multiuserkey==currentusergenkey): #avoid leaking prompts in multiuser pendtxt = handle.get_pending_output() pendtxtStr = ctypes.string_at(pendtxt).decode("UTF-8","ignore") response_body = (json.dumps({"results": [{"text": pendtxtStr}]}).encode())