added field to show recent seed

This commit is contained in:
Concedo 2024-01-02 15:35:04 +08:00
parent 9e0dee769b
commit 94e68fe474
5 changed files with 12 additions and 31 deletions

View file

@ -1,29 +0,0 @@
name: Python check requirements.txt
on:
push:
paths:
- 'scripts/check-requirements.sh'
- 'convert*.py'
- 'requirements.txt'
- 'requirements/*.txt'
pull_request:
paths:
- 'scripts/check-requirements.sh'
- 'convert*.py'
- 'requirements.txt'
- 'requirements/*.txt'
jobs:
python-check-requirements:
runs-on: ubuntu-latest
name: check-requirements
steps:
- name: Check out source repository
uses: actions/checkout@v3
- name: Set up Python environment
uses: actions/setup-python@v4
with:
python-version: "3.11"
- name: Run check-requirements.sh script
run: bash scripts/check-requirements.sh nocleanup

View file

@ -217,6 +217,10 @@ extern "C"
int get_last_token_count() {
return last_token_count;
}
int get_last_seed()
{
return last_seed;
}
int get_total_gens() {
return total_gens;
}

View file

@ -102,5 +102,6 @@ extern bool generation_finished;
extern float last_eval_time;
extern float last_process_time;
extern int last_token_count;
extern int last_seed;
extern int total_gens;
extern stop_reason last_stop_reason;

View file

@ -39,6 +39,7 @@ bool generation_finished;
float last_process_time = 0;
float last_eval_time = 0;
int last_token_count = 0;
int last_seed = -1;
int total_gens = 0;
stop_reason last_stop_reason = stop_reason::INVALID;
std::vector<std::string> generated_tokens;
@ -1531,7 +1532,7 @@ generation_outputs gpttype_generate(const generation_inputs inputs, generation_o
}
if (kcpp_params->seed <= 0 || kcpp_params->seed==0xFFFFFFFF)
{
kcpp_params->seed = time(NULL);
kcpp_params->seed = (((uint32_t)time(NULL)) % 1000000);
}
// tokenize the prompt
@ -2051,6 +2052,7 @@ generation_outputs gpttype_generate(const generation_inputs inputs, generation_o
last_eval_time = pt2;
last_process_time = pt1;
last_token_count = realnpredict;
last_seed = kcpp_params->seed;
total_gens += 1;
snprintf(output.text, sizeof(output.text), "%s", concat_output.c_str());

View file

@ -226,6 +226,7 @@ def init_library():
handle.get_last_eval_time.restype = ctypes.c_float
handle.get_last_process_time.restype = ctypes.c_float
handle.get_last_token_count.restype = ctypes.c_int
handle.get_last_seed.restype = ctypes.c_int
handle.get_total_gens.restype = ctypes.c_int
handle.get_last_stop_reason.restype = ctypes.c_int
handle.abort_generate.restype = ctypes.c_bool
@ -793,7 +794,8 @@ Enter Prompt:<br>
lastc = handle.get_last_token_count()
totalgens = handle.get_total_gens()
stopreason = handle.get_last_stop_reason()
response_body = (json.dumps({"last_process":lastp,"last_eval":laste,"last_token_count":lastc, "total_gens":totalgens, "stop_reason":stopreason, "queue":requestsinqueue, "idle":(0 if modelbusy.locked() else 1), "hordeexitcounter":exitcounter}).encode())
lastseed = handle.get_last_seed()
response_body = (json.dumps({"last_process":lastp,"last_eval":laste,"last_token_count":lastc, "last_seed":lastseed, "total_gens":totalgens, "stop_reason":stopreason, "queue":requestsinqueue, "idle":(0 if modelbusy.locked() else 1), "hordeexitcounter":exitcounter}).encode())
elif self.path.endswith('/api/extra/generate/check'):
pendtxtStr = ""
@ -2214,6 +2216,7 @@ def unload_libs():
del handle.get_last_eval_time
del handle.get_last_process_time
del handle.get_last_token_count
del handle.get_last_seed
del handle.get_total_gens
del handle.get_last_stop_reason
del handle.abort_generate