make help text load faster

This commit is contained in:
anon 2023-06-02 17:52:04 -03:00
parent a6ed390cc6
commit 05a5a485b8

View file

@ -756,8 +756,6 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
int main(int argc, char **argv) int main(int argc, char **argv)
{ {
llama_init_backend();
// own arguments required by this example // own arguments required by this example
gpt_params params; gpt_params params;
server_params sparams; server_params sparams;
@ -775,6 +773,8 @@ int main(int argc, char **argv)
params.model_alias = params.model; params.model_alias = params.model;
} }
llama_init_backend();
fprintf(stderr, "%s: build = %d (%s)\n", __func__, BUILD_NUMBER, BUILD_COMMIT); fprintf(stderr, "%s: build = %d (%s)\n", __func__, BUILD_NUMBER, BUILD_COMMIT);
fprintf(stderr, "system_info: n_threads = %d / %d | %s\n\n", params.n_threads, fprintf(stderr, "system_info: n_threads = %d / %d | %s\n\n", params.n_threads,
std::thread::hardware_concurrency(), llama_print_system_info()); std::thread::hardware_concurrency(), llama_print_system_info());