llama : change llama_load_model_from_file -> llama_model_load_from_file

ggml-ci
This commit is contained in:
Georgi Gerganov 2025-01-03 14:42:28 +02:00
parent 0261d4f02f
commit 138255e761
No known key found for this signature in database
GPG key ID: 449E073F9DC10735
21 changed files with 32 additions and 21 deletions

View file

@ -846,7 +846,7 @@ struct common_init_result common_init_from_params(common_params & params) {
} else if (!params.model_url.empty()) { } else if (!params.model_url.empty()) {
model = common_load_model_from_url(params.model_url, params.model, params.hf_token, mparams); model = common_load_model_from_url(params.model_url, params.model, params.hf_token, mparams);
} else { } else {
model = llama_load_model_from_file(params.model.c_str(), mparams); model = llama_model_load_from_file(params.model.c_str(), mparams);
} }
if (model == NULL) { if (model == NULL) {
@ -1411,7 +1411,7 @@ struct llama_model * common_load_model_from_url(
} }
} }
return llama_load_model_from_file(local_path.c_str(), params); return llama_model_load_from_file(local_path.c_str(), params);
} }
struct llama_model * common_load_model_from_hf( struct llama_model * common_load_model_from_hf(

View file

@ -38,7 +38,7 @@ int main(int argc, char ** argv) {
llama_model_params model_params = common_model_params_to_llama(params); llama_model_params model_params = common_model_params_to_llama(params);
llama_model * model = llama_load_model_from_file(params.model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params.model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
fprintf(stderr , "%s: error: unable to load model\n" , __func__); fprintf(stderr , "%s: error: unable to load model\n" , __func__);

View file

@ -41,7 +41,7 @@ int main(int argc, char ** argv) {
llama_model_params model_params = common_model_params_to_llama(params); llama_model_params model_params = common_model_params_to_llama(params);
llama_model * model = llama_load_model_from_file(params.model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params.model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
LOG_ERR("%s: error: unable to load model\n" , __func__); LOG_ERR("%s: error: unable to load model\n" , __func__);

View file

@ -165,7 +165,7 @@ int main(int argc, char * argv[]) {
llama_backend_init(); llama_backend_init();
llama_model * model = llama_load_model_from_file(params.model.c_str(), mparams); llama_model * model = llama_model_load_from_file(params.model.c_str(), mparams);
// create generation context // create generation context
llama_context * ctx = llama_new_context_with_model(model, cparams); llama_context * ctx = llama_new_context_with_model(model, cparams);

View file

@ -1529,7 +1529,7 @@ int main(int argc, char ** argv) {
llama_model_free(lmodel); llama_model_free(lmodel);
} }
lmodel = llama_load_model_from_file(inst.model.c_str(), inst.to_llama_mparams()); lmodel = llama_model_load_from_file(inst.model.c_str(), inst.to_llama_mparams());
if (lmodel == NULL) { if (lmodel == NULL) {
fprintf(stderr, "%s: error: failed to load model '%s'\n", __func__, inst.model.c_str()); fprintf(stderr, "%s: error: failed to load model '%s'\n", __func__, inst.model.c_str());
return 1; return 1;

View file

@ -221,7 +221,7 @@ static struct llama_model * llava_init(common_params * params) {
llama_model_params model_params = common_model_params_to_llama(*params); llama_model_params model_params = common_model_params_to_llama(*params);
llama_model * model = llama_load_model_from_file(params->model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params->model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
LOG_ERR("%s: unable to load model\n" , __func__); LOG_ERR("%s: unable to load model\n" , __func__);
return NULL; return NULL;

View file

@ -31,7 +31,7 @@ static struct llama_model * llava_init(common_params * params) {
llama_model_params model_params = common_model_params_to_llama(*params); llama_model_params model_params = common_model_params_to_llama(*params);
llama_model * model = llama_load_model_from_file(params->model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params->model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
LOG_ERR("%s: unable to load model\n" , __func__); LOG_ERR("%s: unable to load model\n" , __func__);
return NULL; return NULL;

View file

@ -310,7 +310,7 @@ static struct llama_model * llava_init(common_params * params) {
llama_model_params model_params = common_model_params_to_llama(*params); llama_model_params model_params = common_model_params_to_llama(*params);
llama_model * model = llama_load_model_from_file(params->model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params->model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
LOG_ERR("%s: unable to load model\n" , __func__); LOG_ERR("%s: unable to load model\n" , __func__);
return NULL; return NULL;

View file

@ -63,7 +63,7 @@ int main(int argc, char ** argv) {
llama_model_params model_params = common_model_params_to_llama(params); llama_model_params model_params = common_model_params_to_llama(params);
llama_model * model = llama_load_model_from_file(params.model.c_str(), model_params); llama_model * model = llama_model_load_from_file(params.model.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
LOG_ERR("%s: unable to load model\n" , __func__); LOG_ERR("%s: unable to load model\n" , __func__);

View file

@ -309,7 +309,7 @@ int main(int argc, char ** argv) {
auto mparams = llama_model_default_params(); auto mparams = llama_model_default_params();
mparams.use_mlock = false; mparams.use_mlock = false;
model = llama_load_model_from_file(params.model.c_str(), mparams); model = llama_model_load_from_file(params.model.c_str(), mparams);
if (model == NULL) { if (model == NULL) {
fprintf(stderr, "%s: error: failed to load model '%s'\n", __func__, params.model.c_str()); fprintf(stderr, "%s: error: failed to load model '%s'\n", __func__, params.model.c_str());

View file

@ -664,7 +664,7 @@ class LlamaData {
"\r%*s" "\r%*s"
"\rLoading model", "\rLoading model",
get_terminal_width(), " "); get_terminal_width(), " ");
llama_model_ptr model(llama_load_model_from_file(opt.model_.c_str(), opt.model_params)); llama_model_ptr model(llama_model_load_from_file(opt.model_.c_str(), opt.model_params));
if (!model) { if (!model) {
printe("%s: error: unable to load model from file: %s\n", __func__, opt.model_.c_str()); printe("%s: error: unable to load model from file: %s\n", __func__, opt.model_.c_str());
} }

View file

@ -69,7 +69,7 @@ int main(int argc, char ** argv) {
llama_model_params model_params = llama_model_default_params(); llama_model_params model_params = llama_model_default_params();
model_params.n_gpu_layers = ngl; model_params.n_gpu_layers = ngl;
llama_model * model = llama_load_model_from_file(model_path.c_str(), model_params); llama_model * model = llama_model_load_from_file(model_path.c_str(), model_params);
if (!model) { if (!model) {
fprintf(stderr , "%s: error: unable to load model\n" , __func__); fprintf(stderr , "%s: error: unable to load model\n" , __func__);
return 1; return 1;

View file

@ -83,7 +83,7 @@ int main(int argc, char ** argv) {
llama_model_params model_params = llama_model_default_params(); llama_model_params model_params = llama_model_default_params();
model_params.n_gpu_layers = ngl; model_params.n_gpu_layers = ngl;
llama_model * model = llama_load_model_from_file(model_path.c_str(), model_params); llama_model * model = llama_model_load_from_file(model_path.c_str(), model_params);
if (model == NULL) { if (model == NULL) {
fprintf(stderr , "%s: error: unable to load model\n" , __func__); fprintf(stderr , "%s: error: unable to load model\n" , __func__);

View file

@ -333,7 +333,7 @@ int main(int raw_argc, char ** raw_argv) {
llama_model_params model_params = llama_model_default_params(); llama_model_params model_params = llama_model_default_params();
model_params.vocab_only = true; model_params.vocab_only = true;
llama_model * model = llama_load_model_from_file(model_path, model_params); llama_model * model = llama_model_load_from_file(model_path, model_params);
if (!model) { if (!model) {
fprintf(stderr, "Error: could not load model from file '%s'.\n", model_path); fprintf(stderr, "Error: could not load model from file '%s'.\n", model_path);
return 1; return 1;

View file

@ -413,7 +413,12 @@ extern "C" {
// Call once at the end of the program - currently only used for MPI // Call once at the end of the program - currently only used for MPI
LLAMA_API void llama_backend_free(void); LLAMA_API void llama_backend_free(void);
LLAMA_API struct llama_model * llama_load_model_from_file( DEPRECATED(LLAMA_API struct llama_model * llama_load_model_from_file(
const char * path_model,
struct llama_model_params params),
"use llama_model_load_from_file instead");
LLAMA_API struct llama_model * llama_model_load_from_file(
const char * path_model, const char * path_model,
struct llama_model_params params); struct llama_model_params params);

View file

@ -11456,6 +11456,12 @@ int64_t llama_time_us(void) {
struct llama_model * llama_load_model_from_file( struct llama_model * llama_load_model_from_file(
const char * path_model, const char * path_model,
struct llama_model_params params) { struct llama_model_params params) {
return llama_model_load_from_file(path_model, params);
}
struct llama_model * llama_model_load_from_file(
const char * path_model,
struct llama_model_params params) {
ggml_time_init(); ggml_time_init();
llama_model * model = new llama_model; llama_model * model = new llama_model;

View file

@ -13,7 +13,7 @@ int main(int argc, char ** argv) {
std::thread([&model_path]() { std::thread([&model_path]() {
llama_backend_init(); llama_backend_init();
auto * model = llama_load_model_from_file(model_path, llama_model_default_params()); auto * model = llama_model_load_from_file(model_path, llama_model_default_params());
auto * ctx = llama_new_context_with_model(model, llama_context_default_params()); auto * ctx = llama_new_context_with_model(model, llama_context_default_params());
llama_free(ctx); llama_free(ctx);
llama_model_free(model); llama_model_free(model);

View file

@ -21,7 +21,7 @@ int main(int argc, char *argv[] ) {
(void) ctx; (void) ctx;
return progress > 0.50; return progress > 0.50;
}; };
auto * model = llama_load_model_from_file(model_path, params); auto * model = llama_model_load_from_file(model_path, params);
llama_backend_free(); llama_backend_free();
return model == nullptr ? EXIT_SUCCESS : EXIT_FAILURE; return model == nullptr ? EXIT_SUCCESS : EXIT_FAILURE;
} }

View file

@ -152,7 +152,7 @@ int main(int argc, char **argv) {
mparams.vocab_only = true; mparams.vocab_only = true;
model = llama_load_model_from_file(fname.c_str(), mparams); model = llama_model_load_from_file(fname.c_str(), mparams);
if (model == NULL) { if (model == NULL) {
fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str()); fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str());

View file

@ -46,7 +46,7 @@ int main(int argc, char **argv) {
mparams.vocab_only = true; mparams.vocab_only = true;
model = llama_load_model_from_file(fname.c_str(), mparams); model = llama_model_load_from_file(fname.c_str(), mparams);
if (model == NULL) { if (model == NULL) {
fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str()); fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str());

View file

@ -34,7 +34,7 @@ int main(int argc, char ** argv) {
mparams.vocab_only = true; mparams.vocab_only = true;
model = llama_load_model_from_file(fname.c_str(), mparams); model = llama_model_load_from_file(fname.c_str(), mparams);
if (model == NULL) { if (model == NULL) {
fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str()); fprintf(stderr, "%s: error: failed to load vocab '%s'\n", __func__, fname.c_str());