kompute : remove llama_load_model_from_file_internal
This commit is contained in:
parent
dc08e512cc
commit
7e11fe0880
1 changed files with 4 additions and 9 deletions
13
llama.cpp
13
llama.cpp
|
@ -4146,6 +4146,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam
|
||||||
)
|
)
|
||||||
)) {
|
)) {
|
||||||
// disable Vulkan due to unsupported model architecture or quantization type
|
// disable Vulkan due to unsupported model architecture or quantization type
|
||||||
|
// TODO(cebtenzzre): propagate this error outside of llama_load_model_from_file
|
||||||
params.n_gpu_layers = 0;
|
params.n_gpu_layers = 0;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
@ -10118,11 +10119,9 @@ int64_t llama_time_us(void) {
|
||||||
return ggml_time_us();
|
return ggml_time_us();
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct llama_model * llama_load_model_from_file_internal(
|
struct llama_model * llama_load_model_from_file(
|
||||||
const char * path_model, struct llama_model_params * params_p
|
const char * path_model,
|
||||||
) {
|
struct llama_model_params params) {
|
||||||
auto & params = *params_p;
|
|
||||||
|
|
||||||
ggml_time_init();
|
ggml_time_init();
|
||||||
|
|
||||||
llama_model * model = new llama_model;
|
llama_model * model = new llama_model;
|
||||||
|
@ -10159,10 +10158,6 @@ static struct llama_model * llama_load_model_from_file_internal(
|
||||||
return model;
|
return model;
|
||||||
}
|
}
|
||||||
|
|
||||||
struct llama_model * llama_load_model_from_file(const char * path_model, struct llama_model_params params) {
|
|
||||||
return llama_load_model_from_file_internal(path_model, ¶ms);
|
|
||||||
}
|
|
||||||
|
|
||||||
void llama_free_model(struct llama_model * model) {
|
void llama_free_model(struct llama_model * model) {
|
||||||
delete model;
|
delete model;
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue