Only attempt to use mmap for the lora base model if it is supported
This commit is contained in:
parent
3df343b4f0
commit
63da54e016
1 changed files with 3 additions and 1 deletions
|
@ -1840,7 +1840,9 @@ int llama_apply_lora_from_file_internal(struct llama_context * ctx, const char *
|
||||||
model_loader->ggml_ctx = base_ctx;
|
model_loader->ggml_ctx = base_ctx;
|
||||||
|
|
||||||
// maybe this should in llama_model_loader
|
// maybe this should in llama_model_loader
|
||||||
model_loader->mapping.reset(new llama_mmap(&model_loader->file_loaders.at(0)->file, false));
|
if (model_loader->use_mmap) {
|
||||||
|
model_loader->mapping.reset(new llama_mmap(&model_loader->file_loaders.at(0)->file, /* prefetch */ false));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// read tensors and apply
|
// read tensors and apply
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue