llama : disable 405B max_nodes path due to lack of complaints
ggml-ci
This commit is contained in:
parent
7e27c17572
commit
2d74714535
1 changed files with 4 additions and 4 deletions
|
@ -3657,10 +3657,10 @@ namespace GGUFMeta {
|
|||
using llama_buf_map = std::unordered_map<uint32_t, ggml_backend_buffer_t>;
|
||||
|
||||
// TODO: update when needed or think of some clever automatic way to do this
|
||||
static size_t llama_model_max_nodes(const llama_model & model) {
|
||||
if (model.arch == LLM_ARCH_LLAMA && model.hparams.n_layer > 400) { // llama-3 405B
|
||||
return 32768;
|
||||
}
|
||||
static size_t llama_model_max_nodes(const llama_model & /*model*/) {
|
||||
//if (model.arch == LLM_ARCH_LLAMA && model.hparams.n_layer > ??) { // llama-3 405B
|
||||
// return 32768;
|
||||
//}
|
||||
|
||||
return 8192;
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue