Update default values for n_embd_head_k and n_embd_head_v
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
parent
f56cbce6b7
commit
522e534903
1 changed files with 2 additions and 2 deletions
|
@ -2681,10 +2681,10 @@ static void llm_load_hparams(
|
||||||
// gpt-j n_rot = rotary_dim
|
// gpt-j n_rot = rotary_dim
|
||||||
}
|
}
|
||||||
|
|
||||||
hparams.n_embd_head_k = hparams.n_embd / hparams.n_head_kv;
|
hparams.n_embd_head_k = hparams.n_embd / hparams.n_head;
|
||||||
ml.get_key(LLM_KV_ATTENTION_KEY_LENGTH, hparams.n_embd_head_k, false);
|
ml.get_key(LLM_KV_ATTENTION_KEY_LENGTH, hparams.n_embd_head_k, false);
|
||||||
|
|
||||||
hparams.n_embd_head_v = hparams.n_embd / hparams.n_head_kv;
|
hparams.n_embd_head_v = hparams.n_embd / hparams.n_head;
|
||||||
ml.get_key(LLM_KV_ATTENTION_VALUE_LENGTH, hparams.n_embd_head_v, false);
|
ml.get_key(LLM_KV_ATTENTION_VALUE_LENGTH, hparams.n_embd_head_v, false);
|
||||||
|
|
||||||
// arch-specific KVs
|
// arch-specific KVs
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue