llama: dbrx: hardcode nn.LayerNorm epsilon
This commit is contained in:
parent
74e6d876f6
commit
f8f97e74f9
1 changed files with 1 additions and 0 deletions
|
@ -3986,6 +3986,7 @@ static void llm_load_hparams(
|
||||||
case LLM_ARCH_DBRX:
|
case LLM_ARCH_DBRX:
|
||||||
{
|
{
|
||||||
ml.get_key(LLM_KV_ATTENTION_CLAMP_KQV, hparams.f_clamp_kqv);
|
ml.get_key(LLM_KV_ATTENTION_CLAMP_KQV, hparams.f_clamp_kqv);
|
||||||
|
hparams.f_norm_eps = 1.e-5; // REVIEW is that OK ? https://pytorch.org/docs/stable/generated/torch.nn.LayerNorm.html
|
||||||
|
|
||||||
switch (hparams.n_layer) {
|
switch (hparams.n_layer) {
|
||||||
case 40: model.type = e_model::MODEL_132B; break;
|
case 40: model.type = e_model::MODEL_132B; break;
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue