llama: dbrx: first add the residuals and then do the norm
This commit is contained in:
parent
f30a73bb01
commit
ea8b58c6cd
1 changed files with 5 additions and 5 deletions
10
llama.cpp
10
llama.cpp
|
@ -7122,16 +7122,16 @@ struct llm_build_context {
|
||||||
inpSA = ggml_get_rows(ctx0, inpSA, inp_out_ids);
|
inpSA = ggml_get_rows(ctx0, inpSA, inp_out_ids);
|
||||||
}
|
}
|
||||||
|
|
||||||
cur = llm_build_norm(ctx0, cur, hparams,
|
|
||||||
model.layers[il].attn_out_norm, NULL,
|
|
||||||
LLM_NORM, cb, il);
|
|
||||||
cb(cur, "attn_out_norm", il);
|
|
||||||
|
|
||||||
struct ggml_tensor * ffn_inp = ggml_add(ctx0, cur, inpSA);
|
struct ggml_tensor * ffn_inp = ggml_add(ctx0, cur, inpSA);
|
||||||
cb(ffn_inp, "ffn_inp", il);
|
cb(ffn_inp, "ffn_inp", il);
|
||||||
|
|
||||||
// feed-forward network
|
// feed-forward network
|
||||||
// MoE branch
|
// MoE branch
|
||||||
|
cur = llm_build_norm(ctx0, cur, hparams,
|
||||||
|
model.layers[il].attn_out_norm, NULL,
|
||||||
|
LLM_NORM, cb, il);
|
||||||
|
cb(cur, "attn_out_norm", il);
|
||||||
|
|
||||||
cur = build_moe(cur, n_tokens, il);
|
cur = build_moe(cur, n_tokens, il);
|
||||||
|
|
||||||
cur = ggml_add(ctx0, cur, ffn_inp);
|
cur = ggml_add(ctx0, cur, ffn_inp);
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue