Fix missing layer_inp_i names

This commit is contained in:
Branden Butler 2023-11-01 12:23:30 -05:00
parent b7599f7a56
commit 32078d6fe1

View file

@ -5729,6 +5729,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_mask = build_inp_KQ_mask(); struct ggml_tensor * KQ_mask = build_inp_KQ_mask();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * inpSA = inpL; struct ggml_tensor * inpSA = inpL;
// norm // norm
@ -5907,6 +5908,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_pos = build_inp_KQ_pos(); struct ggml_tensor * KQ_pos = build_inp_KQ_pos();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * inpSA = inpL; struct ggml_tensor * inpSA = inpL;
cur = llm_build_norm(ctx0, inpL, hparams, cur = llm_build_norm(ctx0, inpL, hparams,
@ -6016,6 +6018,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_mask = build_inp_KQ_mask(); struct ggml_tensor * KQ_mask = build_inp_KQ_mask();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * attn_norm; struct ggml_tensor * attn_norm;
attn_norm = llm_build_norm(ctx0, inpL, hparams, attn_norm = llm_build_norm(ctx0, inpL, hparams,
@ -6134,6 +6137,7 @@ struct llm_build_context {
cb(inpL, "inpL", -1); cb(inpL, "inpL", -1);
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
cur = llm_build_norm(ctx0, inpL, hparams, cur = llm_build_norm(ctx0, inpL, hparams,
model.layers[il].attn_norm, model.layers[il].attn_norm,
model.layers[il].attn_norm_b, model.layers[il].attn_norm_b,
@ -6221,6 +6225,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_mask = build_inp_KQ_mask(); struct ggml_tensor * KQ_mask = build_inp_KQ_mask();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * residual = inpL; struct ggml_tensor * residual = inpL;
cur = llm_build_norm(ctx0, inpL, hparams, cur = llm_build_norm(ctx0, inpL, hparams,
@ -6420,6 +6425,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_pos = build_inp_KQ_pos(); struct ggml_tensor * KQ_pos = build_inp_KQ_pos();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * inpSA = inpL; struct ggml_tensor * inpSA = inpL;
cur = llm_build_norm(ctx0, inpL, hparams, cur = llm_build_norm(ctx0, inpL, hparams,
@ -6699,6 +6705,7 @@ struct llm_build_context {
cb(inpL, "inp_norm", -1); cb(inpL, "inp_norm", -1);
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
cur = llm_build_norm(ctx0, inpL, hparams, cur = llm_build_norm(ctx0, inpL, hparams,
model.layers[il].attn_norm, model.layers[il].attn_norm,
model.layers[il].attn_norm_b, model.layers[il].attn_norm_b,
@ -6786,6 +6793,7 @@ struct llm_build_context {
struct ggml_tensor * KQ_pos = build_inp_KQ_pos(); struct ggml_tensor * KQ_pos = build_inp_KQ_pos();
for (int il = 0; il < n_layer; ++il) { for (int il = 0; il < n_layer; ++il) {
ggml_format_name(inpL, "layer_inp_%d", il); //MPI
struct ggml_tensor * attn_norm; struct ggml_tensor * attn_norm;
attn_norm = llm_build_norm(ctx0, inpL, hparams, attn_norm = llm_build_norm(ctx0, inpL, hparams,