From 03da419fc069180fe292dad803ea4f2290bd38dc Mon Sep 17 00:00:00 2001 From: Pierrick HYMBERT Date: Sat, 6 Apr 2024 20:43:46 +0200 Subject: [PATCH] llama: dbrx: remove wrong attn output layer in model arch --- llama.cpp | 1 - 1 file changed, 1 deletion(-) diff --git a/llama.cpp b/llama.cpp index 360d4b086..944317a8e 100644 --- a/llama.cpp +++ b/llama.cpp @@ -935,7 +935,6 @@ static const std::map> LLM_TENSOR_NA { LLM_TENSOR_OUTPUT_NORM, "output_norm" }, { LLM_TENSOR_OUTPUT, "output" }, { LLM_TENSOR_ATTN_QKV, "blk.%d.attn_qkv" }, - { LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" }, { LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" }, { LLM_TENSOR_ATTN_NORM_2, "blk.%d.attn_norm_2" }, { LLM_TENSOR_FFN_GATE_INP, "blk.%d.ffn_gate_inp" },