Readd variable attn_k, attn_q, attn_o after merge
This commit is contained in:
parent
fb2b9ea667
commit
596a4aec86
1 changed files with 3 additions and 0 deletions
|
@ -17328,6 +17328,9 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
|
||||||
n_attn_layer *= 3;
|
n_attn_layer *= 3;
|
||||||
}
|
}
|
||||||
GGML_ASSERT((qs.n_attention_wv == n_attn_layer) && "n_attention_wv is unexpected");
|
GGML_ASSERT((qs.n_attention_wv == n_attn_layer) && "n_attention_wv is unexpected");
|
||||||
|
GGML_ASSERT((qs.n_attention_wk == n_attn_layer) && "n_attention_wk is unexpected");
|
||||||
|
GGML_ASSERT((qs.n_attention_wq == n_attn_layer) && "n_attention_wq is unexpected");
|
||||||
|
GGML_ASSERT((qs.n_attention_wo == n_attn_layer) && "n_attention_wo is unexpected");
|
||||||
}
|
}
|
||||||
|
|
||||||
size_t total_size_org = 0;
|
size_t total_size_org = 0;
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue