model: dbrx: fix missing embedding tensor, mix with output layer
This commit is contained in:
parent
9c7dedb0f3
commit
fe8089871e
1 changed files with 2 additions and 1 deletions
|
@ -24,6 +24,7 @@ class TensorNameMap:
|
||||||
"backbone.embedding", # mamba
|
"backbone.embedding", # mamba
|
||||||
"backbone.embeddings", # mamba-hf
|
"backbone.embeddings", # mamba-hf
|
||||||
"transformer.in_out_embed", # Grok
|
"transformer.in_out_embed", # Grok
|
||||||
|
"transformer.wte.weight", # dbrx
|
||||||
),
|
),
|
||||||
|
|
||||||
# Token type embeddings
|
# Token type embeddings
|
||||||
|
@ -52,7 +53,7 @@ class TensorNameMap:
|
||||||
"output", # llama-pth bloom internlm2
|
"output", # llama-pth bloom internlm2
|
||||||
"word_embeddings_for_head", # persimmon
|
"word_embeddings_for_head", # persimmon
|
||||||
"lm_head.linear", # phi2
|
"lm_head.linear", # phi2
|
||||||
"transformer.wte.weight", # dbrx
|
"lm_head.weight", # dbrx
|
||||||
),
|
),
|
||||||
|
|
||||||
# Output norm
|
# Output norm
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue