model: dbrx: fix missing embedding tensor, mix with output layer
This commit is contained in:
parent
9c7dedb0f3
commit
fe8089871e
1 changed files with 2 additions and 1 deletions
|
@ -24,6 +24,7 @@ class TensorNameMap:
|
|||
"backbone.embedding", # mamba
|
||||
"backbone.embeddings", # mamba-hf
|
||||
"transformer.in_out_embed", # Grok
|
||||
"transformer.wte.weight", # dbrx
|
||||
),
|
||||
|
||||
# Token type embeddings
|
||||
|
@ -52,7 +53,7 @@ class TensorNameMap:
|
|||
"output", # llama-pth bloom internlm2
|
||||
"word_embeddings_for_head", # persimmon
|
||||
"lm_head.linear", # phi2
|
||||
"transformer.wte.weight", # dbrx
|
||||
"lm_head.weight", # dbrx
|
||||
),
|
||||
|
||||
# Output norm
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue