llama: Support MiniCPM-1B (with & w/o longrope) (#10559)
This commit is contained in:
parent
2759916d86
commit
8d0cfd554a
4 changed files with 61 additions and 183 deletions
|
@ -896,6 +896,8 @@ MODEL_TENSORS: dict[MODEL_ARCH, list[MODEL_TENSOR]] = {
|
|||
MODEL_TENSOR.OUTPUT,
|
||||
MODEL_TENSOR.OUTPUT_NORM,
|
||||
MODEL_TENSOR.ROPE_FREQS,
|
||||
MODEL_TENSOR.ROPE_FACTORS_LONG,
|
||||
MODEL_TENSOR.ROPE_FACTORS_SHORT,
|
||||
MODEL_TENSOR.ATTN_NORM,
|
||||
MODEL_TENSOR.ATTN_Q,
|
||||
MODEL_TENSOR.ATTN_K,
|
||||
|
@ -1388,9 +1390,10 @@ class TokenType(IntEnum):
|
|||
|
||||
|
||||
class RopeScalingType(Enum):
|
||||
NONE = 'none'
|
||||
LINEAR = 'linear'
|
||||
YARN = 'yarn'
|
||||
NONE = 'none'
|
||||
LINEAR = 'linear'
|
||||
YARN = 'yarn'
|
||||
LONGROPE = 'longrope'
|
||||
|
||||
|
||||
class PoolingType(IntEnum):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue