@@ -9,7 +9,6 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
9
9
{ LLM_ARCH_LLAMA4, " llama4" },
10
10
{ LLM_ARCH_DECI, " deci" },
11
11
{ LLM_ARCH_FALCON, " falcon" },
12
- { LLM_ARCH_FALCON_H1, " falcon-h1" },
13
12
{ LLM_ARCH_GROK, " grok" },
14
13
{ LLM_ARCH_GPT2, " gpt2" },
15
14
{ LLM_ARCH_GPTJ, " gptj" },
@@ -48,6 +47,7 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
48
47
{ LLM_ARCH_MAMBA, " mamba" },
49
48
{ LLM_ARCH_MAMBA2, " mamba2" },
50
49
{ LLM_ARCH_JAMBA, " jamba" },
50
+ { LLM_ARCH_FALCON_H1, " falcon-h1" },
51
51
{ LLM_ARCH_BAMBA, " bamba" },
52
52
{ LLM_ARCH_XVERSE, " xverse" },
53
53
{ LLM_ARCH_COMMAND_R, " command-r" },
@@ -364,30 +364,6 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
364
364
{ LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
365
365
},
366
366
},
367
- {
368
- LLM_ARCH_FALCON_H1,
369
- {
370
- { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
371
- { LLM_TENSOR_OUTPUT, " output" },
372
- { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
373
- { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
374
- { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
375
- { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
376
- { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
377
- { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
378
- { LLM_TENSOR_SSM_IN, " blk.%d.ssm_in" },
379
- { LLM_TENSOR_SSM_CONV1D, " blk.%d.ssm_conv1d" },
380
- { LLM_TENSOR_SSM_DT, " blk.%d.ssm_dt" },
381
- { LLM_TENSOR_SSM_A, " blk.%d.ssm_a" },
382
- { LLM_TENSOR_SSM_D, " blk.%d.ssm_d" },
383
- { LLM_TENSOR_SSM_NORM, " blk.%d.ssm_norm" },
384
- { LLM_TENSOR_SSM_OUT, " blk.%d.ssm_out" },
385
- { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
386
- { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
387
- { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
388
- { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
389
- },
390
- },
391
367
{
392
368
LLM_ARCH_GROK,
393
369
{
@@ -1083,6 +1059,30 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
1083
1059
{ LLM_TENSOR_FFN_UP_EXPS, " blk.%d.ffn_up_exps" },
1084
1060
},
1085
1061
},
1062
+ {
1063
+ LLM_ARCH_FALCON_H1,
1064
+ {
1065
+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1066
+ { LLM_TENSOR_OUTPUT, " output" },
1067
+ { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1068
+ { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1069
+ { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1070
+ { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1071
+ { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1072
+ { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1073
+ { LLM_TENSOR_SSM_IN, " blk.%d.ssm_in" },
1074
+ { LLM_TENSOR_SSM_CONV1D, " blk.%d.ssm_conv1d" },
1075
+ { LLM_TENSOR_SSM_DT, " blk.%d.ssm_dt" },
1076
+ { LLM_TENSOR_SSM_A, " blk.%d.ssm_a" },
1077
+ { LLM_TENSOR_SSM_D, " blk.%d.ssm_d" },
1078
+ { LLM_TENSOR_SSM_NORM, " blk.%d.ssm_norm" },
1079
+ { LLM_TENSOR_SSM_OUT, " blk.%d.ssm_out" },
1080
+ { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1081
+ { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1082
+ { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1083
+ { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
1084
+ },
1085
+ },
1086
1086
{
1087
1087
LLM_ARCH_BAMBA,
1088
1088
{
@@ -2099,8 +2099,8 @@ bool llm_arch_is_recurrent(const llm_arch & arch) {
2099
2099
2100
2100
bool llm_arch_is_hybrid (const llm_arch & arch) {
2101
2101
switch (arch) {
2102
- case LLM_ARCH_FALCON_H1:
2103
2102
case LLM_ARCH_JAMBA:
2103
+ case LLM_ARCH_FALCON_H1:
2104
2104
case LLM_ARCH_BAMBA:
2105
2105
case LLM_ARCH_GRANITE_MOE_HYBRID:
2106
2106
return true ;
0 commit comments