Skip to content

Commit 44cda75

Browse files
committed
fix: Revert order changes for Falcon H1 to stay consistent with upstream
Branch: GraniteFour Signed-off-by: Gabe Goodhart <ghart@us.ibm.com>
1 parent fa159cf commit 44cda75

File tree

3 files changed

+28
-28
lines changed

3 files changed

+28
-28
lines changed

gguf-py/gguf/constants.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -603,7 +603,6 @@ class MODEL_TENSOR(IntEnum):
603603
MODEL_ARCH.LLAMA4: "llama4",
604604
MODEL_ARCH.DECI: "deci",
605605
MODEL_ARCH.FALCON: "falcon",
606-
MODEL_ARCH.FALCON_H1: "falcon-h1",
607606
MODEL_ARCH.BAICHUAN: "baichuan",
608607
MODEL_ARCH.GROK: "grok",
609608
MODEL_ARCH.GPT2: "gpt2",
@@ -676,6 +675,7 @@ class MODEL_TENSOR(IntEnum):
676675
MODEL_ARCH.DOTS1: "dots1",
677676
MODEL_ARCH.ARCEE: "arcee",
678677
MODEL_ARCH.ERNIE4_5: "ernie4_5",
678+
MODEL_ARCH.FALCON_H1: "falcon-h1",
679679
MODEL_ARCH.HUNYUAN_MOE: "hunyuan-moe",
680680
MODEL_ARCH.SMOLLM3: "smollm3",
681681
}

src/llama-arch.cpp

Lines changed: 26 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,6 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
99
{ LLM_ARCH_LLAMA4, "llama4" },
1010
{ LLM_ARCH_DECI, "deci" },
1111
{ LLM_ARCH_FALCON, "falcon" },
12-
{ LLM_ARCH_FALCON_H1, "falcon-h1" },
1312
{ LLM_ARCH_GROK, "grok" },
1413
{ LLM_ARCH_GPT2, "gpt2" },
1514
{ LLM_ARCH_GPTJ, "gptj" },
@@ -48,6 +47,7 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
4847
{ LLM_ARCH_MAMBA, "mamba" },
4948
{ LLM_ARCH_MAMBA2, "mamba2" },
5049
{ LLM_ARCH_JAMBA, "jamba" },
50+
{ LLM_ARCH_FALCON_H1, "falcon-h1" },
5151
{ LLM_ARCH_BAMBA, "bamba" },
5252
{ LLM_ARCH_XVERSE, "xverse" },
5353
{ LLM_ARCH_COMMAND_R, "command-r" },
@@ -364,30 +364,6 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
364364
{ LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" },
365365
},
366366
},
367-
{
368-
LLM_ARCH_FALCON_H1,
369-
{
370-
{ LLM_TENSOR_TOKEN_EMBD, "token_embd" },
371-
{ LLM_TENSOR_OUTPUT, "output" },
372-
{ LLM_TENSOR_OUTPUT_NORM, "output_norm" },
373-
{ LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" },
374-
{ LLM_TENSOR_ATTN_Q, "blk.%d.attn_q" },
375-
{ LLM_TENSOR_ATTN_K, "blk.%d.attn_k" },
376-
{ LLM_TENSOR_ATTN_V, "blk.%d.attn_v" },
377-
{ LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" },
378-
{ LLM_TENSOR_SSM_IN, "blk.%d.ssm_in" },
379-
{ LLM_TENSOR_SSM_CONV1D, "blk.%d.ssm_conv1d" },
380-
{ LLM_TENSOR_SSM_DT, "blk.%d.ssm_dt" },
381-
{ LLM_TENSOR_SSM_A, "blk.%d.ssm_a" },
382-
{ LLM_TENSOR_SSM_D, "blk.%d.ssm_d" },
383-
{ LLM_TENSOR_SSM_NORM, "blk.%d.ssm_norm" },
384-
{ LLM_TENSOR_SSM_OUT, "blk.%d.ssm_out" },
385-
{ LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" },
386-
{ LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" },
387-
{ LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" },
388-
{ LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" },
389-
},
390-
},
391367
{
392368
LLM_ARCH_GROK,
393369
{
@@ -1083,6 +1059,30 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
10831059
{ LLM_TENSOR_FFN_UP_EXPS, "blk.%d.ffn_up_exps" },
10841060
},
10851061
},
1062+
{
1063+
LLM_ARCH_FALCON_H1,
1064+
{
1065+
{ LLM_TENSOR_TOKEN_EMBD, "token_embd" },
1066+
{ LLM_TENSOR_OUTPUT, "output" },
1067+
{ LLM_TENSOR_OUTPUT_NORM, "output_norm" },
1068+
{ LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" },
1069+
{ LLM_TENSOR_ATTN_Q, "blk.%d.attn_q" },
1070+
{ LLM_TENSOR_ATTN_K, "blk.%d.attn_k" },
1071+
{ LLM_TENSOR_ATTN_V, "blk.%d.attn_v" },
1072+
{ LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" },
1073+
{ LLM_TENSOR_SSM_IN, "blk.%d.ssm_in" },
1074+
{ LLM_TENSOR_SSM_CONV1D, "blk.%d.ssm_conv1d" },
1075+
{ LLM_TENSOR_SSM_DT, "blk.%d.ssm_dt" },
1076+
{ LLM_TENSOR_SSM_A, "blk.%d.ssm_a" },
1077+
{ LLM_TENSOR_SSM_D, "blk.%d.ssm_d" },
1078+
{ LLM_TENSOR_SSM_NORM, "blk.%d.ssm_norm" },
1079+
{ LLM_TENSOR_SSM_OUT, "blk.%d.ssm_out" },
1080+
{ LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" },
1081+
{ LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" },
1082+
{ LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" },
1083+
{ LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" },
1084+
},
1085+
},
10861086
{
10871087
LLM_ARCH_BAMBA,
10881088
{
@@ -2099,8 +2099,8 @@ bool llm_arch_is_recurrent(const llm_arch & arch) {
20992099

21002100
bool llm_arch_is_hybrid(const llm_arch & arch) {
21012101
switch (arch) {
2102-
case LLM_ARCH_FALCON_H1:
21032102
case LLM_ARCH_JAMBA:
2103+
case LLM_ARCH_FALCON_H1:
21042104
case LLM_ARCH_BAMBA:
21052105
case LLM_ARCH_GRANITE_MOE_HYBRID:
21062106
return true;

src/llama-arch.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,6 @@ enum llm_arch {
1313
LLM_ARCH_LLAMA4,
1414
LLM_ARCH_DECI,
1515
LLM_ARCH_FALCON,
16-
LLM_ARCH_FALCON_H1,
1716
LLM_ARCH_BAICHUAN,
1817
LLM_ARCH_GROK,
1918
LLM_ARCH_GPT2,
@@ -52,6 +51,7 @@ enum llm_arch {
5251
LLM_ARCH_MAMBA,
5352
LLM_ARCH_MAMBA2,
5453
LLM_ARCH_JAMBA,
54+
LLM_ARCH_FALCON_H1,
5555
LLM_ARCH_BAMBA,
5656
LLM_ARCH_XVERSE,
5757
LLM_ARCH_COMMAND_R,

0 commit comments

Comments
 (0)