Skip to content

Commit 670e136

Browse files
authored
convert : fix Ernie4.5 MoE without shared experts (ggml-org#14746)
1 parent 760b448 commit 670e136

File tree

1 file changed

+4
-3
lines changed

1 file changed

+4
-3
lines changed

convert_hf_to_gguf.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2910,11 +2910,12 @@ def set_gguf_parameters(self):
29102910
self.gguf_writer.add_expert_used_count(self.hparams["moe_k"])
29112911
self.gguf_writer.add_interleave_moe_layer_step(self.hparams["moe_layer_interval"])
29122912
self.gguf_writer.add_leading_dense_block_count(self.hparams["moe_layer_start_index"])
2913-
self.gguf_writer.add_rope_freq_base(self.hparams["rope_theta"])
29142913
if (moe_intermediate_size := self.hparams.get("moe_intermediate_size")) is not None:
29152914
self.gguf_writer.add_expert_feed_forward_length(moe_intermediate_size)
2916-
if (shared_expert_intermediate_size := self.hparams.get('intermediate_size')) is not None and (num_key_value_heads := self.hparams.get('num_key_value_heads')) is not None:
2917-
self.gguf_writer.add_expert_shared_feed_forward_length(shared_expert_intermediate_size // num_key_value_heads)
2915+
if (shared_expert_count := self.hparams.get('moe_num_shared_experts')) is not None:
2916+
self.gguf_writer.add_expert_shared_count(shared_expert_count)
2917+
if shared_expert_count > 0 and (shared_expert_intermediate_size := self.hparams.get('intermediate_size')) is not None and (num_key_value_heads := self.hparams.get('num_key_value_heads')) is not None:
2918+
self.gguf_writer.add_expert_shared_feed_forward_length(shared_expert_intermediate_size // num_key_value_heads)
29182919

29192920
def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
29202921
# Modify correction bias name as in DeepseekV2

0 commit comments

Comments
 (0)