Skip to content

Commit a90b784

Browse files
committed
fix etp rank related accuracy problem
Signed-off-by: whx-sjtu <2952154980@qq.com>
1 parent 68fb634 commit a90b784

File tree

1 file changed

+6
-16
lines changed

1 file changed

+6
-16
lines changed

vllm_ascend/ops/fused_moe.py

Lines changed: 6 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -586,22 +586,12 @@ def __init__(self,
586586
self.expert_map = None
587587
self.activation = activation
588588

589-
if self.ep_size > 1:
590-
# Create a tensor of size num_experts filled with -1
591-
self.local_num_experts, self.expert_map = determine_expert_map(
592-
self.ep_size,
593-
get_ep_group().rank_in_group, self.global_num_experts)
594-
self.tp_rank = get_etp_group().rank_in_group
595-
self.ep_rank = get_ep_group().rank_in_group
596-
else:
597-
# Adjust TP size for DP attention
598-
# haven't test its functionality yet, may remove in the future
599-
self.tp_rank = self.tp_size * self.dp_rank
600-
self.ep_rank = 0
601-
self.tp_size = self.tp_size * self.dp_size
602-
self.ep_size = 1
603-
self.local_num_experts = self.global_num_experts
604-
self.expert_map = None
589+
# Create a tensor of size num_experts filled with -1
590+
self.local_num_experts, self.expert_map = determine_expert_map(
591+
self.ep_size,
592+
get_ep_group().rank_in_group, self.global_num_experts)
593+
self.tp_rank = get_etp_group().rank_in_group
594+
self.ep_rank = get_ep_group().rank_in_group
605595

606596
if self.scoring_func != "softmax" and not self.use_grouped_topk:
607597
raise ValueError("Only softmax scoring function is supported for "

0 commit comments

Comments
 (0)