Skip to content

Commit e7c0d2d

Browse files
author
weijinqian_v1
committed
[Feature]Moe alltoallv communication optimization for unquantized RL training sence & alltoallv support dpo
Signed-off-by: weijinqian_v1 <weijinqian@huawei.com>
1 parent 91570d8 commit e7c0d2d

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm_ascend/ops/fused_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -991,7 +991,7 @@ def apply(
991991
global_batch_size=self.global_batch_size,
992992
expert_map=expert_map,
993993
ep_group=get_ep_group())
994-
elif fused_moe_state == FusedMoEState.All2AllSeq is not None and is_prefill:
994+
elif fused_moe_state == FusedMoEState.All2AllSeq and is_prefill:
995995
token_dispatcher = kwargs.get('token_dispatcher')
996996
return fused_experts_with_all2allv(token_dispatcher=token_dispatcher,
997997
probs=topk_weights,

0 commit comments

Comments
 (0)