We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 91570d8 commit e7c0d2dCopy full SHA for e7c0d2d
vllm_ascend/ops/fused_moe.py
@@ -991,7 +991,7 @@ def apply(
991
global_batch_size=self.global_batch_size,
992
expert_map=expert_map,
993
ep_group=get_ep_group())
994
- elif fused_moe_state == FusedMoEState.All2AllSeq is not None and is_prefill:
+ elif fused_moe_state == FusedMoEState.All2AllSeq and is_prefill:
995
token_dispatcher = kwargs.get('token_dispatcher')
996
return fused_experts_with_all2allv(token_dispatcher=token_dispatcher,
997
probs=topk_weights,
0 commit comments