Skip to content

Commit f69fcb5

Browse files
author
Dingli
committed
bugfix
Signed-off-by: Dingli <dingli40@huawei.com>
1 parent 78a00c3 commit f69fcb5

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

vllm_ascend/models/deepseek_v2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -228,7 +228,7 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
228228

229229
cv_parallel = VLLM_ENABLE_CV_PARALLEL and not is_prefill
230230

231-
if self.n_shared_experts is not None
231+
if self.n_shared_experts is not None:
232232
if not cv_parallel:
233233
shared_output = self.shared_experts(hidden_states)
234234
else:

vllm_ascend/quantization/w8a8_dynamic.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -127,7 +127,7 @@ def apply_mlp(hidden_states_wrapper: List[torch.Tensor],
127127
output_dtype=torch.bfloat16,
128128
)
129129
if shared_experts.down_proj.reduce_results and shared_experts.down_proj.tp_size > 1:
130-
shared_output = tensor_model_parallel_all_reduce(x)
130+
shared_output = tensor_model_parallel_all_reduce(shared_output)
131131
if shared_experts:
132132
return hidden_states, shared_output
133133
return hidden_states

0 commit comments

Comments
 (0)