Skip to content

Commit d118d63

Browse files
committed
fix codecheck
Signed-off-by: David9857 <985700846@qq.com>
1 parent 9ddb591 commit d118d63

File tree

2 files changed

+16
-18
lines changed

2 files changed

+16
-18
lines changed

vllm_ascend/models/deepseek_v2.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@
3030
import torch
3131
import torch.distributed as dist
3232
import torch_npu
33-
import torchair as tng
33+
import torchair as tng # type: ignore
3434
import vllm.envs as envs
3535
from torch import nn
3636
from transformers import PretrainedConfig
@@ -257,7 +257,8 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
257257
if self.n_shared_experts is not None and cv_parallel:
258258
with tng.scope.npu_stream_switch('cv'):
259259
tng.scope.npu_wait_tensor(shared_hidden_states, router_logits)
260-
x, dynamic_scale = torch_npu.npu_dynamic_quant(shared_hidden_states)
260+
x, dynamic_scale = torch_npu.npu_dynamic_quant(
261+
shared_hidden_states)
261262
gate_up = torch_npu.npu_quant_matmul(
262263
x,
263264
self.shared_experts.gate_up_proj.weight,
@@ -274,8 +275,7 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
274275
enable_force_load_balance=enable_force_load_balance,
275276
shared_experts=self.shared_experts,
276277
shared_gate_up=gate_up,
277-
shared_dynamic_scale=dynamic_scale
278-
)
278+
shared_dynamic_scale=dynamic_scale)
279279
router_hidden_states = router_hidden_states * self.routed_scaling_factor
280280
else:
281281
router_hidden_states = self.experts(

vllm_ascend/quantization/w8a8_dynamic.py

Lines changed: 12 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
import torch
2121
import torch.distributed as dist
2222
import torch_npu
23-
import torchair as tng
23+
import torchair as tng # type: ignore
2424
from vllm.distributed import GroupCoordinator, tensor_model_parallel_all_reduce
2525

2626
import vllm_ascend.envs as envs_ascend
@@ -133,19 +133,17 @@ def apply_mlp(hidden_states_wrapper: List[torch.Tensor],
133133
return hidden_states
134134

135135

136-
def fused_experts_with_mc2(
137-
hidden_states: torch.Tensor,
138-
w1: torch.Tensor,
139-
w2: torch.Tensor,
140-
w1_scale: torch.Tensor,
141-
w2_scale: torch.Tensor,
142-
topk_weights: torch.Tensor,
143-
topk_ids: torch.Tensor,
144-
top_k: int,
145-
expert_map: torch.Tensor = None,
146-
moe_all_to_all_group_name: str = "",
147-
**kwargs
148-
) -> torch.Tensor:
136+
def fused_experts_with_mc2(hidden_states: torch.Tensor,
137+
w1: torch.Tensor,
138+
w2: torch.Tensor,
139+
w1_scale: torch.Tensor,
140+
w2_scale: torch.Tensor,
141+
topk_weights: torch.Tensor,
142+
topk_ids: torch.Tensor,
143+
top_k: int,
144+
expert_map: torch.Tensor = None,
145+
moe_all_to_all_group_name: str = "",
146+
**kwargs) -> torch.Tensor:
149147
global_bs = 0
150148
moe_expert_num = len(expert_map)
151149
# hidden_states = hidden_states.bfloat16()

0 commit comments

Comments
 (0)