Skip to content

Commit 354ff2c

Browse files
committed
update torchair config
Signed-off-by: David9857 <985700846@qq.com> bugfix Signed-off-by: David9857 <985700846@qq.com>
1 parent 051073a commit 354ff2c

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

vllm_ascend/ascend_config.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,6 +53,8 @@ def __init__(self, torchair_graph_config):
5353
"graph_batch_sizes", [])
5454
self.graph_batch_sizes_init = torchair_graph_config.get(
5555
"graph_batch_sizes_init", False)
56+
self.enable_multistream_shared_expert = torchair_graph_config.get(
57+
"enable_multistream_shared_expert", False)
5658

5759
if not isinstance(self.graph_batch_sizes, list):
5860
raise TypeError("graph_batch_sizes must be list[int]")
@@ -105,7 +107,7 @@ def check_ascend_config(vllm_config, enforce_eager):
105107
ascend_config = get_ascend_config()
106108

107109
# Both for V0 and V1 Engine, torchair_graph cannot be enabled with eager mode.
108-
if ascend_config.torchair_graph_config.enabled and not enforce_eager:
110+
if ascend_config.torchair_graph_config.enabled and enforce_eager:
109111
raise RuntimeError(
110112
"Can't enable graph mode and eager mode at the same time. Please set `enforce_eager=False` if you attempt to enable NPU graph mode."
111113
)

vllm_ascend/models/deepseek_v2.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,6 @@
3030
import torch
3131
import torch.distributed as dist
3232
import torch_npu
33-
import torchair as tng # type: ignore
3433
import vllm.envs as envs
3534
from torch import nn
3635
from transformers import PretrainedConfig

0 commit comments

Comments
 (0)