Skip to content

Commit d967a87

Browse files
committed
chore: remove v0.8.4 patch
Signed-off-by: Jade Zheng <zheng.shoujian@outlook.com>
1 parent f26c0e0 commit d967a87

File tree

1 file changed

+2
-11
lines changed

1 file changed

+2
-11
lines changed

vllm_ascend/worker/model_runner_v1.py

Lines changed: 2 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,8 @@
3232
from vllm.attention import AttentionType, get_attn_backend
3333
from vllm.attention.layer import Attention
3434
from vllm.config import CompilationLevel, VllmConfig
35+
from vllm.distributed.kv_transfer import (get_kv_transfer_group,
36+
has_kv_transfer_group)
3537
from vllm.distributed.parallel_state import get_pp_group
3638
from vllm.forward_context import set_forward_context
3739
from vllm.inputs import INPUT_REGISTRY
@@ -55,17 +57,6 @@
5557
from vllm_ascend.attention.attention_v1 import AscendAttentionState
5658
from vllm_ascend.platform import NPUPlatform
5759

58-
if vllm_version_is("0.8.4"):
59-
from vllm.distributed import get_kv_transfer_group
60-
61-
def has_kv_transfer_group() -> bool:
62-
# vLLM 0.8.4 does not support disaggregated prefill. This function is
63-
# added to ensure compatibility with vLLM 0.8.4.
64-
return False
65-
else:
66-
from vllm.distributed.kv_transfer import ( # type: ignore
67-
get_kv_transfer_group, has_kv_transfer_group)
68-
6960
if TYPE_CHECKING:
7061
import xgrammar as xgr # type: ignore[import-untyped]
7162
from vllm.v1.core.sched.output import SchedulerOutput

0 commit comments

Comments
 (0)