Skip to content

Commit eb2591c

Browse files
committed
chore: lint code
Signed-off-by: Jade Zheng <zheng.shoujian@outlook.com>
1 parent ebf2242 commit eb2591c

File tree

2 files changed

+15
-14
lines changed

2 files changed

+15
-14
lines changed

examples/disaggregated-prefill-v1/disagg_prefill_proxy_server.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
import os
55

66
import aiohttp
7-
from quart import Quart, make_response, request
7+
from quart import Quart, make_response, request # type: ignore
88

99
AIOHTTP_TIMEOUT = aiohttp.ClientTimeout(total=6 * 60 * 60)
1010

@@ -47,7 +47,7 @@ async def handle_request():
4747
},
4848
):
4949
# Print the prefill result
50-
print(f"===== Prefill result =====")
50+
print("===== Prefill result =====")
5151
print(prefill_result.decode("utf-8"))
5252
print("==========================")
5353
response = json.loads(prefill_result.decode("utf-8"))

vllm_ascend/distributed/llmdatadist_connector_v1.py

Lines changed: 13 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -4,12 +4,12 @@
44
import struct
55
import time
66
from dataclasses import dataclass
7-
from typing import TYPE_CHECKING, Any, List, Tuple
7+
from typing import TYPE_CHECKING, Any, List, Optional, Tuple, Union
88

99
import requests
1010
import torch
1111
import torch_npu
12-
import torchair
12+
import torchair # type: ignore
1313
from vllm.distributed import get_tensor_model_parallel_rank, get_world_group
1414
from vllm.distributed.kv_transfer.kv_connector.v1.base import (
1515
KVConnectorBase_V1, KVConnectorMetadata, KVConnectorRole)
@@ -67,7 +67,8 @@ class ServerInfo:
6767
role: ServerRole
6868
devices: List[DeviceInfo]
6969

70-
def get_device(self, tp_rank: int, dp_rank: int) -> DeviceInfo:
70+
def get_device(self, tp_rank: int,
71+
dp_rank: int) -> Union[DeviceInfo, None]:
7172
for device in self.devices:
7273
if device.tp_rank == tp_rank and device.dp_rank == dp_rank:
7374
return device
@@ -162,7 +163,7 @@ def __init__(self, vllm_config: "VllmConfig") -> None:
162163
GLOBAL_RANKTABLE, self.prefill_tp, self.decode_tp)
163164

164165
def get_device(self, server_id: str, dp_rank: int,
165-
tp_rank: int) -> DeviceInfo:
166+
tp_rank: int) -> Union[DeviceInfo, None]:
166167
for server in self._servers:
167168
if server.server_id != server_id:
168169
continue
@@ -225,7 +226,7 @@ def _get_first_matching_value(self, config_dict: dict,
225226
return default
226227

227228

228-
_CLUSTER_INFO: "ClusterInfo" = None
229+
_CLUSTER_INFO: Optional["ClusterInfo"] = None
229230

230231

231232
def init_cluster_info(vllm_config: "VllmConfig") -> None:
@@ -272,7 +273,7 @@ def __init__(self, role: llm_datadist.LLMRole, local_rank: int,
272273
local_device_info = self.cluster_info.get_device(
273274
local_server_id, dp_rank, tp_rank)
274275
assert local_device_info is not None, \
275-
f"Could not find local device from cluster info."
276+
"Could not find local device from cluster info."
276277

277278
self.cluster_id = local_device_info.cluster_id
278279
self.local_device_ip = local_device_info.device_ip
@@ -379,8 +380,8 @@ def __init__(self, vllm_config: "VllmConfig",
379380
self.kv_role = llm_datadist.LLMRole.DECODER
380381
else:
381382
raise ValueError(
382-
f"The value of kv_role must be either `kv_producer` or `kv_consumer`, but received {kv_transfer_config.kv_role}."
383-
)
383+
"The value of kv_role must be either `kv_producer` or"
384+
f" `kv_consumer`, but received {kv_transfer_config.kv_role}.")
384385

385386
# Used by scheduler process
386387
self._requests_need_load: dict[str, Request] = {}
@@ -400,7 +401,7 @@ def __init__(self, vllm_config: "VllmConfig",
400401
"local_server_id", None)
401402
assert (
402403
self.local_server_id is not None
403-
), f"Cannot find `local_server_id` from `kv_transfer_config.kv_connector_extra_config`."
404+
), "Cannot find `local_server_id` from `kv_transfer_config.kv_connector_extra_config`."
404405

405406
self.dp_rank = self._vllm_config.parallel_config.data_parallel_rank
406407
self.tp_size = self._vllm_config.parallel_config.tensor_parallel_size
@@ -474,7 +475,7 @@ def start_load_kv(self, forward_context: "ForwardContext",
474475
# this is the cause.
475476
if prefill_infos is None:
476477
logger.error(
477-
f"[rank%d][D]: Failed to get prefill info, redo model forwarding.",
478+
"[rank%d][D]: Failed to get prefill info, redo model forwarding.",
478479
torch.distributed.get_rank())
479480
return None
480481

@@ -853,8 +854,8 @@ def _create_cache_tensors(self,
853854
except LLMException as e:
854855
if e.status_code == LLMStatusCode.LLM_DEVICE_OUT_OF_MEMORY:
855856
logger.warning(
856-
f"allocate_cache failed due to insufficient space in the mbuf memory."
857-
)
857+
"allocate_cache failed due to insufficient space in the"
858+
" mbuf memory.")
858859
time.sleep(0.03) # wait for cache buf to be ready
859860
else:
860861
raise e

0 commit comments

Comments
 (0)