Skip to content

Commit 2208ee4

Browse files
committed
chore: lint code
Signed-off-by: Jade Zheng <zheng.shoujian@outlook.com>
1 parent ebf2242 commit 2208ee4

File tree

2 files changed

+14
-14
lines changed

2 files changed

+14
-14
lines changed

examples/disaggregated-prefill-v1/disagg_prefill_proxy_server.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
import os
55

66
import aiohttp
7-
from quart import Quart, make_response, request
7+
from quart import Quart, make_response, request # type: ignore
88

99
AIOHTTP_TIMEOUT = aiohttp.ClientTimeout(total=6 * 60 * 60)
1010

@@ -47,7 +47,7 @@ async def handle_request():
4747
},
4848
):
4949
# Print the prefill result
50-
print(f"===== Prefill result =====")
50+
print("===== Prefill result =====")
5151
print(prefill_result.decode("utf-8"))
5252
print("==========================")
5353
response = json.loads(prefill_result.decode("utf-8"))

vllm_ascend/distributed/llmdatadist_connector_v1.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -4,12 +4,12 @@
44
import struct
55
import time
66
from dataclasses import dataclass
7-
from typing import TYPE_CHECKING, Any, List, Tuple
7+
from typing import TYPE_CHECKING, Any, List, Optional, Tuple, Union
88

99
import requests
1010
import torch
1111
import torch_npu
12-
import torchair
12+
import torchair # type: ignore
1313
from vllm.distributed import get_tensor_model_parallel_rank, get_world_group
1414
from vllm.distributed.kv_transfer.kv_connector.v1.base import (
1515
KVConnectorBase_V1, KVConnectorMetadata, KVConnectorRole)
@@ -67,7 +67,7 @@ class ServerInfo:
6767
role: ServerRole
6868
devices: List[DeviceInfo]
6969

70-
def get_device(self, tp_rank: int, dp_rank: int) -> DeviceInfo:
70+
def get_device(self, tp_rank: int, dp_rank: int) -> Union[DeviceInfo, None]:
7171
for device in self.devices:
7272
if device.tp_rank == tp_rank and device.dp_rank == dp_rank:
7373
return device
@@ -162,7 +162,7 @@ def __init__(self, vllm_config: "VllmConfig") -> None:
162162
GLOBAL_RANKTABLE, self.prefill_tp, self.decode_tp)
163163

164164
def get_device(self, server_id: str, dp_rank: int,
165-
tp_rank: int) -> DeviceInfo:
165+
tp_rank: int) -> Union[DeviceInfo , None]:
166166
for server in self._servers:
167167
if server.server_id != server_id:
168168
continue
@@ -225,7 +225,7 @@ def _get_first_matching_value(self, config_dict: dict,
225225
return default
226226

227227

228-
_CLUSTER_INFO: "ClusterInfo" = None
228+
_CLUSTER_INFO: Optional["ClusterInfo"] = None
229229

230230

231231
def init_cluster_info(vllm_config: "VllmConfig") -> None:
@@ -272,7 +272,7 @@ def __init__(self, role: llm_datadist.LLMRole, local_rank: int,
272272
local_device_info = self.cluster_info.get_device(
273273
local_server_id, dp_rank, tp_rank)
274274
assert local_device_info is not None, \
275-
f"Could not find local device from cluster info."
275+
"Could not find local device from cluster info."
276276

277277
self.cluster_id = local_device_info.cluster_id
278278
self.local_device_ip = local_device_info.device_ip
@@ -379,8 +379,8 @@ def __init__(self, vllm_config: "VllmConfig",
379379
self.kv_role = llm_datadist.LLMRole.DECODER
380380
else:
381381
raise ValueError(
382-
f"The value of kv_role must be either `kv_producer` or `kv_consumer`, but received {kv_transfer_config.kv_role}."
383-
)
382+
"The value of kv_role must be either `kv_producer` or"
383+
f" `kv_consumer`, but received {kv_transfer_config.kv_role}.")
384384

385385
# Used by scheduler process
386386
self._requests_need_load: dict[str, Request] = {}
@@ -400,7 +400,7 @@ def __init__(self, vllm_config: "VllmConfig",
400400
"local_server_id", None)
401401
assert (
402402
self.local_server_id is not None
403-
), f"Cannot find `local_server_id` from `kv_transfer_config.kv_connector_extra_config`."
403+
), "Cannot find `local_server_id` from `kv_transfer_config.kv_connector_extra_config`."
404404

405405
self.dp_rank = self._vllm_config.parallel_config.data_parallel_rank
406406
self.tp_size = self._vllm_config.parallel_config.tensor_parallel_size
@@ -474,7 +474,7 @@ def start_load_kv(self, forward_context: "ForwardContext",
474474
# this is the cause.
475475
if prefill_infos is None:
476476
logger.error(
477-
f"[rank%d][D]: Failed to get prefill info, redo model forwarding.",
477+
"[rank%d][D]: Failed to get prefill info, redo model forwarding.",
478478
torch.distributed.get_rank())
479479
return None
480480

@@ -853,8 +853,8 @@ def _create_cache_tensors(self,
853853
except LLMException as e:
854854
if e.status_code == LLMStatusCode.LLM_DEVICE_OUT_OF_MEMORY:
855855
logger.warning(
856-
f"allocate_cache failed due to insufficient space in the mbuf memory."
857-
)
856+
"allocate_cache failed due to insufficient space in the"
857+
" mbuf memory.")
858858
time.sleep(0.03) # wait for cache buf to be ready
859859
else:
860860
raise e

0 commit comments

Comments
 (0)