Skip to content

Commit 85eb470

Browse files
committed
fix: typo
Signed-off-by: Jade Zheng <zheng.shoujian@outlook.com>
1 parent cd57ee5 commit 85eb470

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

examples/disaggregated-prefill-v1/offline_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -125,7 +125,7 @@ def run_decode(
125125
max_model_len=40,
126126
)
127127

128-
# Wait for the producer to start the comsumer
128+
# Wait for the producer to start the consumer
129129
print("[Decode] Waiting for prefill node to finish...")
130130
prefill_done.wait()
131131

vllm_ascend/distributed/llmdatadist_connector_v1.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -669,7 +669,7 @@ def wait_for_save(self):
669669
indices = torch.tensor([0], dtype=torch.int64, device="npu")
670670

671671
prefill_info_input = {}
672-
# kv cache should be transfered by request
672+
# kv cache should be transferred by request
673673
for _, request in enumerate(metadata.requests):
674674
if not request.is_store:
675675
continue
@@ -820,7 +820,7 @@ def get_num_new_matched_tokens(
820820
# NOTE: only request in waiting queue will come here. we use datadist
821821
# pull cache to do transfer, so we don't align to block_size in prefill,
822822
# we won't have extra new matched tokens; in decode, new request kv
823-
# cache will be transfered from prefill, so num_computed_tokens = 0, and
823+
# cache will be transferred from prefill, so num_computed_tokens = 0, and
824824
# extra new matched tokens should be len(request.prompt_token_ids) - 1
825825
if self.kv_role == llm_datadist.LLMRole.PROMPT:
826826
return 0

0 commit comments

Comments
 (0)