Skip to content

Commit 6338bd2

Browse files
remove duplicates
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
1 parent 6f7c68d commit 6338bd2

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/v1/attention/backends/mla/rocm_aiter_mla.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -75,12 +75,12 @@ def __init__(self, kv_cache_spec: AttentionSpec, vllm_config: VllmConfig,
7575
self.compilation_config = vllm_config.compilation_config
7676
max_num_pages_per_req = cdiv(vllm_config.model_config.max_model_len,
7777
self.kv_cache_spec.block_size)
78-
max_num_req = vllm_config.scheduler_config.max_num_seqs
79-
max_num_pages = max_num_req * max_num_pages_per_req
78+
max_num_reqs = vllm_config.scheduler_config.max_num_seqs
79+
max_num_pages = max_num_reqs * max_num_pages_per_req
8080

8181
# Preparing persistent buffers
8282
if vllm_config.compilation_config.full_cuda_graph:
83-
max_num_reqs = vllm_config.scheduler_config.max_num_seqs
83+
8484
self.paged_kv_indptr = torch.zeros(max_num_reqs + 1,
8585
dtype=torch.int32,
8686
device=device)

0 commit comments

Comments
 (0)