Skip to content

Commit aa51c7c

Browse files
committed
page_size -> num_element_per_page
Signed-off-by: Thomas Parnell <tpa@zurich.ibm.com>
1 parent 6abace4 commit aa51c7c

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/v1/worker/gpu_model_runner.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2444,14 +2444,14 @@ def _reshape_kv_cache_tensors(
24442444
elif isinstance(kv_cache_spec, MambaSpec):
24452445
raw_tensor = kv_cache_raw_tensors[layer_name]
24462446
dtype = kv_cache_spec.dtype
2447-
page_size = kv_cache_spec.page_size_bytes // get_dtype_size(
2448-
dtype)
2447+
num_element_per_page = (kv_cache_spec.page_size_bytes //
2448+
get_dtype_size(dtype))
24492449
state_tensors = []
24502450
storage_offset = 0
24512451
for shape in kv_cache_spec.shapes:
24522452
target_shape = (num_blocks, *shape)
24532453
stride = torch.empty(target_shape).stride()
2454-
target_stride = (page_size, *stride[1:])
2454+
target_stride = (num_element_per_page, *stride[1:])
24552455
tensor = torch.as_strided(
24562456
raw_tensor.view(dtype),
24572457
size=target_shape,

0 commit comments

Comments
 (0)