Skip to content

Commit 5268d6a

Browse files
minor cleanups
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
1 parent c99f908 commit 5268d6a

File tree

2 files changed

+2
-5
lines changed

2 files changed

+2
-5
lines changed

tests/v1/attention/test_attention_backends.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ def create_and_prepopulate_kv_cache(
9595
device: torch.device,
9696
num_blocks: int,
9797
common_attn_metadata: CommonAttentionMetadata,
98-
randomize_blocks: bool = True) -> tuple[torch.Tensor, torch.Tensor]:
98+
randomize_blocks: bool = True) -> torch.Tensor:
9999
"""Create and prepopulate a KV cache with context data.
100100
101101
Args:

vllm/v1/attention/backends/flash_attn.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
33
"""Attention layer with FlashAttention."""
44
from dataclasses import dataclass
5-
from typing import TYPE_CHECKING, Any, ClassVar, Optional
5+
from typing import Any, ClassVar, Optional
66

77
import numpy as np
88
import torch
@@ -30,9 +30,6 @@
3030
make_local_attention_virtual_batches)
3131
from vllm.v1.kv_cache_interface import AttentionSpec
3232

33-
if TYPE_CHECKING:
34-
pass
35-
3633
logger = init_logger(__name__)
3734

3835
# NOTE(woosuk): This is an arbitrary number. Tune it if needed.

0 commit comments

Comments
 (0)