Skip to content

Commit e627f0a

Browse files
committed
reformat
Signed-off-by: Leo Tian <leo.tian@centml.ai>
1 parent a064ccb commit e627f0a

File tree

1 file changed

+3
-5
lines changed

1 file changed

+3
-5
lines changed

vllm/v1/worker/gpu_model_runner.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -40,9 +40,8 @@
4040
from vllm.sampling_params import SamplingType
4141
from vllm.sequence import IntermediateTensors
4242
from vllm.utils import (STR_DTYPE_TO_TORCH_DTYPE, DeviceMemoryProfiler,
43-
GiB_bytes, LazyLoader, async_tensor_h2d, cdiv,
44-
check_use_alibi, get_dtype_size,
45-
is_pin_memory_available)
43+
GiB_bytes, LazyLoader, cdiv, check_use_alibi,
44+
get_dtype_size, is_pin_memory_available)
4645
from vllm.v1.attention.backends.mamba_attn import Mamba2AttentionBackend
4746
from vllm.v1.attention.backends.utils import (AttentionMetadataBuilder,
4847
CommonAttentionMetadata)
@@ -1694,7 +1693,7 @@ def execute_model(
16941693
finished_recving=finished_recving,
16951694
num_nans_in_logits=num_nans_in_logits,
16961695
)
1697-
1696+
16981697
def get_valid_sampled_token_ids(
16991698
self, max_gen_len: int, sampled_token_ids: torch.Tensor,
17001699
discard_sampled_tokens_req_indices: np.ndarray) -> list[list[int]]:
@@ -1715,7 +1714,6 @@ def get_valid_sampled_token_ids(
17151714

17161715
return valid_sampled_token_ids
17171716

1718-
17191717
def kv_connector_no_forward(
17201718
self, scheduler_output: "SchedulerOutput") -> ModelRunnerOutput:
17211719
# KV send/recv even if no work to do.

0 commit comments

Comments
 (0)