Skip to content

Commit aca33c5

Browse files
committed
fix: n>1 logprobs streaming edge case
Signed-off-by: Travis Johnson <tsjohnso@us.ibm.com>
1 parent 8359f4c commit aca33c5

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

vllm/entrypoints/openai/serving_chat.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -573,7 +573,8 @@ async def chat_completion_stream_generator(
573573
if finish_reason_sent[i]:
574574
continue
575575

576-
if request.logprobs and request.top_logprobs is not None:
576+
if request.logprobs and request.top_logprobs is not None \
577+
and output.token_ids:
577578
assert output.logprobs is not None, (
578579
"Did not output logprobs")
579580
logprobs = self._create_chat_logprobs(

vllm/entrypoints/openai/serving_completion.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -358,7 +358,7 @@ async def completion_stream_generator(
358358
# Chunked prefill case, don't return empty chunks
359359
continue
360360

361-
if request.logprobs is not None:
361+
if request.logprobs is not None and output.token_ids:
362362
assert out_logprobs is not None, (
363363
"Did not output logprobs")
364364
logprobs = self._create_completion_logprobs(

0 commit comments

Comments
 (0)