Skip to content

Commit 66017f5

Browse files
authored
Bug fixes (#173)
This PR fixes two small bugs. - Double counting of prompt tokens in the calculation of total tokens_per_second - The warmup percent was not working when using max duration (worked for max requests). Traced this to a - which should have been a + in benchmark/aggregator.py
1 parent 6fff1bf commit 66017f5

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

src/guidellm/benchmark/aggregator.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -403,7 +403,7 @@ def add_result(
403403
in_warmup_duration = (
404404
self.args.warmup_duration
405405
and result.request_info.worker_start
406-
<= (global_start_time - self.args.warmup_duration)
406+
<= (global_start_time + self.args.warmup_duration)
407407
)
408408

409409
if in_warmup_number or in_warmup_duration:

src/guidellm/benchmark/benchmark.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -816,7 +816,7 @@ def from_stats(
816816
for req in total_with_output_first
817817
],
818818
iter_counts=[
819-
req.prompt_tokens + req.output_tokens
819+
req.output_tokens
820820
for req in total_with_output_first
821821
],
822822
first_iter_counts=[

0 commit comments

Comments
 (0)