-
Notifications
You must be signed in to change notification settings - Fork 15
Setup vLLM benchmark CI for H100 #32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
jobs: | ||
benchmark-h100: | ||
name: Run vLLM benchmarks | ||
runs-on: linux.aws.h100.4 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
for my own knowledge, is this mean instance with 4 h100?
how many of those we have now?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We have 4 of them atm. Also, FYI, there is one 8xH100 runner too.
The new workflow can be run periodically every 2 hours or on demand by setting the commit from vLLM main branch to benchmark. It works as follows:
public.ecr.aws/q9t5s3a7/vllm-ci-postmerge-repo:<SHA>
Some more PRs are coming after this:
Testing
The results are showing up on the dashboard now https://hud.pytorch.org/benchmark/llms?startTime=Fri%2C%2023%20May%202025%2019%3A19%3A35%20GMT&stopTime=Fri%2C%2030%20May%202025%2019%3A19%3A35%20GMT&granularity=day&lBranch=main&lCommit=7f21e8052b5f3948c8a59514a8dc1e9c5eef70d6&rBranch=main&rCommit=7f21e8052b5f3948c8a59514a8dc1e9c5eef70d6&repoName=vllm-project%2Fvllm&benchmarkName=&modelName=All%20Models&backendName=All%20Backends&modeName=All%20Modes&dtypeName=All%20DType&deviceName=All%20Devices&archName=All%20Platforms