-
Notifications
You must be signed in to change notification settings - Fork 37
Issues: neuralmagic/guidellm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Sweep Mode Improvements - get closer to approaching peak-throughput
load-request
load-request workstream
priority-high
#155
opened May 8, 2025 by
rgreenberg1
Standardize dataset profiles for reasoning benchmarking
priority-high
#154
opened May 8, 2025 by
rgreenberg1
Support Audio to text multi-modal benchmarking in GuideLLM
priority-high
#153
opened May 8, 2025 by
rgreenberg1
Support VL Model Benchmarking
multi-modal
Support for benchmarking new multi-modal models
priority-high
#152
opened May 8, 2025 by
rgreenberg1
Event loop is closed error.
bug
Something isn't working
priority-high
#147
opened May 5, 2025 by
phalani-paladugu
GuideLLM Sweep Benchmark: Low KV Cache Utilization and Limited QPS on H100 GPUs
#136
opened Apr 25, 2025 by
kaushikmitr
Sweep mode with --max-seconds fails ungracefully if calculated RPS is too low
priority-low
#115
opened Apr 15, 2025 by
sjmonson
[Bug] Possibly incorrect output dump ? (The actual data appears to be fine)
good first issue
Good for newcomers
priority-medium
#114
opened Apr 15, 2025 by
thameem-abbas
[FeatureReuqest] Auto-sweep with model saturation detection
enhancement
New feature or request
#108
opened Apr 8, 2025 by
markVaykhansky
[FeatureRequest] Constraint input & output sizes no matter the dataset
#106
opened Apr 8, 2025 by
markVaykhansky
[Feature Request] Testing with defined prefix lengths
priority-high
#104
opened Apr 4, 2025 by
thameem-abbas
Enable Batch Inferencing Benchmarking Support
priority-medium
#102
opened Mar 31, 2025 by
rgreenberg1
Automatic performance stabilizization - so no need to set max-requests or max-seconds
analysis
analysis workstream - relating to adding in features to do analysis on extrapolating results
automation
automation workstream
priority-high
#101
opened Mar 31, 2025 by
rgreenberg1
Land GuideLLM UI in Upstream Repository
priority-high
UI
Front-end workstream
#98
opened Mar 12, 2025 by
rgreenberg1
INFO log level is to verbose
enhancement
New feature or request
good first issue
Good for newcomers
priority-low
#97
opened Mar 11, 2025 by
sjmonson
GuideLLM v0.2.0 - CI/CD Finalization, Documentation Expansion, and Backend Support
Have GuideLLM kick off a vLLM server automatically to avoid having the user install vLLM and assign the target themselves
enhancement
New feature or request
#95
opened Mar 9, 2025 by
rgreenberg1
Integrate lm-eval into GuideLLM
accuracy-eval
accuracy-evaluation workstream
#90
opened Mar 6, 2025 by
rgreenberg1
Enable vLLM scheduler specific Metrics
metrics
Metrics workstream
#86
opened Mar 4, 2025 by
rgreenberg1
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.