Skip to content

Yeok-c/vllm-top

Repository files navigation

vllm-top

PyPI version

Demo

vllm-top is a Python package for monitoring and displaying metrics from the vLLM service. It provides a comprehensive dashboard to visualize both current state and historical performance, making it easy to track and analyze service behavior over time.


🚀 Features

  • Task State Visibility: Instantly see GPU Cache Usage, Running and Waiting requests to help debug bottlenecks and improve throughput.
  • Minimalist Monitoring: Lightweight dashboard that parses metrics directly from Prometheus.
  • Quick Setup: No extra configuration — just pip install and run.

📦 Installation

Install via pip:

pip install vllm-top

🛠️ Usage

Start monitoring:

vllm-top

Change update interval (in seconds):

vllm-top --interval 5

Get a one-time snapshot:

vllm-top --snapshot

🤝 Contributing

Contributions are welcome! Please submit a pull request or open an issue for enhancements or bug fixes.


📄 License

Licensed under the MIT License. See the LICENSE file for details.


📜 Changelog

See CHANGELOG.md for a detailed

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages