Discover and explore top open-source AI tools and projects—updated daily.
sgl-projectLLM serving performance benchmarking
Top 99.8% on SourcePulse
Genai-bench provides a unified, accurate, and user-friendly solution for comprehensive token-level performance evaluation of Large Language Model (LLM) serving systems. It targets engineers and researchers needing to deeply understand and optimize LLM deployment performance. The tool delivers detailed insights into metrics such as throughput, latency (TTFT, E2E, TPOT), error rates, and requests per second (RPS) across diverse traffic scenarios and concurrency levels, facilitating informed infrastructure decisions.
How It Works
The project employs a dual-interface approach: a robust Command Line Interface (CLI) for initiating and managing benchmarks, and an interactive Live UI Dashboard for real-time progress monitoring and metric visualization. It focuses on granular, token-level analysis of LLM serving systems. Post-benchmark, an Experiment Analyzer automatically generates detailed Excel reports containing pricing and raw metrics, alongside flexible plot configurations for visualizing performance trends.
Quick Start & Requirements
pip install genai-bench.pyproject.toml).Highlighted Details
Maintenance & Community
The README does not specify maintainers, community channels (e.g., Discord, Slack), sponsorships, or a public roadmap. Contribution guidelines are provided within the repository.
Licensing & Compatibility
The project is released under the MIT license, which is highly permissive and generally suitable for commercial use and integration into closed-source projects.
Limitations & Caveats
The provided documentation does not explicitly detail limitations, unsupported platforms, alpha status, or specific caveats regarding the benchmarking process or supported model backends beyond the general your-backend placeholder in usage examples.
3 days ago
Inactive
ray-project
JinjieNi
InferenceMAX
benchmark-action
ray-project