Discover and explore top open-source AI tools and projects—updated daily.
leworkBenchmark LLM concurrency and stress test performance
Top 99.8% on SourcePulse
Summary
This project offers an LLM concurrent performance testing tool designed for automated stress testing and detailed performance report generation. It is targeted at engineers and researchers evaluating LLM deployment performance under varying loads, providing insights into throughput, latency, and stability. The tool helps in understanding how an LLM service scales and performs under pressure.
How It Works
The tool employs a multi-stage concurrency testing approach, systematically increasing load from low to high (1-300 concurrent requests) to identify performance bottlenecks. It automates data collection, analysis, and the generation of comprehensive statistical reports, supporting both short and long text scenarios. The core logic in llm_benchmark.py manages request handling, connection pooling, and detailed metric collection, including support for streaming responses.
Quick Start & Requirements
pip install -r requirements.txt. Alternatively, use Docker by building the image (docker build -t llm-benchmark .) or pulling a pre-built one (docker pull samge/llm-benchmark).requirements.txt dependencies. Docker is optional but recommended for ease of use.python run_benchmarks.py --llm_url <URL> --api_key <KEY> --model <MODEL_NAME> [--use_long_context]python llm_benchmark.py --llm_url <URL> --api_key <KEY> --model <MODEL_NAME> --num_requests <N> --concurrency <C>$PWD/output:/app/output).Highlighted Details
Maintenance & Community
The provided README does not contain information regarding maintainers, community channels (like Discord/Slack), or project roadmaps.
Licensing & Compatibility
Limitations & Caveats
The tool requires a running LLM endpoint to test against, specified via --llm_url. Specific model names and API keys may be necessary depending on the LLM service. The README does not detail specific hardware requirements beyond standard Python/Docker environments, nor does it mention known bugs or alpha/beta status.
4 months ago
Inactive
ray-project