Discover and explore top open-source AI tools and projects—updated daily.
Benchmark any foundation model on AWS generative AI services
Top 99.6% on SourcePulse
This tool addresses the need to benchmark foundation models (FMs) across various AWS generative AI services, enabling users to determine optimal price-performance and model accuracy for their workloads. It is targeted at engineers and researchers who need to evaluate and select FMs for deployment on AWS platforms like SageMaker, Bedrock, EKS, or EC2. The primary benefit is providing a standardized and flexible method for performance and accuracy testing, simplifying the decision-making process for generative AI deployments.
How It Works
FMBench operates as a Python package that can be run on any AWS platform with Python support. It utilizes configuration files to define the FM, deployment strategy (including instance types and inference containers like DeepSpeed, TensorRT, and HuggingFace TGI), and benchmarking tests. The tool supports benchmarking against models deployed directly via FMBench or through a "Bring your own endpoint" mode. It measures both performance (latency, transactions per minute) and model accuracy using a panel of LLM evaluators.
Quick Start & Requirements
uv pip install -U fmbench
.Highlighted Details
g5
, p4d
, p5
, Inf2
), inference containers, and parameters like tensor parallelism and rolling batch.fmbench-orchestrator
for automating benchmarking across multiple EC2 instances.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
5 months ago
Inactive