Text-to-3D generation benchmark
Top 35.3% on sourcepulse
T3Bench provides a comprehensive benchmark for evaluating text-to-3D generation models. It addresses the need for standardized assessment by offering 300 diverse text prompts across three complexity levels, along with novel automatic metrics for quality and text alignment. This benchmark is designed for researchers and developers in the 3D generation field.
How It Works
T3Bench leverages multi-view images generated from 3D content to assess quality and text alignment. The quality metric combines multi-view text-image scores with regional convolution to detect inconsistencies. The alignment metric uses multi-view captioning and LLM evaluation to measure text-3D consistency, aiming for efficient and reliable evaluation.
Quick Start & Requirements
pip install -r requirements.txt
.run_t3.py
), extracting meshes (run_mesh.py
), quality evaluation (run_eval_quality.py
), and alignment evaluation (run_caption.py
, run_eval_alignment.py
).Highlighted Details
Maintenance & Community
The project acknowledges contributions from open-source works like ThreeStudio, Cap3D, Stable-DreamFusion, ImageReward, and LAVIS. Further community or maintenance details are not explicitly provided in the README.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The benchmark relies on the ThreeStudio implementation, which may introduce specific dependencies or limitations. The evaluation metrics are automatic and their correlation with human judgment, while claimed to be close, may still have nuances not fully captured.
1 year ago
1+ week