Evaluation metric for text-to-3D generative models
Top 95.4% on sourcepulse
This repository provides an implementation for evaluating text-to-3D generative models using GPT-4V as a human-aligned evaluator. It offers a framework for generating ELO scores for new methods by comparing them against existing ones, targeting researchers and developers in the 3D generation field.
How It Works
The core of the system leverages GPT-4V's ability to analyze rendered images of 3D models. For each text prompt, users must provide 120 RGB and normal map renderings of their generated 3D models. These renders are then fed to GPT-4V, which acts as an impartial judge, assigning scores that contribute to an ELO rating for the evaluated method. This approach aims to automate and standardize the qualitative evaluation process, which is often subjective and time-consuming.
Quick Start & Requirements
pip install --upgrade openai tqdm numpy Pillow gdown
python gpt_eval_alpha.py
with appropriate arguments.Highlighted Details
Maintenance & Community
The project is associated with CVPR 2024 and acknowledges contributions from several prominent 3D generation projects, including threestudio, mvdream, and shap-e. Further utilities and visualization tools are planned.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. However, given its reliance on OpenAI's API, users must adhere to OpenAI's terms of service. Compatibility with commercial or closed-source projects is not specified.
Limitations & Caveats
The evaluation process is dependent on the OpenAI API, which may incur costs and is subject to OpenAI's rate limits and availability. The quality of evaluation is directly tied to GPT-4V's capabilities and potential biases. The project is described as "alpha" in the evaluation command, suggesting potential instability or ongoing development.
1 year ago
1+ week