Research paper for zero-shot one image to 3D object generation
Top 16.7% on sourcepulse
This repository provides Zero-1-to-3, a novel method for generating 3D objects from a single input image. It addresses the challenge of zero-shot novel view synthesis and 3D reconstruction, targeting researchers and developers in computer vision and graphics. The primary benefit is enabling high-quality 3D asset creation from minimal input.
How It Works
Zero-1-to-3 leverages a finetuned Stable Diffusion model to generate novel views of an object from a single input image. It explicitly models camera pose changes, trained on a large dataset of 3D object renderings (Objaverse). This approach alleviates the "Janus problem" (viewpoint ambiguity) inherent in text-to-image models by ensuring consistency and accuracy across synthesized viewpoints, facilitating 3D reconstruction.
Quick Start & Requirements
conda create -n zero123 python=3.9
, conda activate zero123
), pip install -r requirements.txt
, pip install -e taming-transformers/
, pip install -e CLIP/
.105000.ckpt
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The training script is preliminary and configured for an 8x A100 system, requiring adjustments for smaller GPU setups. Hyperparameters for 3D reconstruction are not extensively tuned.
1 year ago
1 day