Discover and explore top open-source AI tools and projects—updated daily.
cvlab-columbiaResearch paper for zero-shot one image to 3D object generation
Top 15.8% on SourcePulse
This repository provides Zero-1-to-3, a novel method for generating 3D objects from a single input image. It addresses the challenge of zero-shot novel view synthesis and 3D reconstruction, targeting researchers and developers in computer vision and graphics. The primary benefit is enabling high-quality 3D asset creation from minimal input.
How It Works
Zero-1-to-3 leverages a finetuned Stable Diffusion model to generate novel views of an object from a single input image. It explicitly models camera pose changes, trained on a large dataset of 3D object renderings (Objaverse). This approach alleviates the "Janus problem" (viewpoint ambiguity) inherent in text-to-image models by ensuring consistency and accuracy across synthesized viewpoints, facilitating 3D reconstruction.
Quick Start & Requirements
conda create -n zero123 python=3.9, conda activate zero123), pip install -r requirements.txt, pip install -e taming-transformers/, pip install -e CLIP/.105000.ckpt).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The training script is preliminary and configured for an 8x A100 system, requiring adjustments for smaller GPU setups. Hyperparameters for 3D reconstruction are not extensively tuned.
2 years ago
Inactive
ashawkey