PyTorch code for multi-view diffusion-based 3D generation research
Top 83.6% on sourcepulse
This repository provides the official PyTorch implementation for "3D-Adapter: Geometry-Consistent Multi-View Diffusion for High-Quality 3D Generation" and "Generic 3D Diffusion Adapter Using Controlled Multi-View Editing." It enables high-quality 3D asset generation and editing through controlled multi-view diffusion, targeting researchers and developers in computer vision and graphics.
How It Works
The project leverages multi-view diffusion models to achieve geometry-consistent 3D generation. It acts as an adapter, integrating with existing diffusion pipelines to guide the generation process across multiple views, ensuring spatial coherence and high fidelity in the resulting 3D assets. The approach utilizes off-the-shelf models for optimization-based adapters, requiring no further training for this variant.
Quick Start & Requirements
pip install -r requirements.txt
. A conda environment with Python 3.10, PyTorch 2.1.2, and CUDA 12.1 is recommended. FFmpeg and x264 are optional for video export.tiny-cuda-nn
.python app.py --unload-models
to start the Gradio Web UI. A GPU with at least 24GB VRAM is required.Highlighted Details
Maintenance & Community
The project is associated with Stanford University, Apparate Labs, and UCSD. GRM-based 3D-Adapter models are pending release alongside GRM.
Licensing & Compatibility
The repository does not explicitly state a license in the provided README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
GRM-based 3D-Adapters are not yet released. Certain packages may require specific configuration for Windows installation. API documentation may contain inaccuracies in data types and default values.
7 months ago
Inactive