Text-to-3D model for generating detailed 3D assets
Top 66.2% on sourcepulse
RichDreamer addresses the challenge of generating high-detail 3D assets from text prompts by leveraging a generalizable Normal-Depth diffusion model. It targets researchers and practitioners in computer graphics and AI who need to create detailed 3D content, offering improved realism and fidelity over existing text-to-3D methods.
How It Works
The core of RichDreamer is a diffusion model conditioned on both normal and depth maps, which are generated alongside the 3D representation. This approach allows the diffusion process to explicitly control and enhance geometric details and surface properties. The model can generate multi-view consistent normal and depth maps, which are then used to reconstruct a 3D mesh, potentially via NeRF or DMTet representations, leading to richer details.
Quick Start & Requirements
conda create -n rd
, conda activate rd
), and install dependencies (pip install -r requirements_3d.txt
).registry.cn-hangzhou.aliyuncs.com/ailab-public/aigc3d
) or build from docker/Dockerfile
.Highlighted Details
Maintenance & Community
The project is associated with ModelScope and Damo_XR_Lab. Links to related projects like normal-depth-diffusion
and gobjaverse
are provided.
Licensing & Compatibility
The repository does not explicitly state a license in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The README mentions that optimizing high-resolution DMTet spheres directly can be challenging and may require multiple GPUs, although a single GPU optimization trick is provided. Some dependencies might require specific versions or manual setup if not using Docker.
10 months ago
1 week