Generate immersive 3D worlds from text or pixels
New!
Top 31.0% on sourcepulse
HunyuanWorld-1.0 generates immersive, explorable, and interactive 3D worlds from text or pixel inputs. It targets developers in virtual reality, game development, and interactive content creation, offering a novel approach to 3D scene generation that combines panoramic proxies with mesh export and disentangled object representations.
How It Works
The framework utilizes a semantically layered 3D mesh representation, leveraging panoramic images as 360° world proxies. This approach facilitates semantic-aware decomposition and reconstruction, enabling the generation of diverse 3D worlds. Key advantages include 360° immersive experiences, mesh export for compatibility with existing graphics pipelines, and disentangled object representations for enhanced interactivity.
Quick Start & Requirements
docker/HunyuanWorld.yaml
), installing dependencies including Real-ESRGAN and ZIM, and logging into Hugging Face. Draco installation is also recommended for exporting to Draco format.modelviewer.html
is included for local visualization.Highlighted Details
Maintenance & Community
The project was released on July 26, 2025, with a technical report. Community channels include WeChat, Xiaohongshu, X (formerly Twitter), and Discord.
Licensing & Compatibility
The repository acknowledges contributions from various open-source projects. Specific licensing for HunyuanWorld-1.0 is not explicitly stated in the README, but compatibility with commercial use or closed-source linking would require clarification.
Limitations & Caveats
The README mentions that certain scenes may fail to load in the ModelViewer due to hardware limitations. A TensorRT version and RGBD video diffusion are listed as future open-source plans, indicating these features are not yet available.
1 day ago
Inactive