Embodied AI brain model for robotics
Top 59.9% on SourcePulse
RoboBrain 2.0 is an advanced open-source embodied AI system designed to unify perception, reasoning, and planning for complex robotic tasks. Targeting researchers and developers in embodied AI, it offers a powerful solution for creating generalist embodied agents capable of understanding and interacting with physical environments.
How It Works
RoboBrain 2.0 features a heterogeneous architecture combining a vision encoder with a large language model (LLM). It processes multi-modal inputs, including images, long videos, and structured scene graphs, alongside complex task instructions. The LLM decoder performs chain-of-thought reasoning to output structured plans, spatial relations, and coordinates, enabling capabilities like spatial understanding and temporal decision-making.
Quick Start & Requirements
pip install -r requirements.txt
within a conda
environment (Python 3.10 recommended).BAAI/RoboBrain2.0-7B
) from Hugging Face or ModelScope.Highlighted Details
Maintenance & Community
The project is associated with BAAI (Beijing Academy of Artificial Intelligence). Contact information via WeChat and RedNote is provided.
Licensing & Compatibility
The README does not explicitly state a license. Further investigation is required for commercial use or closed-source linking.
Limitations & Caveats
The README does not detail specific limitations or known issues. The project appears to be actively developed with recent updates in July 2025.
5 days ago
Inactive