Discover and explore top open-source AI tools and projects—updated daily.
BeingBeyondVision-language-action model for robot learning
Top 87.1% on SourcePulse
Being-H0.5 is a foundational Vision-Language-Action (VLA) model designed to enhance cross-embodiment generalization in robot control. It scales human-centric learning using the UniHand-2.0 dataset and a unified action space, aiming to provide robust and adaptable robotic manipulation capabilities for researchers and practitioners.
How It Works
Being-H0.5 is a foundational Vision-Language-Action (VLA) model focused on scaling human-centric robot learning. It utilizes the UniHand-2.0 dataset and a unified action space to achieve robust cross-embodiment generalization, enabling policies to adapt across diverse robotic hardware. The architecture integrates visual perception, language understanding, and action generation for interpreting instructions and environmental states for robotic control.
Quick Start & Requirements
requirements.txt and flash-attn.cuda:0).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 days ago
Inactive
octo-models
Physical-Intelligence