Discover and explore top open-source AI tools and projects—updated daily.
taco-groupOpen-source implementation of Waymo EMMA for autonomous driving
Top 42.5% on SourcePulse
OpenEMMA is an open-source framework for end-to-end motion planning in autonomous driving, replicating Waymo's EMMA model. It targets researchers and developers, leveraging pre-trained Vision Language Models (VLMs) to integrate text and camera inputs for predicting future waypoints and generating decision rationales.
How It Works
OpenEMMA integrates multimodal inputs, including front-view camera images and textual descriptions, with large Vision Language Models (VLMs) like GPT-4 and LLaVA. This approach allows the model to understand complex driving scenarios, predict future trajectories, and provide human-readable explanations for its decisions, aiming for more interpretable and robust autonomous driving systems.
Quick Start & Requirements
pip install openemmapip install -r requirements.txtcudatoolkit (tested with CUDA 12.4.0).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as a reproduction and may require significant computational resources for training and inference, particularly for larger VLM models. The README does not specify performance benchmarks or hardware requirements beyond CUDA.
5 months ago
1 week