Discover and explore top open-source AI tools and projects—updated daily.
taco-groupOpen-source implementation of Waymo EMMA for autonomous driving
Top 39.6% on SourcePulse
OpenEMMA is an open-source framework for end-to-end motion planning in autonomous driving, replicating Waymo's EMMA model. It targets researchers and developers, leveraging pre-trained Vision Language Models (VLMs) to integrate text and camera inputs for predicting future waypoints and generating decision rationales.
How It Works
OpenEMMA integrates multimodal inputs, including front-view camera images and textual descriptions, with large Vision Language Models (VLMs) like GPT-4 and LLaVA. This approach allows the model to understand complex driving scenarios, predict future trajectories, and provide human-readable explanations for its decisions, aiming for more interpretable and robust autonomous driving systems.
Quick Start & Requirements
pip install openemmapip install -r requirements.txtcudatoolkit (tested with CUDA 12.4.0).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as a reproduction and may require significant computational resources for training and inference, particularly for larger VLM models. The README does not specify performance benchmarks or hardware requirements beyond CUDA.
11 months ago
1 week