Open-source implementation of Waymo EMMA for autonomous driving
Top 46.7% on sourcepulse
OpenEMMA is an open-source framework for end-to-end motion planning in autonomous driving, replicating Waymo's EMMA model. It targets researchers and developers, leveraging pre-trained Vision Language Models (VLMs) to integrate text and camera inputs for predicting future waypoints and generating decision rationales.
How It Works
OpenEMMA integrates multimodal inputs, including front-view camera images and textual descriptions, with large Vision Language Models (VLMs) like GPT-4 and LLaVA. This approach allows the model to understand complex driving scenarios, predict future trajectories, and provide human-readable explanations for its decisions, aiming for more interpretable and robust autonomous driving systems.
Quick Start & Requirements
pip install openemma
pip install -r requirements.txt
cudatoolkit
(tested with CUDA 12.4.0).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as a reproduction and may require significant computational resources for training and inference, particularly for larger VLM models. The README does not specify performance benchmarks or hardware requirements beyond CUDA.
2 months ago
1 week