Benchmark dataset for embodied question answering (EQA) research
Top 88.6% on sourcepulse
OpenEQA introduces a new formulation for Embodied Question Answering (EQA), enabling agents to answer questions about environments by leveraging episodic memory or active exploration. It targets researchers in Embodied AI and conversational agents, providing a benchmark dataset and an LLM-powered evaluation protocol to challenge foundation models.
How It Works
The project defines EQA as understanding an environment to answer questions in natural language. This understanding is achieved either through recalling past experiences (episodic memory) or by actively exploring the physical space. The OpenEQA dataset, comprising over 1600 human-generated question-answer pairs across 180 real-world environments, supports both these approaches. An automatic evaluation protocol using GPT-4 is also provided, demonstrating high correlation with human judgment.
Quick Start & Requirements
conda create -n openeqa python=3.9
, conda activate openeqa
, pip install -r requirements.txt
, pip install -e .
Highlighted Details
Maintenance & Community
The project is from Facebook Research, with notable contributors including Arjun Majumdar, Dhruv Batra, and Franziska Meier. No community links (Discord/Slack) are provided in the README.
Licensing & Compatibility
Limitations & Caveats
The README indicates that current foundation models significantly lag behind human-level performance on this benchmark, suggesting it poses a considerable challenge. Episode histories require a separate download process.
10 months ago
Inactive