Benchmark for embodied multi-agent task planning using LLMs in Habitat simulator
Top 86.7% on sourcepulse
This repository provides the PARTNR benchmark for evaluating Large Planning Models (LPMs) in human-robot collaboration and instruction-following tasks within the Habitat simulator. It offers a framework for agents to interpret natural language commands and interact with simulated environments, targeting researchers and developers in embodied AI and robotics.
How It Works
The system models agents, tools, and skills within the Habitat simulator. Agents interact with the environment using a set of tools, which abstract perception and low-level actions. Planners, often LLM-based, leverage a hierarchical WorldGraph representing the environment's state to select appropriate tools for task execution. A simulated perception pipeline feeds local detections into the world model, enabling agents to reason about their surroundings.
Quick Start & Requirements
python -m habitat_llm.examples.planner_demo
with specified config files (e.g., baselines/decentralized_zero_shot_react_summary.yaml
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The skill_runner
currently supports only sequential skill execution for one agent at a time, limiting simultaneous multi-agent interaction within that specific tool. Constrained generation is not supported for the OpenAI backend.
3 months ago
1+ week