Discover and explore top open-source AI tools and projects—updated daily.
Research paper source code for cooperative embodied agents using LLMs
Top 95.2% on SourcePulse
This repository provides the source code for the ICLR 2024 paper "Building Cooperative Embodied Agents Modularly with Large Language Models." It enables the creation of modular, LLM-powered embodied agents capable of complex cooperative tasks in simulated environments, targeting researchers and developers in embodied AI and multi-agent systems.
How It Works
The project leverages large language models (LLMs) to control embodied agents, breaking down complex tasks into modular components. Agents interact within simulated environments like ThreeDWorld Multi-Agent Transport (TDW-MAT) and Communicative Watch-And-Help (C-WAH), utilizing LLMs for decision-making, communication, and task execution. This modular LLM-driven approach facilitates emergent cooperative behaviors and simplifies the development of sophisticated embodied AI.
Quick Start & Requirements
pip install -e .
within the tdw_mat
directory. A demo scene can be run with python demo/demo_scene.py
.pip install -r requirements.txt
within the cwah
directory.Highlighted Details
Maintenance & Community
The project is associated with UMass-Embodied-AGI and has recent updates addressing navigation issues in TDW-MAT. The paper was published in ICLR 2024.
Licensing & Compatibility
The repository itself does not explicitly state a license. The included VirtualHome repository is under the MIT license. Compatibility for commercial use depends on the licenses of the underlying LLMs and simulators used.
Limitations & Caveats
The setup for both environments can be complex, requiring specific simulator versions and configurations. The reliance on external LLM APIs (like GPT-4) introduces potential costs and dependencies. The project is presented as research code, and stability for production use is not guaranteed.
5 months ago
Inactive