Research paper on executable code actions for LLM agents
Top 31.2% on sourcepulse
This repository provides the official implementation for "Executable Code Actions Elicit Better LLM Agents," introducing CodeAct, a unified action space for LLM agents that leverages executable Python code. It aims to improve agent performance by allowing dynamic revision of actions based on execution results, targeting researchers and developers building sophisticated LLM-powered agents.
How It Works
CodeAct integrates a Python interpreter to execute code actions, enabling multi-turn interactions where agents can dynamically revise or emit new actions based on observations like code execution outcomes. This approach consolidates LLM actions into a unified, executable space, outperforming traditional Text and JSON action formats in benchmarks like M3 ToolEval.
Quick Start & Requirements
./scripts/chat/start_vllm.sh
.llama.cpp
build and optionally converting models to GGUF format. Run ./server -m <model_path>
../scripts/chat/code_execution/start_jupyter_server.sh 8081
.scripts/chat/demo.py
for CLI or configure chat-ui
for a web interface.git lfs
.Highlighted Details
Maintenance & Community
The project is associated with ICML 2024. Links to Hugging Face for data and models are provided. No specific community channels (Discord/Slack) are mentioned in the README.
Licensing & Compatibility
The repository does not explicitly state a license. The models are hosted on Hugging Face, which typically uses specific licenses per model. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The primary LLM serving method (vLLM) requires NVIDIA GPUs and Docker. While llama.cpp support is available for laptop inference, it involves model conversion and compilation steps. The project appears to be released alongside a research paper, and its long-term maintenance status is not detailed.
1 year ago
1+ week