Async framework for deploying agentic workflows to production
Top 22.2% on sourcepulse
LlamaDeploy is an async-first framework for deploying, scaling, and productionizing agentic multi-service systems built with llama_index
workflows. It targets developers who want to transition notebook-based agent prototypes to robust, scalable cloud services with minimal code changes, offering a HTTP API for integration.
How It Works
LlamaDeploy utilizes a microservices architecture with a hub-and-spoke design. Workflows defined in llama_index
are exposed as individual services. This approach allows for independent scaling of components, flexible swapping of underlying technologies (e.g., message queues), and inherent fault tolerance through built-in retry mechanisms. The control plane manages inter-service state, simplifying complex, multi-step agentic processes.
Quick Start & Requirements
pip install llama_deploy
Highlighted Details
llama_index
workflows to production with minimal code modification.Maintenance & Community
llama-agents
.Licensing & Compatibility
Limitations & Caveats
The README does not specify the project's license, which is crucial for determining commercial use compatibility.
4 days ago
1 day