llama_deploy  by run-llama

Async framework for deploying agentic workflows to production

Created 1 year ago
2,054 stars

Top 21.7% on SourcePulse

GitHubView on GitHub
Project Summary

LlamaDeploy is an async-first framework for deploying, scaling, and productionizing agentic multi-service systems built with llama_index workflows. It targets developers who want to transition notebook-based agent prototypes to robust, scalable cloud services with minimal code changes, offering a HTTP API for integration.

How It Works

LlamaDeploy utilizes a microservices architecture with a hub-and-spoke design. Workflows defined in llama_index are exposed as individual services. This approach allows for independent scaling of components, flexible swapping of underlying technologies (e.g., message queues), and inherent fault tolerance through built-in retry mechanisms. The control plane manages inter-service state, simplifying complex, multi-step agentic processes.

Quick Start & Requirements

Highlighted Details

  • Seamlessly deploys llama_index workflows to production with minimal code modification.
  • Microservices architecture enables independent scaling and fault tolerance.
  • Hub-and-spoke design allows for flexible component swapping and extensibility.
  • Async-first design supports high-concurrency and real-time applications.

Maintenance & Community

  • Formerly known as llama-agents.
  • Focus shifted to bridging local development and remote execution of agents as services.

Licensing & Compatibility

  • License not specified in the README.

Limitations & Caveats

The README does not specify the project's license, which is crucial for determining commercial use compatibility.

Health Check
Last Commit

2 weeks ago

Responsiveness

1 day

Pull Requests (30d)
2
Issues (30d)
0
Star History
8 stars in the last 30 days

Explore Similar Projects

Feedback? Help us improve.