API protocol for LLM agent production deployment
Top 72.3% on sourcepulse
This project defines a framework-agnostic API protocol for serving Large Language Model (LLM) agents in production. It aims to standardize interactions around core concepts like runs (stateless or background executions), threads (multi-turn conversations with state management), and a key-value store for long-term memory. The protocol is designed for developers building and deploying LLM-powered applications, offering a consistent interface for agent execution, state persistence, and introspection.
How It Works
The protocol is built around RESTful endpoints defined in an OpenAPI specification. It separates agent execution into "runs" which can be ephemeral (stateless, single-shot) or persistent (associated with a thread). Threads provide a mechanism for managing conversational history, state revisions, and concurrency control. A key-value "store" API allows agents to persist and retrieve arbitrary data, enabling long-term memory capabilities. Introspection endpoints enable querying agent capabilities and schemas.
Quick Start & Requirements
Highlighted Details
/runs/wait
, /runs/stream
) and multi-turn threads (/threads
, /threads/{thread_id}/runs
)./agents/search
, /agents/{agent_id}/schemas
)./store/items
, /store/namespaces
) for persistent memory.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 months ago
1+ week