API unifier for LLMs, converting to OpenAI format with load balancing
Top 37.0% on sourcepulse
This project provides a unified API gateway for managing multiple Large Language Model (LLM) services, abstracting away provider-specific differences and offering features like load balancing and OpenAI-compatible endpoints. It's designed for users who need to integrate various LLM backends without complex frontend interfaces, benefiting from a single point of access and flexible configuration.
How It Works
Uni-API acts as a reverse proxy, accepting requests in OpenAI format and routing them to configured backend LLM providers. It uniformly converts responses to the OpenAI format. Key features include sophisticated load balancing strategies (channel-level weighted, regional, sequential, round-robin), automatic retries, channel cooling mechanisms for failed providers, fine-grained model timeout settings, and permission controls.
Quick Start & Requirements
docker run --user root -p 8001:8000 --name uni-api -dit -v ./api.yaml:/home/api.yaml yym68686/uni-api:latest
CONFIG_URL
environment variable for remote configuration.api.yaml
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 day ago
1 day