Discover and explore top open-source AI tools and projects—updated daily.
bestruiruiUnified LLM API gateway with intelligent load balancing
Top 37.9% on SourcePulse
Summary
Octopus addresses the need for unified management and efficient access to multiple Large Language Model (LLM) APIs for individual users. It aggregates various LLM provider channels, offering load balancing, protocol conversion, and cost tracking through an elegant web UI. This service benefits users by simplifying API integration, ensuring service stability, and providing insights into LLM usage costs.
How It Works
The service functions as an LLM API gateway, connecting to multiple LLM providers as "channels." These channels can be organized into "groups," allowing unified access under a single model name. Octopus supports several load balancing strategies (Round Robin, Random, Failover, Weighted) to distribute requests efficiently across configured channels. It also performs seamless protocol conversion between OpenAI Chat, OpenAI Responses, and Anthropic API formats, simplifying integration with diverse LLM backends. Model pricing and availability are automatically synchronized from external sources like models.dev.
Quick Start & Requirements
docker run -d --name octopus -v /path/to/data:/app/data -p 8080:8080 bestrui/octopus), Docker Compose (wget https://raw.githubusercontent.com/bestrui/octopus/refs/heads/dev/docker-compose.yml && docker compose up -d), or download pre-compiled binaries from Releases.admin, Password: admin (immediate change recommended).Highlighted Details
models.dev.Maintenance & Community
The project acknowledges contributions from looplj/axonhub for its LLM API adaptation module and sst/models.dev for AI model pricing data. No specific community channels (like Discord/Slack) or details on active maintainers/sponsorships are provided in the README.
Licensing & Compatibility
The provided README text does not specify the software license. Users should verify licensing terms before adoption, especially for commercial use.
Limitations & Caveats
Proper shutdown procedures (e.g., Ctrl+C, SIGTERM) are critical to prevent data loss for in-memory statistics; forced termination (kill -9) is strongly discouraged. Default administrative credentials require immediate modification for security. Building from source involves a multi-step process including frontend compilation.
3 days ago
Inactive
theopenco
songquanpeng