Discover and explore top open-source AI tools and projects—updated daily.
aurora-developSimple API server for LLM access
Top 46.1% on SourcePulse
Duck2api provides a unified API endpoint for various large language models, acting as a proxy to simplify integration for developers. It supports models like Claude-3-Haiku, Llama-3.3-70b, Mixtral-8x7b, and GPT-4o-mini, offering a single interface to interact with different LLM providers.
How It Works
Duck2api functions as a reverse proxy, forwarding requests to specified LLM APIs. It abstracts away the complexities of individual model integrations, presenting a consistent API structure. This approach allows users to switch between different LLM backends without modifying their client applications, promoting flexibility and reducing vendor lock-in.
Quick Start & Requirements
git clone https://github.com/aurora-develop/duck2api && cd duck2api && go build -o duck2api && chmod +x ./duck2api && ./duck2apidocker run -d --name duck2api -p 8080:8080 ghcr.io/aurora-develop/duck2api:latestdocker-compose.yml and run docker-compose up -d.curl --location 'http://<your-server-ip>:8080/v1/chat/completions' --header 'Content-Type: application/json' --data '{ "model": "gpt-4o-mini", "messages": [{"role": "user", "content": "Say this is a test!"}], "stream": true }'http://<your-server-ip>:8080/web.Highlighted Details
/v1/chat/completions API endpoint.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project mentions that gpt-3.5-turbo is no longer supported due to the DuckDuckGo API removing support. The README does not detail specific performance benchmarks or advanced configuration options beyond environment variables.
10 months ago
Inactive
modelcontextprotocol
PrefectHQ