Self-hosted API for LLM experimentation, mimicking OpenAI endpoints
Top 83.8% on sourcepulse
SimpleAI provides a self-hosted API for large language models, acting as an alternative to proprietary services. It targets developers and researchers who want to experiment with various AI models, create benchmarks, or handle specific use cases without relying on external providers. The primary benefit is flexibility and independence in AI model deployment and integration.
How It Works
SimpleAI exposes OpenAI-compatible API endpoints (text completion, chat completion, embeddings) via gRPC. This architectural choice decouples the API layer from model inference, allowing for multi-language support and easier integration of diverse models. Users define and expose their models by implementing specific Python interfaces and configuring them in models.toml
, pointing to gRPC services that host the actual model inference.
Quick Start & Requirements
pip install simple_ai_server
or from source: pip install git+https://github.com/lhenault/simpleAI
simple_ai init
to create models.toml
, then simple_ai serve
to start the server.Highlighted Details
Maintenance & Community
make
and poetry
.Licensing & Compatibility
Limitations & Caveats
Image, audio, file, fine-tune, and moderation endpoints are not yet supported. The project is actively under development, and some features may be experimental or incomplete. Customization for CORS or API versioning requires custom server scripts.
1 year ago
1 day