Local GenAI server for private, offline AI
Top 78.4% on sourcepulse
Edgen provides a local, private GenAI server that acts as a drop-in replacement for OpenAI's API, targeting developers and users who need on-device inference for LLMs, speech-to-text, and other AI models. It offers data privacy, reliability, and cost savings by leveraging user hardware, eliminating the need for cloud infrastructure and API keys.
How It Works
Built in Rust, Edgen is natively compiled for Windows, macOS, and Linux, abstracting the complexities of AI model optimization across different hardware and platforms. It supports model caching to avoid redundant downloads and offers modularity for easy integration of new models and runtimes. The server implements an OpenAI-compatible API, exposing endpoints for chat completions, audio transcriptions, and embeddings, with plans for image generation and multimodal chat.
Quick Start & Requirements
serve
command starts the server (default behavior).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Image generation and multimodal chat completions are listed as future features. The README notes that Vulkan, CUDA, and Metal GPU features cannot be enabled simultaneously.
1 year ago
1 week