AI inference router for specialized workflows
Top 48.2% on sourcepulse
WilmerAI acts as a sophisticated intermediary for Large Language Models (LLMs), enabling users to route prompts to specialized LLM workflows based on domain or persona. It allows multiple LLMs to collaborate on generating a single response, enhancing output quality and enabling complex AI assistant configurations. The project is targeted at users who want to orchestrate multiple LLMs for advanced tasks, including RAG and iterative response refinement.
How It Works
WilmerAI routes prompts through user-defined workflows, which are sequences of LLM calls. These workflows can incorporate custom Python scripts, external APIs like the Offline Wikipedia API, and conditional logic. The system supports distributing LLM inference across multiple machines and leverages Ollama's model hotswapping to maximize VRAM usage on systems with limited GPU memory. It exposes OpenAI and Ollama compatible API endpoints for seamless integration with various front-end applications.
Quick Start & Requirements
pip install -r requirements.txt
, then start the server with python server.py
. Alternatively, use provided .bat
(Windows) or .sh
(macOS) scripts.Highlighted Details
Maintenance & Community
This is a personal project under heavy development, maintained by "Socg" in their free time. Updates may take a week or two. Contact is available via WilmerAI.Project@gmail.com.
Licensing & Compatibility
WilmerAI is licensed under the GNU General Public License v3.0 or later. This license permits redistribution and modification but requires derived works to also be licensed under the GPL, potentially restricting commercial use or linking with closed-source applications.
Limitations & Caveats
WilmerAI does not currently track or report token usage, requiring users to monitor costs via their LLM API dashboards. The project is in heavy development, and the README explicitly states it may contain bugs or incomplete code. The quality of WilmerAI's output is highly dependent on the connected LLMs and the user's configuration of prompts and presets. Linux support is not provided due to a lack of testing.
4 days ago
1 day