airunner  by Capsize-Games

Local AI inference engine for text, images, speech, and agent workflows

Created 2 years ago
1,232 stars

Top 32.0% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

AI Runner is an open-source, local-first application designed to run various AI models, including LLMs, Stable Diffusion, TTS, and STT, without cloud dependencies. It targets developers and end-users seeking an all-in-one, offline inference engine for prototyping, private data processing, or custom UI development.

How It Works

AI Runner leverages HuggingFace and Llama-index libraries to provide a unified interface for diverse AI tasks. Its architecture supports local LLM inference, Stable Diffusion image generation, text-to-speech, and speech-to-text. The project emphasizes a "local-first" approach, ensuring data privacy and offline functionality, while offering a plugin and extension API for customization and integration into other Python projects.

Quick Start & Requirements

  • Install/Run: Docker is recommended. After installing NVIDIA Container Toolkit, run ./src/airunner/bin/docker.sh airunner.
  • Prerequisites: NVIDIA GPU (RTX 3060 minimum, RTX 4090 recommended), Docker, NVIDIA Container Toolkit.
  • Storage: 22 GB minimum, 50 GB recommended for models.
  • Links: Installation Wiki, Discord

Highlighted Details

  • Unified interface for LLMs, Stable Diffusion, TTS, and STT.
  • Real-time voice conversations with LLMs.
  • Supports Stable Diffusion 1.5, SDXL, Turbo, ControlNet, LoRA, and embeddings.
  • Plugin and extension API for custom features and UI.
  • Can be installed as a Python library (pip install airunner).
  • Offers memory optimization techniques like TF32 Mode and VAE/Attention Slicing.

Maintenance & Community

The project welcomes contributions and provides a Discord server for questions and ideas. Detailed contribution guidelines are available in the repository.

Licensing & Compatibility

The repository does not explicitly state a license in the README. This requires further investigation for commercial use or closed-source linking.

Limitations & Caveats

The README does not specify a license, which is a critical factor for adoption, especially for commercial or closed-source projects. While Docker simplifies setup, it requires the NVIDIA Container Toolkit for GPU acceleration.

Health Check
Last Commit

3 days ago

Responsiveness

1 day

Pull Requests (30d)
24
Issues (30d)
2
Star History
19 stars in the last 30 days

Explore Similar Projects

Starred by Chris Lattner Chris Lattner(Author of LLVM, Clang, Swift, Mojo, MLIR; Cofounder of Modular), Tobi Lutke Tobi Lutke(Cofounder of Shopify), and
11 more.

modular by modular

0.1%
25k
AI toolchain unifying fragmented AI deployment workflows
Created 2 years ago
Updated 1 day ago
Feedback? Help us improve.