Kolosal  by KolosalAI

Desktop app for local LLM training and inference

Created 11 months ago
297 stars

Top 89.4% on SourcePulse

GitHubView on GitHub
Project Summary

Kolosal AI is an open-source desktop application for running large language models (LLMs) offline on personal devices. It targets users seeking a lightweight, privacy-focused alternative to cloud-based AI services, enabling local inference and custom model training on a wide range of hardware.

How It Works

Kolosal AI is built using C++17 and CMake, compiled into a compact ~20 MB executable. It leverages the Genta Personal Engine, which is based on llama.cpp, to support various LLMs like Mistral, LLaMA, and Qwen. The application is designed for universal hardware compatibility, running on CPUs with AVX2 instructions and supporting AMD and NVIDIA GPUs, with an optional Vulkan backend for GPU acceleration.

Quick Start & Requirements

  • Install: Clone the repository, ensure dependencies are met, configure with CMake (e.g., cmake -S .. -B . -DCMAKE_BUILD_TYPE=Release), and build (e.g., cmake --build . --config Release).
  • Prerequisites: CMake 3.14+, C++17 compiler (MSVC, GCC 7+), OpenSSL, CURL. Vulkan SDK is optional for Vulkan backend.
  • Setup: Requires manual download/placement of OpenSSL and CURL binaries if not system-installed.
  • Docs: https://github.com/Genta-Technology/Kolosal

Highlighted Details

  • Lightweight (~20 MB compiled) and portable, suitable for edge devices.
  • Universal hardware support: AVX2 CPUs, AMD & NVIDIA GPUs.
  • Powered by Genta Personal Engine (based on llama.cpp).
  • Supports popular LLMs (Mistral, LLaMA, Qwen).
  • Facilitates local dataset generation and model training.

Maintenance & Community

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatible with commercial use and closed-source linking.
  • External dependencies use their own licenses (MIT, zlib, Public Domain).

Limitations & Caveats

The build process requires manual management of external dependencies like OpenSSL and CURL if not system-installed. The Windows-specific resource file (resource.rc) may require modification for Linux/macOS builds.

Health Check
Last Commit

3 months ago

Responsiveness

1 week

Pull Requests (30d)
0
Issues (30d)
2
Star History
7 stars in the last 30 days

Explore Similar Projects

Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Gabriel Almeida Gabriel Almeida(Cofounder of Langflow), and
2 more.

torchchat by pytorch

0.1%
4k
PyTorch-native SDK for local LLM inference across diverse platforms
Created 1 year ago
Updated 1 week ago
Starred by Omar Sanseviero Omar Sanseviero(DevRel at Google DeepMind), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
11 more.

petals by bigscience-workshop

0.1%
10k
Run LLMs at home, BitTorrent-style
Created 3 years ago
Updated 1 year ago
Feedback? Help us improve.