Kolosal  by KolosalAI

Desktop app for local LLM training and inference

created 9 months ago
290 stars

Top 91.7% on sourcepulse

GitHubView on GitHub
Project Summary

Kolosal AI is an open-source desktop application for running large language models (LLMs) offline on personal devices. It targets users seeking a lightweight, privacy-focused alternative to cloud-based AI services, enabling local inference and custom model training on a wide range of hardware.

How It Works

Kolosal AI is built using C++17 and CMake, compiled into a compact ~20 MB executable. It leverages the Genta Personal Engine, which is based on llama.cpp, to support various LLMs like Mistral, LLaMA, and Qwen. The application is designed for universal hardware compatibility, running on CPUs with AVX2 instructions and supporting AMD and NVIDIA GPUs, with an optional Vulkan backend for GPU acceleration.

Quick Start & Requirements

  • Install: Clone the repository, ensure dependencies are met, configure with CMake (e.g., cmake -S .. -B . -DCMAKE_BUILD_TYPE=Release), and build (e.g., cmake --build . --config Release).
  • Prerequisites: CMake 3.14+, C++17 compiler (MSVC, GCC 7+), OpenSSL, CURL. Vulkan SDK is optional for Vulkan backend.
  • Setup: Requires manual download/placement of OpenSSL and CURL binaries if not system-installed.
  • Docs: https://github.com/Genta-Technology/Kolosal

Highlighted Details

  • Lightweight (~20 MB compiled) and portable, suitable for edge devices.
  • Universal hardware support: AVX2 CPUs, AMD & NVIDIA GPUs.
  • Powered by Genta Personal Engine (based on llama.cpp).
  • Supports popular LLMs (Mistral, LLaMA, Qwen).
  • Facilitates local dataset generation and model training.

Maintenance & Community

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatible with commercial use and closed-source linking.
  • External dependencies use their own licenses (MIT, zlib, Public Domain).

Limitations & Caveats

The build process requires manual management of external dependencies like OpenSSL and CURL if not system-installed. The Windows-specific resource file (resource.rc) may require modification for Linux/macOS builds.

Health Check
Last commit

2 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
3
Star History
58 stars in the last 90 days

Explore Similar Projects

Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Anil Dash Anil Dash(Former CEO of Glitch), and
15 more.

llamafile by Mozilla-Ocho

0.2%
23k
Single-file LLM distribution and runtime via `llama.cpp` and Cosmopolitan Libc
created 1 year ago
updated 1 month ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Pietro Schirano Pietro Schirano(Founder of MagicPath), and
1 more.

SillyTavern by SillyTavern

3.2%
17k
LLM frontend for power users
created 2 years ago
updated 3 days ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Nat Friedman Nat Friedman(Former CEO of GitHub), and
32 more.

llama.cpp by ggml-org

0.4%
84k
C/C++ library for local LLM inference
created 2 years ago
updated 22 hours ago
Feedback? Help us improve.