Kolo  by MaxHastings

CLI tool for local LLM fine-tuning automation

Created 7 months ago
320 stars

Top 84.6% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

Kolo streamlines LLM fine-tuning by automating environment setup and providing a unified interface for popular tools like Unsloth, Torchtune, Llama.cpp, and Ollama. It targets AI researchers and developers seeking a rapid, hassle-free local fine-tuning experience, reducing setup time to minutes.

How It Works

Kolo leverages Docker to create a consistent, pre-configured environment, eliminating dependency conflicts. It integrates Unsloth for faster training and lower VRAM usage, Torchtune for PyTorch-native fine-tuning (including AMD GPU and CPU support), and Llama.cpp for GGUF conversion and quantization. Ollama manages model deployment, and Open WebUI provides a testing interface. This stacked approach offers flexibility and performance for local LLM experimentation.

Quick Start & Requirements

  • Install: Requires Docker Desktop (Windows/Linux) and WSL2/HyperV on Windows. AMD GPU users need ROCm installed on Linux.
  • Build: ./build_image.ps1 (or ./build_image_amd.ps1 for AMD).
  • Run: ./create_and_run_container.ps1 (or ./create_and_run_container_amd.ps1 for AMD).
  • Data: ./copy_training_data.ps1
  • Train: ./train_model_unsloth.ps1 or ./train_model_torchtune.ps1
  • Install Model: ./install_model.ps1
  • Test: Access localhost:8080 in a browser.
  • Docs: Fine Tune Training Guide (linked within README).

Highlighted Details

  • Supports both NVIDIA (CUDA 12.1+) and AMD GPUs (Linux only via ROCm).
  • Offers fine-tuning via Unsloth (faster, lower VRAM) and Torchtune (PyTorch native).
  • Includes scripts for model installation, uninstallation, and listing.
  • Provides SSH and SFTP access to the Docker container for advanced users.

Maintenance & Community

  • Community support via Discord (link provided in README).

Licensing & Compatibility

  • The README does not explicitly state the project's license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

  • AMD GPU support is restricted to Linux; Windows WSL2 is not supported.
  • Torchtune requires Hugging Face account and model access permissions.
  • Re-training with the same output directory requires manual deletion via ./delete_model.ps1.
Health Check
Last Commit

6 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
4 stars in the last 30 days

Explore Similar Projects

Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
6 more.

xTuring by stochasticai

0.0%
3k
SDK for fine-tuning and customizing open-source LLMs
Created 2 years ago
Updated 1 day ago
Starred by Casper Hansen Casper Hansen(Author of AutoAWQ), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
5 more.

xtuner by InternLM

0.5%
5k
LLM fine-tuning toolkit for research
Created 2 years ago
Updated 1 day ago
Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Roy Frostig Roy Frostig(Coauthor of JAX; Research Scientist at Google DeepMind), and
19 more.

ml-engineering by stas00

0.4%
15k
Open book for LLM/VLM training engineers
Created 5 years ago
Updated 1 day ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Stefan van der Walt Stefan van der Walt(Core Contributor to scientific Python ecosystem), and
12 more.

litgpt by Lightning-AI

0.1%
13k
LLM SDK for pretraining, finetuning, and deploying 20+ high-performance LLMs
Created 2 years ago
Updated 5 days ago
Starred by Junyang Lin Junyang Lin(Core Maintainer at Alibaba Qwen), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
25 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.