h2o-llmstudio  by h2oai

LLM Studio: framework for LLM fine-tuning via GUI or CLI

Created 2 years ago
4,634 stars

Top 10.7% on SourcePulse

GitHubView on GitHub
Project Summary

H2O LLM Studio provides a no-code GUI and framework for fine-tuning large language models, targeting users who want to customize LLMs without extensive coding. It simplifies the process of adapting state-of-the-art models using various hyperparameters and modern techniques like LoRA and 8-bit training, enabling efficient model customization and evaluation.

How It Works

The framework supports fine-tuning LLMs using techniques such as Low-Rank Adaptation (LoRA) and 8-bit quantization for reduced memory footprint. It also incorporates Reinforcement Learning (RL) alternatives like Direct Preference Optimization (DPO), Identity Preference Optimization (IPO), and KTO for model tuning based on preference data. The system offers advanced evaluation metrics, visual performance tracking, and integrations with Neptune and Weights & Biases.

Quick Start & Requirements

  • Install: Recommended: pipenv install via make setup. Alternative: make setup-no-flash (disables Flash Attention 2). Docker installation is also available.
  • Prerequisites: Ubuntu 16.04+, NVIDIA GPU (>= 470.57.02 drivers), Python 3.10 recommended. For larger models, 24GB+ GPU memory is advised. CUDA 12.1 or 12.4 is recommended for DeepSpeed.
  • Resources: Setup involves installing Python, drivers, and dependencies. Docker requires NVIDIA Container Toolkit.
  • Links: Documentation: https://docs.h2o.ai/h2o-llmstudio/

Highlighted Details

  • Supports Causal Regression and Classification modeling.
  • Integrates DeepSpeed for distributed training on multi-GPU setups (requires NVLink).
  • Offers CLI for fine-tuning, interactive chat, and Hugging Face Hub publishing.
  • Recent updates include DPO/IPO/KTO optimization and removal of RLHF.

Maintenance & Community

  • Active development with recent PRs addressing new features and deprecations.
  • Discord server available for community discussion.
  • Model checkpoints and datasets are available on H2O.ai's Hugging Face page.

Licensing & Compatibility

  • Licensed under Apache 2.0.
  • Compatible with commercial use and closed-source linking.

Limitations & Caveats

The project notes that due to rapid development, full backward compatibility is not guaranteed, and users are advised to pin versions. RLHF is being deprecated and will be fully removed in a future release.

Health Check
Last Commit

14 hours ago

Responsiveness

1 day

Pull Requests (30d)
5
Issues (30d)
1
Star History
281 stars in the last 30 days

Explore Similar Projects

Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera) and Stas Bekman Stas Bekman(Author of "Machine Learning Engineering Open Book"; Research Engineer at Snowflake).

InternEvo by InternLM

0.2%
407
Lightweight training framework for model pre-training
Created 1 year ago
Updated 4 weeks ago
Starred by Didier Lopes Didier Lopes(Founder of OpenBB), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
5 more.

mlx-lm by ml-explore

26.1%
2k
Python package for LLM text generation and fine-tuning on Apple silicon
Created 6 months ago
Updated 1 day ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Stefan van der Walt Stefan van der Walt(Core Contributor to scientific Python ecosystem), and
12 more.

litgpt by Lightning-AI

0.1%
13k
LLM SDK for pretraining, finetuning, and deploying 20+ high-performance LLMs
Created 2 years ago
Updated 5 days ago
Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), and
36 more.

unsloth by unslothai

0.6%
46k
Finetuning tool for LLMs, targeting speed and memory efficiency
Created 1 year ago
Updated 14 hours ago
Feedback? Help us improve.