h2o-llmstudio  by h2oai

LLM Studio: framework for LLM fine-tuning via GUI or CLI

created 2 years ago
4,362 stars

Top 11.4% on sourcepulse

GitHubView on GitHub
Project Summary

H2O LLM Studio provides a no-code GUI and framework for fine-tuning large language models, targeting users who want to customize LLMs without extensive coding. It simplifies the process of adapting state-of-the-art models using various hyperparameters and modern techniques like LoRA and 8-bit training, enabling efficient model customization and evaluation.

How It Works

The framework supports fine-tuning LLMs using techniques such as Low-Rank Adaptation (LoRA) and 8-bit quantization for reduced memory footprint. It also incorporates Reinforcement Learning (RL) alternatives like Direct Preference Optimization (DPO), Identity Preference Optimization (IPO), and KTO for model tuning based on preference data. The system offers advanced evaluation metrics, visual performance tracking, and integrations with Neptune and Weights & Biases.

Quick Start & Requirements

  • Install: Recommended: pipenv install via make setup. Alternative: make setup-no-flash (disables Flash Attention 2). Docker installation is also available.
  • Prerequisites: Ubuntu 16.04+, NVIDIA GPU (>= 470.57.02 drivers), Python 3.10 recommended. For larger models, 24GB+ GPU memory is advised. CUDA 12.1 or 12.4 is recommended for DeepSpeed.
  • Resources: Setup involves installing Python, drivers, and dependencies. Docker requires NVIDIA Container Toolkit.
  • Links: Documentation: https://docs.h2o.ai/h2o-llmstudio/

Highlighted Details

  • Supports Causal Regression and Classification modeling.
  • Integrates DeepSpeed for distributed training on multi-GPU setups (requires NVLink).
  • Offers CLI for fine-tuning, interactive chat, and Hugging Face Hub publishing.
  • Recent updates include DPO/IPO/KTO optimization and removal of RLHF.

Maintenance & Community

  • Active development with recent PRs addressing new features and deprecations.
  • Discord server available for community discussion.
  • Model checkpoints and datasets are available on H2O.ai's Hugging Face page.

Licensing & Compatibility

  • Licensed under Apache 2.0.
  • Compatible with commercial use and closed-source linking.

Limitations & Caveats

The project notes that due to rapid development, full backward compatibility is not guaranteed, and users are advised to pin versions. RLHF is being deprecated and will be fully removed in a future release.

Health Check
Last commit

1 day ago

Responsiveness

Inactive

Pull Requests (30d)
13
Issues (30d)
2
Star History
92 stars in the last 90 days

Explore Similar Projects

Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Jiayi Pan Jiayi Pan(Author of SWE-Gym; AI Researcher at UC Berkeley), and
5 more.

Liger-Kernel by linkedin

0.6%
5k
Triton kernels for efficient LLM training
created 1 year ago
updated 1 day ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Omar Sanseviero Omar Sanseviero(DevRel at Google DeepMind), and
5 more.

TensorRT-LLM by NVIDIA

0.6%
11k
LLM inference optimization SDK for NVIDIA GPUs
created 1 year ago
updated 14 hours ago
Feedback? Help us improve.