Finetune_LLMs  by mallorbc

LLM finetuning code using DeepSpeed, LoRA, or QLoRA

Created 4 years ago
457 stars

Top 66.2% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides code for fine-tuning Large Language Models (LLMs) using a curated quotes dataset. It targets users familiar with LLM fine-tuning who need a streamlined process, offering DeepSpeed, LoRA, and QLoRA methods.

How It Works

The project leverages Hugging Face's transformers library and integrates DeepSpeed for distributed training, enabling efficient fine-tuning of large models. LoRA and QLoRA are included for parameter-efficient fine-tuning, reducing memory requirements and training time. The code is adapted from an existing repository to support a wider range of models and methods.

Quick Start & Requirements

  • Install/Run: Use the provided build_image.sh and run_image.sh scripts for Docker-based setup.
  • Prerequisites:
    • NVIDIA GPU (>= 24GB VRAM, fp16 support recommended; A100 suggested).
    • Linux OS (Ubuntu recommended).
    • Modern Docker and NVIDIA Docker.
    • Latest NVIDIA drivers and CUDA drivers.
  • Setup: Building the Docker image can take several minutes.
  • Docs: Original Repo, Huggingface Model Walkthrough

Highlighted Details

  • Supports DeepSpeed, LoRA, and QLoRA fine-tuning methods.
  • Includes a formatted dataset (quotes_dataset) for fine-tuning.
  • Dockerized workflow simplifies setup and dependency management.
  • Original code adapted for broader model and method compatibility.

Maintenance & Community

  • The repository mentions paid professional assistance is available via email.
  • Links to older video walkthroughs are provided, but users are advised to use modern methods.

Licensing & Compatibility

  • The repository does not explicitly state a license.

Limitations & Caveats

The project's Docker images may require updates for newer CUDA versions if build errors occur. Older video walkthroughs are not recommended for current usage.

Health Check
Last Commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
1
Issues (30d)
0
Star History
1 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Junyang Lin Junyang Lin(Core Maintainer at Alibaba Qwen), and
3 more.

Alpaca-CoT by PhoebusSi

0.1%
3k
IFT platform for instruction collection, parameter-efficient methods, and LLMs
Created 2 years ago
Updated 1 year ago
Starred by Casper Hansen Casper Hansen(Author of AutoAWQ), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
5 more.

xtuner by InternLM

0.5%
5k
LLM fine-tuning toolkit for research
Created 2 years ago
Updated 1 day ago
Starred by Junyang Lin Junyang Lin(Core Maintainer at Alibaba Qwen), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
25 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.