LLM finetuning code using DeepSpeed, LoRA, or QLoRA
Top 67.1% on sourcepulse
This repository provides code for fine-tuning Large Language Models (LLMs) using a curated quotes dataset. It targets users familiar with LLM fine-tuning who need a streamlined process, offering DeepSpeed, LoRA, and QLoRA methods.
How It Works
The project leverages Hugging Face's transformers
library and integrates DeepSpeed for distributed training, enabling efficient fine-tuning of large models. LoRA and QLoRA are included for parameter-efficient fine-tuning, reducing memory requirements and training time. The code is adapted from an existing repository to support a wider range of models and methods.
Quick Start & Requirements
build_image.sh
and run_image.sh
scripts for Docker-based setup.Highlighted Details
quotes_dataset
) for fine-tuning.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project's Docker images may require updates for newer CUDA versions if build errors occur. Older video walkthroughs are not recommended for current usage.
1 year ago
1 week