Fine-tuning tool for LLMs using QLoRA
Top 98.2% on sourcepulse
This repository provides a streamlined approach to fine-tuning large language models (LLMs) using the QLoRA technique. It targets AI researchers and developers looking to efficiently adapt pre-trained LLMs for specific tasks or datasets, enabling custom model creation with reduced computational resources.
How It Works
The project leverages QLoRA, a parameter-efficient fine-tuning method that quantizes pre-trained LLMs to 4-bit precision and uses Low-Rank Adaptation (LoRA) to fine-tune only a small subset of adapter weights. This significantly reduces memory requirements and computational cost during training while maintaining performance comparable to full fine-tuning.
Quick Start & Requirements
pip install -r requirements.txt
python train.py <config_file>
(e.g., python train.py configs/llama3_8b_chat_uncensored.yaml
)Highlighted Details
llama.cpp
.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Users on Python 3.7 may encounter compatibility issues with newer transformers
and peft
versions, requiring specific older package versions for stable operation.
1 year ago
1 day