Discover and explore top open-source AI tools and projects—updated daily.
georgesungFine-tuning tool for LLMs using QLoRA
Top 96.2% on SourcePulse
This repository provides a streamlined approach to fine-tuning large language models (LLMs) using the QLoRA technique. It targets AI researchers and developers looking to efficiently adapt pre-trained LLMs for specific tasks or datasets, enabling custom model creation with reduced computational resources.
How It Works
The project leverages QLoRA, a parameter-efficient fine-tuning method that quantizes pre-trained LLMs to 4-bit precision and uses Low-Rank Adaptation (LoRA) to fine-tune only a small subset of adapter weights. This significantly reduces memory requirements and computational cost during training while maintaining performance comparable to full fine-tuning.
Quick Start & Requirements
pip install -r requirements.txtpython train.py <config_file> (e.g., python train.py configs/llama3_8b_chat_uncensored.yaml)Highlighted Details
llama.cpp.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Users on Python 3.7 may encounter compatibility issues with newer transformers and peft versions, requiring specific older package versions for stable operation.
1 year ago
Inactive
huggingface
InternLM
tloen
tatsu-lab
meta-llama