LLM training tool for Qwen2.5, Llama3, Gemma, and other models
Top 8.0% on sourcepulse
Firefly is a comprehensive, one-stop training tool for large language models, designed for researchers and developers looking to pre-train, fine-tune, and apply Direct Preference Optimization (DPO) to a wide array of popular open-source LLMs. It offers efficient training methods like LoRA and QLoRA, significantly reducing resource requirements and accelerating the fine-tuning process, making advanced LLM customization accessible even with limited hardware.
How It Works
Firefly supports full parameter training, LoRA, and QLoRA, enabling users to choose the most suitable method based on their available resources. It integrates with Unsloth for accelerated training and reduced memory usage, boasting up to 47.32% speedup and 39.13% memory savings. The tool aligns fine-tuning templates with various official chat models, ensuring compatibility and consistent performance across a broad spectrum of LLMs.
Quick Start & Requirements
pip install requirements.txt
pip install git+https://github.com/unslothai/unsloth.git
and other specified dependencies.bitsandbytes
, peft
, xformers
. CUDA is implicitly required for GPU acceleration.Highlighted Details
firefly-mixtral-8x7b
scoring 70.16.firefly-train-1.1M
).Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Open-sourced models may produce factual inaccuracies, biased, or unethical content due to a lack of specific safety fine-tuning. Capabilities in coding and reasoning are noted as areas for improvement. Use is restricted to non-harmful purposes.
9 months ago
1 week