Framework for efficient LoRA fine-tuning of multiple LLMs
Top 84.0% on sourcepulse
mLoRA is an open-source framework for efficiently fine-tuning multiple Large Language Models (LLMs) using LoRA and its variants. It targets researchers and practitioners needing to train numerous adapters concurrently, offering significant computational and memory savings by sharing a base model and employing pipeline parallelism.
How It Works
mLoRA utilizes a pipeline parallelism algorithm to distribute the fine-tuning process across multiple GPUs. This approach allows for concurrent training of multiple LoRA adapters, each sharing a common base LLM. The framework supports various LoRA variants (QLoRA, LoRA+, VeRA, DoRA) and preference alignment algorithms (DPO, CPO, CIT), enabling flexible and efficient multi-adapter fine-tuning.
Quick Start & Requirements
pip install .
(Python >= 3.12) or use the official Docker image (yezhengmaolove/mlora:latest
).python mlora_train.py --base_model <model_name> --config <config_path>
demo/
folder.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project mentions "Document Help Document[TODO]" and "How to develop a new adapter", indicating that some documentation may be incomplete or under development.
5 months ago
1 day