Parameter-efficient fine-tuning via high-rank updating (MoRA)
Top 79.2% on sourcepulse
MoRA (Modular Rank-wise Adaptation) is a parameter-efficient fine-tuning (PEFT) technique designed to enhance the efficiency and effectiveness of adapting large language models. It targets researchers and practitioners working with LLMs who need to fine-tune models with fewer parameters and computational resources, offering a more flexible and potentially superior alternative to standard LoRA.
How It Works
MoRA decomposes the low-rank update matrices into multiple smaller, low-rank matrices, allowing for a more granular control over the adaptation process. It introduces two types of updates: Type 1 shares parameters across multiple low-rank matrices for large ranks, reducing parameter count, while Type 6 uses a novel RoPE-based approach for smaller ranks, offering improved performance. This modularity allows for a trade-off between parameter efficiency and expressiveness.
Quick Start & Requirements
pip install -e ./peft-mora
.peft
library.deepspeed
.bf16
and 16bit
precision.Highlighted Details
peft
library.merge_and_unload()
.deepspeed
for distributed training.Maintenance & Community
The project is based on popular libraries like peft
, alpaca-lora
, and ReLoRA
. Further community engagement details (Discord, Slack, roadmap) are not explicitly provided in the README.
Licensing & Compatibility
The README does not explicitly state the license. Given its reliance on Hugging Face peft
and alpaca-lora
, it is likely compatible with common open-source licenses, but explicit verification is recommended for commercial use.
Limitations & Caveats
The project is presented as an implementation within peft-mora
, suggesting it might be an experimental or research-oriented extension rather than a fully integrated feature of the main peft
library. The specific performance benefits and stability compared to standard LoRA or other PEFT methods would require further evaluation.
1 year ago
1 day