PyTorch package for parameter-efficient fine-tuning via adaptive budget allocation
Top 82.8% on sourcepulse
AdaLoRA provides a parameter-efficient fine-tuning method that adaptively allocates a budget of trainable parameters across layers. It targets researchers and practitioners seeking to reduce the computational cost and memory footprint of fine-tuning large language models, enabling efficient adaptation of models like DeBERTa and BART.
How It Works
AdaLoRA employs Singular Value Decomposition (SVD) to decompose weight matrices into smaller, trainable low-rank matrices. Its core innovation is the RankAllocator
, which dynamically adjusts the rank (number of singular values) for each layer based on its importance during training. This adaptive budget allocation, combined with orthogonality regularization, aims to optimize parameter efficiency and performance.
Quick Start & Requirements
pip install -e loralib/
.nn.Linear
with loralib.SVDLinear
and using loralib.RankAllocator
within the training loop.NLU/
and NLG_QA/
directories.Highlighted Details
loralib/adalora.py
.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 years ago
1 week