Adapter tuning research for visual recognition tasks
Top 81.6% on sourcepulse
Mona is a novel adapter-based tuning method designed to surpass the performance of full fine-tuning in visual recognition tasks, particularly for challenging areas like instance and semantic segmentation. It targets researchers and practitioners in computer vision seeking efficient and high-performing alternatives to traditional fine-tuning.
How It Works
Mona, or Multi-cognitive Visual Adapter, is a parameter-efficient tuning method that integrates multiple adapter modules. This approach is designed to enhance transfer learning efficiency and performance by offering a competitive alternative to full fine-tuning, aiming to break previous performance ceilings in delta-tuning methods.
Quick Start & Requirements
dist_train.sh
scripts with modified configuration files specifying data_root
and load_from
paths.Highlighted Details
Maintenance & Community
The project is associated with the CVPR 2025 conference. Key contributors include Dongshuo Yin, Leiyi Hu, and Bin Li. It acknowledges contributions from Swin-Transformer, mmclassification, NOAH, LoRA, and Adaptformer.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is presented as a CVPR 2025 submission, suggesting it may be in a research or early-stage development phase. Specific limitations or unsupported platforms are not detailed in the README.
1 month ago
1 week