MFT framework for code LLM multitask fine-tuning (KDD 2024 paper)
Top 49.8% on sourcepulse
MFTCoder is a framework for high-accuracy, efficient multi-task fine-tuning of Large Language Models (LLMs), particularly for code-related tasks. It targets researchers and developers working with code LLMs, offering a robust solution for improving model performance across diverse coding challenges through its multi-task fine-tuning approach.
How It Works
MFTCoder employs a multi-task fine-tuning (MFT) strategy to train LLMs on multiple datasets simultaneously, aiming for improved generalization and performance on unseen tasks. It supports various efficient fine-tuning techniques like LoRA, QLoRA, and full-parameter training, enabling the fine-tuning of large models with reduced computational resources. The framework also incorporates methods like Self-paced Loss and the newer Convergence Balancer (CoBa) to manage convergence across different tasks, ensuring balanced training outcomes.
Quick Start & Requirements
bash sh init_env.sh
.Highlighted Details
Maintenance & Community
The project is actively developed by the AI Native team at Ant Group, with recent updates in October 2024 introducing preference alignment methods (DPO/RPO/ORPO) and new convergence balancing techniques. The team has a strong publication record at top AI conferences.
Licensing & Compatibility
The project appears to be open-source, with models available on HuggingFace and ModelScope. Specific licensing details for the framework and models are not explicitly stated in the README, but compatibility for commercial use would require verification.
Limitations & Caveats
While the framework supports various models and fine-tuning methods, specific details on the limitations of each supported model or the exact resource requirements for full-parameter training are not elaborated. The project is actively evolving, with new features and improvements being released regularly.
7 months ago
1+ week