LLM fine-tuning toolkit for research
Top 10.8% on sourcepulse
XTuner is a comprehensive toolkit for fine-tuning large language models (LLMs) and visual-language models (VLMs), designed for efficiency and flexibility. It supports a wide array of models including InternLM, Llama, Mistral, Qwen, and Phi, catering to researchers and developers needing to adapt these models for specific tasks. XTuner enables efficient fine-tuning techniques like QLoRA and full-parameter tuning, even on limited hardware, and integrates with popular distributed training frameworks like DeepSpeed.
How It Works
XTuner leverages optimized kernels (FlashAttention, Triton) and DeepSpeed integration for high-throughput training. Its architecture supports various fine-tuning methods (QLoRA, LoRA, full-parameter) and data processing pipelines, allowing users to customize training from continuous pre-training to instruction and agent fine-tuning. It also facilitates multi-modal VLM pre-training and fine-tuning using architectures like LLaVA.
Quick Start & Requirements
pip install -U xtuner
or pip install -U 'xtuner[deepspeed]'
. Source install: git clone https://github.com/InternLM/xtuner.git && cd xtuner && pip install -e '.[all]'
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project's rapid development pace means new models and features are frequently added, potentially leading to breaking changes or requiring frequent updates to dependencies. Specific hardware requirements may vary significantly based on the model size and fine-tuning method employed.
3 weeks ago
1 day