LLM training toolkit for efficient collaborative tuning
Top 71.3% on sourcepulse
CoLLiE is a comprehensive toolkit for training large language models (LLMs) from scratch, designed for researchers and practitioners. It streamlines the entire LLM training pipeline, from data preprocessing and fine-tuning to model saving and metric monitoring, aiming to accelerate training, improve quality, and reduce costs.
How It Works
CoLLiE builds upon DeepSpeed and PyTorch, integrating advanced parallelization strategies (DP, PP, TP, ZeRO) with efficient fine-tuning methods like LOMO and LoRA, and Flash Attention. This combination allows for collaborative and efficient LLM tuning, offering a user-friendly interface with highly customizable options for both beginners and experienced users.
Quick Start & Requirements
pip install collie-lm
torchrun
for distributed training.Highlighted Details
Maintenance & Community
The project has been accepted into EMNLP System Demonstrations (Dec 2023). Community links are not explicitly provided in the README, but a "Community" section is present.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The README mentions "zero_allow_untested_optimizer" in the DeepSpeed configuration, suggesting potential instability with certain optimizer configurations. Specific hardware benchmarks are provided, but general performance claims are not quantified across all supported models and configurations.
11 months ago
1 day