CoLLiE  by OpenMOSS

LLM training toolkit for efficient collaborative tuning

Created 2 years ago
416 stars

Top 70.6% on SourcePulse

GitHubView on GitHub
Project Summary

CoLLiE is a comprehensive toolkit for training large language models (LLMs) from scratch, designed for researchers and practitioners. It streamlines the entire LLM training pipeline, from data preprocessing and fine-tuning to model saving and metric monitoring, aiming to accelerate training, improve quality, and reduce costs.

How It Works

CoLLiE builds upon DeepSpeed and PyTorch, integrating advanced parallelization strategies (DP, PP, TP, ZeRO) with efficient fine-tuning methods like LOMO and LoRA, and Flash Attention. This combination allows for collaborative and efficient LLM tuning, offering a user-friendly interface with highly customizable options for both beginners and experienced users.

Quick Start & Requirements

  • Install: pip install collie-lm
  • Prerequisites: PyTorch >= 1.13, CUDA >= 11.6, Linux OS.
  • Setup: Installation is straightforward via pip. The README provides a detailed example for training the MOSS model using LOMO and ZeRO-3, requiring torchrun for distributed training.
  • Docs: https://github.com/OpenMOSS/CoLLiE (Examples and tutorials are linked within the README).

Highlighted Details

  • Supports major LLM architectures including MOSS, InternLM, LLaMA, and ChatGLM.
  • Integrates efficient techniques like LOMO, LoRA, and Flash Attention.
  • Offers robust monitoring tools for step time, token generation speed, memory usage, and loss.
  • Includes evaluators for perplexity and generation metrics.

Maintenance & Community

The project has been accepted into EMNLP System Demonstrations (Dec 2023). Community links are not explicitly provided in the README, but a "Community" section is present.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

The README mentions "zero_allow_untested_optimizer" in the DeepSpeed configuration, suggesting potential instability with certain optimizer configurations. Specific hardware benchmarks are provided, but general performance claims are not quantified across all supported models and configurations.

Health Check
Last Commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
1 stars in the last 30 days

Explore Similar Projects

Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
6 more.

xTuring by stochasticai

0.0%
3k
SDK for fine-tuning and customizing open-source LLMs
Created 2 years ago
Updated 1 day ago
Starred by Théophile Gervet Théophile Gervet(Cofounder of Genesis AI), Jason Knight Jason Knight(Director AI Compilers at NVIDIA; Cofounder of OctoML), and
6 more.

lingua by facebookresearch

0.1%
5k
LLM research codebase for training and inference
Created 11 months ago
Updated 2 months ago
Starred by Casper Hansen Casper Hansen(Author of AutoAWQ), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
5 more.

xtuner by InternLM

0.5%
5k
LLM fine-tuning toolkit for research
Created 2 years ago
Updated 1 day ago
Feedback? Help us improve.