ChatLearn  by alibaba

Training framework for large-scale alignment tasks

Created 2 years ago
425 stars

Top 69.4% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

ChatLearn is a flexible and efficient framework for large-scale alignment training of language models, targeting researchers and practitioners. It simplifies the process of implementing alignment techniques like RLHF and DPO, offering significant performance improvements and scalability for complex model configurations.

How It Works

ChatLearn provides a user-friendly interface that abstracts away complex distributed execution, resource scheduling, and data flow management. It supports diverse alignment algorithms (RLHF, DPO, OnlineDPO, GRPO) and allows users to define custom training flows. A key advantage is its support for multiple distributed acceleration backends, including Megatron-LM, DeepSpeed, and vLLM, enabling flexible choices for training and inference acceleration. It also features advanced parallel strategy configuration and efficient GPU memory sharing (EMS) for optimized resource utilization.

Quick Start & Requirements

  • Installation and quick start instructions are available in the official documentation.
  • Prerequisites include Python and potentially specific libraries depending on the chosen backend (e.g., Megatron-LM, DeepSpeed, vLLM).
  • Setup time and resource requirements will vary based on model size and chosen backends.

Highlighted Details

  • Achieves up to 208% speedup in RLHF training throughput compared to optimized DeepSpeed-Chat and OpenRLHF at 70B+70B scale.
  • Supports large-scale alignment training up to 300B+300B parameter configurations.
  • Integrates with Megatron-LM for training/inference and vLLM for inference acceleration.
  • Features Efficient Memory Sharing (EMS) for inter-model memory optimization.

Maintenance & Community

  • Developed by Alibaba Cloud PAI platform.
  • Open to hiring and collaborations; contact wanglin.zj@alibaba-inc.com.
  • Community discussion via DingTalk group: 98090003312.

Licensing & Compatibility

  • The README does not explicitly state the license. Further investigation is required for commercial use or closed-source linking.

Limitations & Caveats

  • The project is newly released (August 2024), and its long-term maintenance and community adoption are yet to be established.
  • The license is not specified, which may pose a barrier for commercial adoption.
Health Check
Last Commit

13 hours ago

Responsiveness

1 day

Pull Requests (30d)
33
Issues (30d)
2
Star History
15 stars in the last 30 days

Explore Similar Projects

Starred by Yaowei Zheng Yaowei Zheng(Author of LLaMA-Factory), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
1 more.

VeOmni by ByteDance-Seed

3.4%
1k
Framework for scaling multimodal model training across accelerators
Created 5 months ago
Updated 3 weeks ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Pawel Garbacki Pawel Garbacki(Cofounder of Fireworks AI), and
11 more.

Liger-Kernel by linkedin

0.6%
6k
Triton kernels for efficient LLM training
Created 1 year ago
Updated 1 day ago
Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), Stas Bekman Stas Bekman(Author of "Machine Learning Engineering Open Book"; Research Engineer at Snowflake), and
25 more.

gpt-neox by EleutherAI

0.2%
7k
Framework for training large-scale autoregressive language models
Created 4 years ago
Updated 2 days ago
Starred by Clement Delangue Clement Delangue(Cofounder of Hugging Face), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
20 more.

accelerate by huggingface

0.3%
9k
PyTorch training helper for distributed execution
Created 4 years ago
Updated 1 day ago
Feedback? Help us improve.