ChatLearn  by alibaba

Training framework for large-scale alignment tasks

created 1 year ago
393 stars

Top 74.3% on sourcepulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

ChatLearn is a flexible and efficient framework for large-scale alignment training of language models, targeting researchers and practitioners. It simplifies the process of implementing alignment techniques like RLHF and DPO, offering significant performance improvements and scalability for complex model configurations.

How It Works

ChatLearn provides a user-friendly interface that abstracts away complex distributed execution, resource scheduling, and data flow management. It supports diverse alignment algorithms (RLHF, DPO, OnlineDPO, GRPO) and allows users to define custom training flows. A key advantage is its support for multiple distributed acceleration backends, including Megatron-LM, DeepSpeed, and vLLM, enabling flexible choices for training and inference acceleration. It also features advanced parallel strategy configuration and efficient GPU memory sharing (EMS) for optimized resource utilization.

Quick Start & Requirements

  • Installation and quick start instructions are available in the official documentation.
  • Prerequisites include Python and potentially specific libraries depending on the chosen backend (e.g., Megatron-LM, DeepSpeed, vLLM).
  • Setup time and resource requirements will vary based on model size and chosen backends.

Highlighted Details

  • Achieves up to 208% speedup in RLHF training throughput compared to optimized DeepSpeed-Chat and OpenRLHF at 70B+70B scale.
  • Supports large-scale alignment training up to 300B+300B parameter configurations.
  • Integrates with Megatron-LM for training/inference and vLLM for inference acceleration.
  • Features Efficient Memory Sharing (EMS) for inter-model memory optimization.

Maintenance & Community

  • Developed by Alibaba Cloud PAI platform.
  • Open to hiring and collaborations; contact wanglin.zj@alibaba-inc.com.
  • Community discussion via DingTalk group: 98090003312.

Licensing & Compatibility

  • The README does not explicitly state the license. Further investigation is required for commercial use or closed-source linking.

Limitations & Caveats

  • The project is newly released (August 2024), and its long-term maintenance and community adoption are yet to be established.
  • The license is not specified, which may pose a barrier for commercial adoption.
Health Check
Last commit

1 day ago

Responsiveness

1 day

Pull Requests (30d)
25
Issues (30d)
0
Star History
54 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.