easydistill  by modelscope

Toolkit for efficient knowledge distillation of large language models

Created 8 months ago
256 stars

Top 98.7% on SourcePulse

GitHubView on GitHub
Project Summary

<2-3 sentences summarising what the project addresses and solves, the target audience, and the benefit.> EasyDistill is a pioneering toolkit for knowledge distillation (KD) of large language models (LLMs), enabling smaller models to emulate larger ones efficiently. It targets NLP researchers and practitioners, offering a versatile, user-friendly platform to streamline KD, support diverse methodologies, and facilitate practical industrial solutions.

How It Works

The toolkit supports black-box and white-box KD, featuring data synthesis, SFT, ranking optimization, and RL. It accommodates System 1 (fast, intuitive) and System 2 (slow, analytical) cognitive models via a modular architecture and simple CLI, facilitating experimentation and industrial integration with platforms like Alibaba Cloud PAI.

Quick Start & Requirements

Clone the repo (git clone https://github.com/modelscope/easydistill), navigate (cd EasyDistill), and install with python setup.py install. Run jobs via easydistill --config <config-file-path>. Specific hardware requirements like GPU memory are configurable within job settings.

Highlighted Details

  • DistilQwen Series: Distilled LLMs from Qwen models, offering reduced size with high performance for resource-constrained environments.
  • Adaptive Thinking Models: DistilQwen-ThoughtX/Y models feature improved reasoning, using the OmniThought dataset with RV/CD scores for optimal CoT generation.
  • System 1 & 2 Models: DistilQwen2/2.5 enhance instruction following. DistilQwen2.5-R1/DS3 focus on reasoning, using DeepSeek-R1 as a teacher and employing CogPO/CoT simplification.
  • Public Datasets: Includes released instruction-following (DistilQwen_100K/1M) and Chain-of-Thought (OmniThought, OmniThought-0528) datasets, annotated with RV/CD scores.

Maintenance & Community

The project actively releases new models and features, with recent updates including OmniThoughtV and multi-modal KD. Community discussions are welcomed via a DingTalk group (117440002081). Integration with Alibaba Cloud PAI is supported.

Licensing & Compatibility

The primary license is Apache License 2.0, permissive for commercial use. However, some included code may originate from other repositories under different licenses; consult the NOTICE file for specifics.

Limitations & Caveats

The README does not explicitly detail known limitations or bugs. Users should review the NOTICE file for potential licensing complexities from incorporated third-party code.

Health Check
Last Commit

1 month ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
1
Star History
27 stars in the last 30 days

Explore Similar Projects

Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), Eric Zhu Eric Zhu(Coauthor of AutoGen; Research Scientist at Microsoft Research), and
8 more.

tinker-cookbook by thinking-machines-lab

0.5%
3k
Advanced LLM fine-tuning SDK and example cookbook
Created 6 months ago
Updated 2 days ago
Starred by Stas Bekman Stas Bekman(Author of "Machine Learning Engineering Open Book"; Research Engineer at Snowflake), Edward Sun Edward Sun(Research Scientist at Meta Superintelligence Lab), and
1 more.

awesome-knowledge-distillation by dkozlov

0.2%
4k
Collection of knowledge distillation resources
Created 9 years ago
Updated 1 month ago
Starred by Michael Han Michael Han(Cofounder of Unsloth), Sebastian Raschka Sebastian Raschka(Author of "Build a Large Language Model (From Scratch)"), and
19 more.

DeepSeek-R1 by deepseek-ai

0.1%
92k
Reasoning models research paper
Created 1 year ago
Updated 7 months ago
Feedback? Help us improve.