Paper list for LLM knowledge distillation
Top 35.0% on sourcepulse
This repository is a curated collection of research papers on Knowledge Distillation (KD) for Large Language Models (LLMs), aimed at researchers and practitioners seeking to transfer capabilities from large proprietary models to smaller ones or enable self-improvement. It provides a structured overview of KD techniques, categorized by algorithms, skill transfer, and domain-specific applications, serving as a comprehensive resource for understanding and implementing LLM distillation.
How It Works
The collection is organized around a taxonomy that breaks down KD into "Knowledge Elicitation" (extracting knowledge from teacher LLMs) and "Distillation Algorithms" (transferring knowledge to student models). It further explores "Skill Distillation" for enhancing specific cognitive abilities (e.g., reasoning, alignment) and "Verticalization Distillation" for domain-specific applications (e.g., law, medicine). This structured approach allows users to navigate the diverse landscape of KD research efficiently.
Quick Start & Requirements
This repository is a curated list of papers and does not involve direct code execution or installation. It serves as a reference guide.
Highlighted Details
Maintenance & Community
The repository is maintained by Xiaohan Xu and collaborators, with contact information provided for contributions and feedback. Users are encouraged to open issues/PRs or email to suggest missing papers or taxonomies.
Licensing & Compatibility
The repository itself is not licensed for software use. The linked papers have their own respective licenses and terms of use, which users must adhere to.
Limitations & Caveats
The collection primarily focuses on generative LLMs and explicitly notes that encoder-based KD is not included, though it is being tracked. Some entries may lack direct code links.
4 months ago
1 week