Discover and explore top open-source AI tools and projects—updated daily.
TencentModel compression toolkit for efficient AI
Top 65.5% on SourcePulse
Model compression is addressed by AngelSlim, a toolkit engineered for enhanced usability, comprehensiveness, and efficiency, targeting engineers and researchers working with large AI models. It provides a unified framework for applying various compression techniques, enabling more accessible and performant model deployment. The toolkit aims to streamline the model compression workflow, making advanced techniques readily available.
How It Works
The toolkit integrates mainstream compression algorithms, including quantization (e.g., FP8, INT4, NVFP4, Tequila) and speculative decoding (Eagle3), into a unified, user-friendly framework. It focuses on performance optimization across the end-to-end compression workflow, from training to deployment. AngelSlim continuously researches and incorporates novel compression algorithms, offering a path to significantly reduce model size and inference costs while maintaining accuracy.
Quick Start & Requirements
pip install angelslim or clone and python setup.py install.Highlighted Details
Maintenance & Community
The project shows active development with frequent releases (e.g., v0.3, v0.2) and ongoing additions of new models and algorithms. Community engagement is facilitated through WeChat, Discord, and GitHub Issues for discussions and support.
Licensing & Compatibility
The code is stated to be open-sourced under "License for AngelSlim." The specific terms and compatibility for commercial use or closed-source linking are not detailed, requiring further clarification.
Limitations & Caveats
Some advanced features, such as token pruning for VLMs and audio models, are listed as "Under Development." The absence of a clearly defined, standard open-source license may pose adoption challenges for certain use cases.
23 hours ago
Inactive
SafeAILab
NVIDIA
mit-han-lab
intel
vllm-project