Discover and explore top open-source AI tools and projects—updated daily.
Neural network compression research toolkit
Top 11.1% on SourcePulse
Summary
IntelLabs/distiller is a Python package designed for neural network compression research, offering tools for sparsity, quantization, and knowledge distillation. It targets researchers and engineers seeking to reduce model size, improve inference speed, and lower energy consumption in deep learning models. The package provides a PyTorch environment for prototyping and analyzing various compression algorithms.
How It Works
Distiller facilitates network compression through a flexible PyTorch framework. It supports diverse techniques including element-wise and structured weight pruning (e.g., kernel-wise, filter-wise, channel-wise), automatic model compression (AMC) via sensitivity analysis, and various quantization methods (post-training, quantization-aware) with customizable bit-widths. The library also integrates knowledge distillation and allows for flexible scheduling of compression tasks, with configurations defined in YAML files.
Quick Start & Requirements
git clone https://github.com/IntelLabs/distiller.git
), create and activate a Python virtual environment (python3 -m venv env
, source env/bin/activate
), then install in development mode (cd distiller
, pip3 install -e .
).Highlighted Details
Maintenance & Community
This project will no longer be maintained by Intel and has been identified as having known security escapes. Intel has ceased all development, maintenance, bug fixes, and contributions. The project is effectively discontinued.
Licensing & Compatibility
Limitations & Caveats
The primary limitation is the project's discontinuation by Intel due to identified security escapes, rendering it unsupported and potentially unsafe. Furthermore, the tested environment is outdated, and users may face compatibility issues with modern hardware and software stacks.
2 years ago
Inactive