Discover and explore top open-source AI tools and projects—updated daily.
NVlabsFP8 training framework for memory efficiency
Top 100.0% on SourcePulse
This project addresses the memory bottleneck in training large AI models by introducing COAT, a novel FP8 quantization method for optimizer states and activations. It targets researchers and engineers seeking to train larger models on limited hardware, offering significant memory reduction and speedup while maintaining accuracy.
How It Works
COAT enhances FP8 training efficiency through two core innovations. First, Dynamic Range Expansion is applied to optimizer states, dynamically adjusting their representation range to better align with FP8 capabilities, thereby minimizing quantization errors. Second, Mixed-Granularity Activation Quantization optimizes activation memory by employing per-tensor quantization for linear layers and more granular strategies (like VS-Quant) for non-linear layers. This dual approach effectively reduces memory footprint and accelerates training without compromising model performance.
Quick Start & Requirements
Installation is available via pip (pip install fp8-coat) or from source by cloning the repository and running the provided environment_setup.sh script to create a Conda environment. COAT supports Llama 2/3 models and integrates seamlessly with Hugging Face's transformers Trainer.
Highlighted Details
Maintenance & Community
The project includes a "To-Do List" indicating ongoing development, with plans for TorchTitan and FSDP2 support. No explicit community channels (e.g., Discord, Slack) or detailed contributor information are prominently featured in the README.
Licensing & Compatibility
The repository's README does not explicitly state a software license. This omission requires further investigation for commercial use or integration into closed-source projects.
Limitations & Caveats
Ongoing development means certain features, such as support for TorchTitan and FSDP2, are still under development. The absence of a clearly defined license is a significant caveat for adoption.
3 months ago
Inactive
microsoft
ELS-RD
SJTU-IPADS
huggingface
unslothai