Discover and explore top open-source AI tools and projects—updated daily.
ai-dynamoInference toolkit for optimizing PyTorch models on NVIDIA GPUs
Top 97.4% on SourcePulse
Summary
NVIDIA AITune is an inference toolkit designed for tuning and deploying deep learning models, optimized for NVIDIA GPUs. It addresses the need for significantly improved inference speed and efficiency across diverse AI workloads by automating the compilation and conversion of PyTorch models and pipelines. Using a unified Python API, it enables seamless tuning with backends like TensorRT, Torch-TensorRT, TorchAO, and Torch Inductor, preparing models for production with minimal code changes.
How It Works
AITune optimizes PyTorch models at the nn.Module level via two modes: Ahead-of-Time (AOT) for greater control and Just-in-Time (JIT) for zero code modification. AOT requires explicit code wrapping, while JIT can be enabled via an environment variable or import for on-the-fly tuning. The system supports multiple backends (TensorRT, Torch-TensorRT, TorchAO, Torch Inductor) and employs strategies like FirstWinsStrategy or HighestThroughputStrategy to automatically select optimal backend configurations based on performance metrics. It automates model export, conversion, correctness testing, and profiling.
Quick Start & Requirements
pip install --extra-index-url https://pypi.nvidia.com aitune. Source install supported.Highlighted Details
AUTOWRAPT_BOOTSTRAP=aitune_enable_jit_tuning or import aitune.torch.jit.enable.export NVTX_ENABLE=1).Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 month ago
Inactive
microsoft
cfregly
Lightning-AI
mindspore-ai
openvinotoolkit