ort  by pytorch

PyTorch extension for ONNX Runtime model acceleration

Created 4 years ago
364 stars

Top 77.2% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

This library accelerates PyTorch model training and inference using ONNX Runtime and Intel® OpenVINO™. It targets PyTorch developers seeking to reduce training time, scale large models, and optimize inference performance, particularly on Intel hardware.

How It Works

The library provides ORTModule for training, which converts PyTorch models to ONNX format and leverages ONNX Runtime for accelerated execution. It also includes optimized optimizers like FusedAdam and FP16_Optimizer, and a LoadBalancingDistributedSampler for efficient data loading in distributed training. For inference, ORTInferenceModule enables ONNX Runtime with the OpenVINO™ Execution Provider, targeting Intel CPUs, GPUs, and VPUs with options for precision (FP32/FP16) and backend.

Quick Start & Requirements

Highlighted Details

  • Reduces PyTorch training time and GPU cost for large transformer models.
  • Supports FusedAdam, FP16_Optimizer, and LoadBalancingDistributedSampler for training efficiency.
  • Inference acceleration via OpenVINO™ Execution Provider on Intel hardware (CPU, GPU, VPU).
  • Offers Mixture of Experts (MoE) layer implementation, usable standalone or with ONNX Runtime.

Maintenance & Community

  • Actively maintained by PyTorch and ONNX Runtime teams.
  • CI checks for API stability.
  • Contribution guide available.

Licensing & Compatibility

  • MIT License.
  • Compatible with commercial and closed-source applications.

Limitations & Caveats

Training acceleration currently requires NVIDIA or AMD GPUs. Inference package has specific OS and Python version requirements.

Health Check
Last Commit

6 months ago

Responsiveness

1 week

Pull Requests (30d)
0
Issues (30d)
0
Star History
1 stars in the last 30 days

Explore Similar Projects

Starred by Travis Addair Travis Addair(Cofounder of Predibase), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
10 more.

hummingbird by microsoft

0.0%
3k
Compiler for trained ML models into tensor computation
Created 5 years ago
Updated 2 months ago
Starred by Clement Delangue Clement Delangue(Cofounder of Hugging Face), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
20 more.

accelerate by huggingface

0.3%
9k
PyTorch training helper for distributed execution
Created 4 years ago
Updated 1 day ago
Feedback? Help us improve.