tvm  by apache

Compiler stack for deep learning systems

Created 9 years ago
12,629 stars

Top 4.0% on SourcePulse

GitHubView on GitHub
Project Summary

Apache TVM is an open-source compiler stack designed to bridge the gap between high-level deep learning frameworks and diverse hardware backends, enabling optimized execution across CPUs, GPUs, and specialized accelerators. It serves researchers and developers seeking to maximize performance and efficiency for machine learning models.

How It Works

TVM operates by providing an end-to-end compilation flow for deep learning models. Its core innovation lies in its cross-level design, featuring TensorIR for tensor-level representation and Relax for graph-level representation. This allows for joint optimization of computational graphs, tensor programs, and libraries, with a focus on Python-first transformations for enhanced accessibility and customizability.

Quick Start & Requirements

Highlighted Details

  • End-to-end compilation for deep learning systems.
  • Cross-level design with TensorIR (tensor-level) and Relax (graph-level) representations.
  • Python-first transformations for accessibility and customization.
  • Foundation for building domain-specific vertical compilers (e.g., for LLMs).

Maintenance & Community

TVM follows the Apache committer model, aiming for community-driven maintenance. Further details can be found in the Contributor Guide.

Licensing & Compatibility

TVM is licensed under the Apache-2.0 license, permitting commercial use and integration with closed-source projects.

Limitations & Caveats

The project's design has evolved significantly from its initial research origins, with the current version focusing on TensorIR and Relax.

Health Check
Last Commit

2 days ago

Responsiveness

1 day

Pull Requests (30d)
97
Issues (30d)
13
Star History
109 stars in the last 30 days

Explore Similar Projects

Starred by Yaowei Zheng Yaowei Zheng(Author of LLaMA-Factory), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
1 more.

VeOmni by ByteDance-Seed

3.4%
1k
Framework for scaling multimodal model training across accelerators
Created 5 months ago
Updated 3 weeks ago
Starred by Nat Friedman Nat Friedman(Former CEO of GitHub), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
15 more.

FasterTransformer by NVIDIA

0.1%
6k
Optimized transformer library for inference
Created 4 years ago
Updated 1 year ago
Feedback? Help us improve.