TensorRT_Tutorial  by LitLeo

TensorRT tutorials and resources

created 8 years ago
1,030 stars

Top 37.1% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides a comprehensive set of Chinese translated documentation and tutorials for NVIDIA's TensorRT, a high-performance deep learning inference optimizer and runtime. It aims to help users, particularly those new to TensorRT or facing documentation challenges, understand and effectively utilize TensorRT for accelerating deep learning models, with a focus on INT8 quantization and custom plugin development.

How It Works

The project offers translated versions of the TensorRT User Guide and detailed explanations of TensorRT samples. It also includes practical usage experiences and blog posts covering topics like INT8 quantization, FP16 precision, custom layer implementation, and model conversion strategies. The content is structured to guide users from basic TensorRT usage to advanced techniques like creating custom plugins and optimizing performance.

Quick Start & Requirements

  • Installation: No direct installation command is provided as this is a documentation and tutorial repository. Users will need to install TensorRT separately from NVIDIA's developer site.
  • Prerequisites: NVIDIA GPU, CUDA Toolkit, and TensorRT itself. Familiarity with C++ and deep learning concepts is beneficial.
  • Resources: Links to TensorRT download pages, official documentation, and GTC presentations are provided.

Highlighted Details

  • Detailed explanations and translations for TensorRT versions up to 8.5.3.
  • Focus on INT8 quantization, FP16 precision, and custom plugin development.
  • Includes practical examples and blog posts on optimizing inference.
  • Covers various model conversion methods and acceleration strategies.

Maintenance & Community

The project was initiated in 2017 and has seen updates, including translations for TensorRT 8.5.3 in late 2023. A QQ group (483063470) is available for community interaction. The repository also mentions recruitment for AI heterogeneous acceleration internships at Tencent Beijing AILAB.

Licensing & Compatibility

The repository itself does not specify a license. The content is primarily for educational and informational purposes, translating and explaining NVIDIA's TensorRT, which is subject to NVIDIA's own licensing terms.

Limitations & Caveats

Some older tutorial chapters (1-2) suggest referring to the latest video versions, implying potential updates or improvements in those specific areas. The project is a community effort focused on translation and explanation, not a direct software package to be installed.

Health Check
Last commit

1 year ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
13 stars in the last 90 days

Explore Similar Projects

Starred by Nat Friedman Nat Friedman(Former CEO of GitHub), Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), and
6 more.

FasterTransformer by NVIDIA

0.1%
6k
Optimized transformer library for inference
created 4 years ago
updated 1 year ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Omar Sanseviero Omar Sanseviero(DevRel at Google DeepMind), and
5 more.

TensorRT-LLM by NVIDIA

0.6%
11k
LLM inference optimization SDK for NVIDIA GPUs
created 1 year ago
updated 11 hours ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Eric Zhang Eric Zhang(Founding Engineer at Modal), and
5 more.

flux by black-forest-labs

0.4%
24k
Inference code for FLUX image generation & editing models
created 1 year ago
updated 3 days ago
Feedback? Help us improve.