Discover and explore top open-source AI tools and projects—updated daily.
triton-inference-serverTriton backend tools for model execution
Top 79.0% on SourcePulse
This repository provides common source, scripts, and utilities for developing custom backends for the Triton Inference Server. It targets developers building custom inference logic or integrating new frameworks with Triton, enabling efficient model execution and pre/post-processing.
How It Works
Backends are implemented as shared libraries adhering to the Triton Backend API, which defines interfaces for managing backend, model, and instance lifecycles, as well as handling inference requests and responses. This API allows backends to interact with Triton for request processing, tensor data access, and response generation, supporting both single and decoupled response patterns.
Quick Start & Requirements
mkdir build && cd build && cmake -DCMAKE_INSTALL_PREFIX:PATH=$(pwd)/install .. && make installcommon and core repositories. Specific tags can be set via CMake arguments (e.g., -DTRITON_COMMON_REPO_TAG=[tag]).CMakeLists.txt, rather than building this repository directly.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
11 hours ago
Inactive
VoltaML
imaurer
ashleve
xorbitsai
modelscope
huggingface
tensorflow