Discover and explore top open-source AI tools and projects—updated daily.
tensorflowPerformant, modular runtime for TensorFlow
Top 45.7% on SourcePulse
A performant and modular runtime for TensorFlow, TFRT (TensorFlow Runtime) aims to provide a unified, extensible infrastructure layer for diverse hardware, focusing on efficient multithreaded CPU utilization and asynchronous programming models. It targets researchers experimenting with new models, developers seeking production serving performance, and hardware makers looking to integrate their devices into TensorFlow. TFRT offers low-level efficiency and aims for best-in-class performance across various domains.
How It Works
TFRT leverages MLIR (Multi-Level Intermediate Representation) for optimizing and compiling TensorFlow graphs into a Binary Executable Format (BEF). This BEF is then executed by the bef_executor. The runtime is designed for fully asynchronous programming and efficient use of host CPUs, providing a low-level, efficient execution layer. Its modular design allows for extensibility, particularly for integrating custom hardware backends.
Quick Start & Requirements
bazel build //tools:bef_executor, bazel build //tools:tfrt_translate. Execution: bazel-bin/tools/tfrt_translate -mlir-to-bef path/to/program.mlir | bazel-bin/tools/bef_executor.pip install libclang.Highlighted Details
Maintenance & Community
TFRT is currently not open to contributions; workflows for accepting contributions are under development. General discussions can occur via a mailing list, and bugs/feature requests are tracked on GitHub issues.
Licensing & Compatibility
Licensed under the Apache License 2.0, which is generally permissive for commercial use and integration into closed-source projects.
Limitations & Caveats
TFRT is an early-stage project and not yet ready for general use. Current platform support is limited to Ubuntu-16.04, with other platforms planned for future releases. The project is not yet open for external contributions.
2 months ago
Inactive
 Shengjia Zhao(Chief Scientist at Meta Superintelligence Lab), 
google
grahamjenson
google-research
triton-inference-server
tensorflow