Discover and explore top open-source AI tools and projects—updated daily.
C++ library for efficient on-device LLM execution
Top 81.8% on SourcePulse
LiteRT-LM is a C++ library designed for efficient on-device execution of language model pipelines across edge platforms. It targets developers building applications that require local LLM inference, offering cross-platform compatibility and hardware acceleration. The library aims to simplify the deployment of complex LLM workflows, enabling greater flexibility and performance on diverse hardware.
How It Works
LiteRT-LM builds upon the LiteRT runtime, providing a C++ API to manage LLM pipelines. It supports customizability for specific features and leverages hardware acceleration (CPU, GPU, NPU) for performance gains. Models are converted to a proprietary .litertlm
format, which is optimized for efficient loading and execution on target devices.
Quick Start & Requirements
.litertlm
format.Highlighted Details
GenerateContent
and granular RunPrefill
/RunDecode
C++ APIs for flexible inference control..litertlm
format is an evolution of .task
files, designed for better compression and metadata inclusion.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
.litertlm
model format is proprietary and specific to this library.16 hours ago
Inactive