PyTorch-native SDK for local LLM inference across diverse platforms
Top 13.7% on sourcepulse
torchchat enables running PyTorch Large Language Models (LLMs) locally across servers, desktops, and mobile devices. It targets developers and power users seeking a flexible, PyTorch-native solution for LLM deployment, offering Python, C++, and mobile (iOS/Android) interfaces with performance optimizations.
How It Works
torchchat leverages PyTorch's native capabilities, including eager execution, compilation via AOT Inductor for optimized desktop/server deployment, and ExecuTorch for mobile optimization. This PyTorch-centric approach emphasizes simplicity, extensibility, and correctness, allowing for modular integration and customization of LLM execution.
Quick Start & Requirements
./install/install_requirements.sh
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The eval
feature is noted as a work in progress. Some model access requires requesting permission via Hugging Face. The README includes a disclaimer about potential performance and compatibility differences compared to original model versions.
3 weeks ago
1 day