Discover and explore top open-source AI tools and projects—updated daily.
pytorchPyTorch-native SDK for local LLM inference across diverse platforms
Top 13.2% on SourcePulse
torchchat enables running PyTorch Large Language Models (LLMs) locally across servers, desktops, and mobile devices. It targets developers and power users seeking a flexible, PyTorch-native solution for LLM deployment, offering Python, C++, and mobile (iOS/Android) interfaces with performance optimizations.
How It Works
torchchat leverages PyTorch's native capabilities, including eager execution, compilation via AOT Inductor for optimized desktop/server deployment, and ExecuTorch for mobile optimization. This PyTorch-centric approach emphasizes simplicity, extensibility, and correctness, allowing for modular integration and customization of LLM execution.
Quick Start & Requirements
./install/install_requirements.sh.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The eval feature is noted as a work in progress. Some model access requires requesting permission via Hugging Face. The README includes a disclaimer about potential performance and compatibility differences compared to original model versions.
4 months ago
1 day
b4rtaz
nomic-ai