AI cluster for running models on diverse devices
Top 1.3% on sourcepulse
exo enables users to create distributed AI inference clusters using everyday devices, including smartphones, Macs, and Raspberry Pis. It targets individuals and businesses looking to run large language models locally, offering a unified, peer-to-peer approach to distributed computing without a master-worker architecture. The primary benefit is leveraging existing hardware for powerful AI inference, with a ChatGPT-compatible API for easy integration.
How It Works
exo employs dynamic model partitioning, intelligently splitting AI models across available devices based on network topology and individual device resources. This allows for the execution of models larger than any single device could handle. The system uses automatic device discovery and a peer-to-peer (P2P) networking model, ensuring any connected device can contribute to the cluster. The default partitioning strategy is ring memory weighted partitioning, where each device processes layers proportional to its memory capacity.
Quick Start & Requirements
git clone https://github.com/exo-explore/exo.git && cd exo && pip install -e .
or source install.sh
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is experimental software with expected bugs. The iOS implementation is currently behind and requires manual access requests. PyTorch and Radio/Bluetooth discovery modules are listed as under development.
4 months ago
1 day