Toolkit for training/running LLMs on Apple Silicon using MLX
Top 94.9% on sourcepulse
SiLLM is a toolkit designed to simplify the training and inference of Large Language Models (LLMs) on Apple Silicon hardware, leveraging the MLX framework. It caters to researchers and developers looking to experiment with LLMs locally on Macs, offering features for model loading, fine-tuning (LoRA, DPO), and deployment.
How It Works
SiLLM builds upon MLX, providing a streamlined interface for common LLM operations. It supports loading models from various formats (Huggingface, Torch, GGUF, MLX) and includes implementations for LoRA and DPO fine-tuning. The toolkit also offers experimental features like speculative decoding and beam search, aiming to make advanced LLM techniques accessible on Apple Silicon.
Quick Start & Requirements
pip install sillm-mlx
git clone https://github.com/armbues/SiLLM.git
cd SiLLM/app && pip install -r requirements.txt && python -m chainlit run app.py -w
Highlighted Details
Maintenance & Community
The project is actively maintained by armbues and acknowledges contributions from the MLX community.
Licensing & Compatibility
Licensed under the MIT License, permitting commercial use and integration with closed-source projects.
Limitations & Caveats
The project includes experimental features, which may be subject to change or instability. Specific performance benchmarks are not detailed in the README.
1 month ago
1 day