AI infra for non-NVIDIA GPUs, enabling LLM fine-tuning
Top 57.7% on sourcepulse
Felafax provides an AI infrastructure framework for fine-tuning and continued training of open-source Large Language Models (LLMs) on non-NVIDIA hardware, primarily Google TPUs. It targets ML researchers and developers seeking cost-effective and scalable LLM training solutions, offering a simplified workflow and enabling efficient utilization of diverse hardware accelerators.
How It Works
Felafax leverages JAX and its XLA backend for efficient computation across various hardware, including TPUs, AWS Trainium, and AMD/Intel GPUs. This approach allows for seamless scaling from single-core VMs to large TPU pods and supports advanced features like model and data sharding for handling large models and datasets. The framework supports both full-precision and LoRA fine-tuning.
Quick Start & Requirements
pip install pipx
followed by pipx install felafax-cli
.pipx
. Supports Llama 3.1 (1B, 3B, 8B, 70B, 405B) models.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
6 months ago
1+ week