felafax  by felafax

AI infra for non-NVIDIA GPUs, enabling LLM fine-tuning

created 1 year ago
566 stars

Top 57.7% on sourcepulse

GitHubView on GitHub
Project Summary

Felafax provides an AI infrastructure framework for fine-tuning and continued training of open-source Large Language Models (LLMs) on non-NVIDIA hardware, primarily Google TPUs. It targets ML researchers and developers seeking cost-effective and scalable LLM training solutions, offering a simplified workflow and enabling efficient utilization of diverse hardware accelerators.

How It Works

Felafax leverages JAX and its XLA backend for efficient computation across various hardware, including TPUs, AWS Trainium, and AMD/Intel GPUs. This approach allows for seamless scaling from single-core VMs to large TPU pods and supports advanced features like model and data sharding for handling large models and datasets. The framework supports both full-precision and LoRA fine-tuning.

Quick Start & Requirements

  • Install: pip install pipx followed by pipx install felafax-cli.
  • Authentication: Requires a token from preview.felafax.ai.
  • Prerequisites: Python, pipx. Supports Llama 3.1 (1B, 3B, 8B, 70B, 405B) models.
  • Resources: Free TPU resources available on Google Colab.
  • Docs: felafax.ai

Highlighted Details

  • Supports Llama 3.1 JAX implementation, converted from PyTorch for performance.
  • Offers a CLI for easy fine-tuning setup, job monitoring, and model interaction.
  • Demonstrated 405B Llama 3.1 fine-tuning on 8x AMD MI300x GPUs using JAX sharding.
  • Enables free fine-tuning on Google Colab TPUs.

Maintenance & Community

Licensing & Compatibility

  • The README does not explicitly state a license. Compatibility for commercial use is not specified.

Limitations & Caveats

  • The 405B model fine-tuning was performed in JAX eager mode due to infrastructure constraints, indicating potential for significant performance improvements with JIT compilation.
  • The project appears to be in active development, with features like the 405B model fine-tuning noted as "New!".
Health Check
Last commit

6 months ago

Responsiveness

1+ week

Pull Requests (30d)
0
Issues (30d)
0
Star History
10 stars in the last 90 days

Explore Similar Projects

Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), and
2 more.

xTuring by stochasticai

0.0%
3k
SDK for fine-tuning and customizing open-source LLMs
created 2 years ago
updated 10 months ago
Feedback? Help us improve.