felafax  by felafax

AI infra for non-NVIDIA GPUs, enabling LLM fine-tuning

Created 1 year ago
566 stars

Top 56.8% on SourcePulse

GitHubView on GitHub
Project Summary

Felafax provides an AI infrastructure framework for fine-tuning and continued training of open-source Large Language Models (LLMs) on non-NVIDIA hardware, primarily Google TPUs. It targets ML researchers and developers seeking cost-effective and scalable LLM training solutions, offering a simplified workflow and enabling efficient utilization of diverse hardware accelerators.

How It Works

Felafax leverages JAX and its XLA backend for efficient computation across various hardware, including TPUs, AWS Trainium, and AMD/Intel GPUs. This approach allows for seamless scaling from single-core VMs to large TPU pods and supports advanced features like model and data sharding for handling large models and datasets. The framework supports both full-precision and LoRA fine-tuning.

Quick Start & Requirements

  • Install: pip install pipx followed by pipx install felafax-cli.
  • Authentication: Requires a token from preview.felafax.ai.
  • Prerequisites: Python, pipx. Supports Llama 3.1 (1B, 3B, 8B, 70B, 405B) models.
  • Resources: Free TPU resources available on Google Colab.
  • Docs: felafax.ai

Highlighted Details

  • Supports Llama 3.1 JAX implementation, converted from PyTorch for performance.
  • Offers a CLI for easy fine-tuning setup, job monitoring, and model interaction.
  • Demonstrated 405B Llama 3.1 fine-tuning on 8x AMD MI300x GPUs using JAX sharding.
  • Enables free fine-tuning on Google Colab TPUs.

Maintenance & Community

Licensing & Compatibility

  • The README does not explicitly state a license. Compatibility for commercial use is not specified.

Limitations & Caveats

  • The 405B model fine-tuning was performed in JAX eager mode due to infrastructure constraints, indicating potential for significant performance improvements with JIT compilation.
  • The project appears to be in active development, with features like the 405B model fine-tuning noted as "New!".
Health Check
Last Commit

7 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
2 stars in the last 30 days

Explore Similar Projects

Starred by Clement Delangue Clement Delangue(Cofounder of Hugging Face), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
20 more.

accelerate by huggingface

0.3%
9k
PyTorch training helper for distributed execution
Created 4 years ago
Updated 1 day ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Stefan van der Walt Stefan van der Walt(Core Contributor to scientific Python ecosystem), and
12 more.

litgpt by Lightning-AI

0.1%
13k
LLM SDK for pretraining, finetuning, and deploying 20+ high-performance LLMs
Created 2 years ago
Updated 5 days ago
Starred by Junyang Lin Junyang Lin(Core Maintainer at Alibaba Qwen), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
25 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.