CLI tool for streamlined post-training of AI models
Top 5.1% on sourcepulse
Axolotl is a comprehensive toolkit for streamlining the post-training of large language models, targeting AI researchers and engineers. It simplifies complex fine-tuning tasks like LoRA, QLoRA, and full fine-tuning, enabling efficient customization of pre-trained models.
How It Works
Axolotl leverages a YAML configuration system to manage the entire model training pipeline, from data preprocessing to inference and evaluation. This approach offers a unified and reproducible workflow. It integrates with performance-enhancing libraries such as xformers, Flash Attention, and various multi-GPU strategies (FSDP, DeepSpeed), aiming to maximize training speed and efficiency.
Quick Start & Requirements
pip install --no-build-isolation axolotl[flash-attn,deepspeed]
axolotl fetch examples
.axolotl train examples/llama-3/lora-1b.yml
.Highlighted Details
Maintenance & Community
The project is sponsored by Modal. Community support is available via Discord.
Licensing & Compatibility
Licensed under the Apache 2.0 License, permitting commercial use and integration with closed-source projects.
Limitations & Caveats
Support for certain models (e.g., Mixtral-MoE, Falcon, Pythia) with specific optimizations like GPTQ or Flash Attention is marked as untested or not supported in the provided table.
21 hours ago
1 day