Discover and explore top open-source AI tools and projects—updated daily.
Minimal, efficient framework for LLM training
Top 97.0% on SourcePulse
Summary
Flame is a minimal, efficient training framework built on torchtitan
for scaling large language models (LLMs). It targets engineers and researchers seeking high performance and ease of use, offering features like zero-cost data preprocessing and advanced parallelism for faster LLM development.
How It Works
Flame leverages torchtitan
to provide a streamlined training experience. Its core design emphasizes efficiency through zero-cost data preprocessing, including online tokenization and dataset shuffling, and supports multiple datasets. The framework is built for scalability, with features like 4D parallelism planned for future releases, aiming to accelerate LLM training pipelines.
Quick Start & Requirements
Installation involves cloning the repository and running pip install .
. Key dependencies include specific versions of flash-linear-attention
and torchtitan
(commit 0b44d4c
). Dataset preparation utilizes the datasets
library for loading, such as HuggingFaceFW/fineweb-edu
. Training is initiated via bash train.sh
, configurable with numerous command-line arguments. Recommended for torch.compile
usage are torch>=2.6
and triton>=3.0
. Multi-node training is supported, with environment variables like MASTER_ADDR
and MASTER_PORT
needing configuration or handled by job schedulers.
Highlighted Details
--training.varlen
to pack variable-length documents into fixed sequences, eliminating padding and improving efficiency.torch.compile
Integration: Supports PyTorch 2.0+ compilation via --training.compile
for potential speedups, though potential conflicts with fused kernels exist.torchao
for potential memory and speed benefits.Maintenance & Community
The provided README does not detail specific community channels (e.g., Discord, Slack), active maintainers beyond the authors listed in the citation, or sponsorship information.
Licensing & Compatibility
The repository's license is not specified in the provided README content. This lack of information presents an adoption blocker, particularly for commercial use or integration into closed-source projects.
Limitations & Caveats
The integration of torch.compile
may encounter conflicts with Flame's fused kernels, requiring up-to-date dependencies. Dataset streaming can be unstable due to network dependencies; local downloads are recommended for reliable training. 4D parallelism is listed as "coming soon." Pipeline parallelism requires manual definition of split points. The absence of explicit licensing information is a significant caveat.
1 month ago
Inactive