Discover and explore top open-source AI tools and projects—updated daily.
EvolvingLMMs-LabUnified engine for training large-scale multimodal AI models
Top 65.2% on SourcePulse
A simple, unified multimodal model training engine designed for lean, flexible, and scalable development. It targets researchers and engineers needing an efficient framework to train complex multimodal models at scale, offering significant optimizations for both distributed training and memory/compute efficiency.
How It Works
LMMS-Engine employs a modular architecture using Factory and Builder patterns for extensibility. Its core strength lies in advanced optimization techniques, including PyTorch 2.0+ FSDP2 for distributed training, Ulysses Sequence Parallelism for handling ultra-long contexts, and Triton-fused kernels (Liger) for substantial memory reduction. It also integrates novel optimizers like Muon and efficient attention mechanisms such as Native Sparse Attention (NSA) and Flash Attention with unpadding, aiming to maximize Model FLOPs Utilization (MFU).
Quick Start & Requirements
Installation involves cloning the repository, synchronizing dependencies with uv sync, and optionally installing performance-enhancing packages like flash-attn and liger-kernel. Training is launched using torchrun or accelerate launch with a configuration YAML file.
git clone https://github.com/EvolvingLMMs-Lab/lmms-engine.git && cd lmms-engine && uv syncHighlighted Details
Maintenance & Community
The project is developed by LMMs-Lab, with a website available at https://lmms-lab.com/. No specific community channels (like Discord/Slack) or public roadmaps are detailed in the README.
Licensing & Compatibility
This project is licensed under the Apache 2.0 License, which permits commercial use and modification.
Limitations & Caveats
Native Sparse Attention (NSA) is currently only supported for the BAGEL model. Some features, like Ulysses Sequence Parallel (USP) for the BAGEL model, are marked as "TBD" in the example table, indicating ongoing development or incomplete integration. As a v0.1 release, expect potential for rapid changes and evolving feature sets.
1 day ago
Inactive
jzhang38
feifeibear
HazyResearch
microsoft
MiniMax-AI