Discover and explore top open-source AI tools and projects—updated daily.
JetAstraScalable sequence generation via diffusion and autoregression synergy
Top 97.4% on SourcePulse
Summary
SDAR (Synergy of Diffusion and AutoRegression) is a large-scale language model family that merges autoregressive (AR) and discrete diffusion modeling. It targets researchers and practitioners seeking efficient, high-performance LLMs, offering competitive accuracy with significantly faster inference speeds (2-4x) and strong reasoning capabilities.
How It Works
SDAR combines the training efficiency of AR methods with the parallel decoding of diffusion models. This synergistic approach allows for training scalability while enabling highly parallelized, faster generation. The core innovation lies in this hybrid paradigm, positioning SDAR as a powerful diffusion-based language model that rivals state-of-the-art AR models, particularly excelling in generalist and specialist roles.
Quick Start & Requirements
Installation involves cloning the repository, initializing submodules, and installing dependencies like transformers>=4.52.4 and flash-attn. GPU acceleration is essential. The project offers multiple inference engines: a built-in script, the optimized JetEngine (achieving 1800+ tokens/sec on A800, 3700+ on H100), and integration with lmdeploy. Fine-tuning is supported via a framework powered by LlamaFactory.
Highlighted Details
Maintenance & Community
The project is actively developed, with core contributors including Shuang Cheng, Yihan Bian, Dawei Liu, and Biqing Qi. While specific community channels like Discord/Slack are not detailed in the README snippet, contact information for key researchers is provided for inquiries. The roadmap indicates ongoing feature development.
Licensing & Compatibility
SDAR is released under the MIT license, which is permissive for commercial use and integration into closed-source projects.
Limitations & Caveats
The project is explicitly described as being in an "early experimental state." The developers are actively working on further systematic development and welcome collaborations.
2 weeks ago
Inactive
hao-ai-lab
hao-ai-lab
SafeAILab
bytedance