Discover and explore top open-source AI tools and projects—updated daily.
meituan-longcatA 560B parameter MoE language model optimized for efficiency and agentic tasks
Top 31.4% on SourcePulse
LongCat-Flash-Chat is a 560 billion parameter language model designed for efficient computation and high performance, particularly in agentic tasks. It targets developers and researchers seeking a powerful, scalable LLM with advanced reasoning and tool-use capabilities, offering competitive performance against leading models.
How It Works
LongCat-Flash utilizes a Mixture-of-Experts (MoE) architecture with a "zero-computation experts" mechanism, dynamically activating 18.6B-31.3B parameters (average ~27B) per token based on context. This, combined with a Shortcut-connected MoE (ScMoE) design, expands the computation-communication overlap window, enabling efficient scaling and high-throughput inference (over 100 TPS). Its training strategy includes hyperparameter transfer, a model-growth initialization, a stability suite (router-gradient balancing, z-loss), and deterministic computation for reproducibility and error detection. Agentic capabilities are enhanced through a multi-stage training pipeline, including data fusion, extended context length (128k), and a multi-agent synthesis framework for generating complex tasks.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model has not been evaluated for every possible downstream application and may exhibit performance variations across languages. Developers must carefully assess accuracy, safety, and fairness, and comply with all applicable laws and regulations, especially in sensitive or high-risk scenarios.
3 weeks ago
Inactive
allenai