Discover and explore top open-source AI tools and projects—updated daily.
Agentic models for deep research and complex problem-solving
Top 85.6% on SourcePulse
MiroThinker is an open-source series of agentic large language models designed for deep research and complex, long-horizon problem-solving. Built upon the Qwen3 architecture, it offers models in 8B, 14B, and 32B parameter sizes, with both Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) variants. MiroThinker excels in tasks requiring tool use, code execution, web browsing, and document processing, demonstrating state-of-the-art performance among open-source models on benchmarks like GAIA.
How It Works
MiroThinker integrates advanced capabilities such as task decomposition, multi-hop reasoning, and retrieval-augmented generation. It leverages the MiroFlow framework, which provides a robust environment for agent development, featuring enhanced conversation management, flexible tool integration (supporting both open-source and commercial tools), and comprehensive benchmark evaluations. The models are trained on the MiroVerse dataset and utilize the MiroTrain framework for efficient training.
Quick Start & Requirements
pf4*
), and set up the environment using uv sync
.uv
package manager, and API keys for various services (Serper, E2B, OpenAI, Anthropic, etc., depending on tool configuration)..env
file. Serving models can be done via SGLang or quantized methods (llama.cpp, Ollama).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model's Chinese language capabilities are currently limited due to the predominantly English nature of the MiroVerse-v0.1 dataset, with plans to improve this in future versions. Performance metrics are reported using both "Best Pass@1" and "Pass@1 (Avg@8)" for stability and peak performance comparison.
14 hours ago
Inactive