MiMo  by XiaomiMiMo

LLM for reasoning, pre-trained and post-trained for math/code tasks

created 3 months ago
1,515 stars

Top 27.8% on sourcepulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

MiMo is a series of 7B parameter language models specifically designed to excel at reasoning tasks, including mathematics and code generation. It targets researchers and developers seeking high-performance models that can compete with much larger architectures, offering a pre-trained base model and fine-tuned versions for enhanced reasoning capabilities.

How It Works

MiMo employs a dual-pronged approach: optimized pre-training and a novel post-training recipe. The base model is pre-trained on approximately 25 trillion tokens with a focus on reasoning patterns, incorporating Multiple-Token Prediction (MTP) for improved performance and inference speed. The post-training phase utilizes a curated dataset of 130K math and code problems, employing rule-based accuracy rewards and a test difficulty-driven reward system to mitigate sparse rewards and stabilize RL training.

Quick Start & Requirements

  • Installation: Inference is officially supported via a fork of vLLM (0.7.3). Hugging Face Transformers can also be used.
  • Prerequisites: Python, vLLM (forked version recommended), Hugging Face Transformers.
  • Resources: Requires significant VRAM for 7B models; specific requirements depend on inference setup.
  • Links: HuggingFace Models, Technical Report (Note: Link is illustrative, actual report link may vary).

Highlighted Details

  • MiMo-7B-RL surpasses larger 32B models on several reasoning benchmarks.
  • Achieves performance comparable to OpenAI's o1-mini on math and code tasks.
  • Features a "Seamless Rollout Engine" for accelerated RL training (2.29x faster).
  • Incorporates Multiple-Token Prediction (MTP) for enhanced inference.

Maintenance & Community

  • Developed by the Xiaomi LLM-Core Team.
  • Contact: mimo@xiaomi.com or GitHub issues.

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatibility: Permissive license suitable for commercial use and integration into closed-source projects.

Limitations & Caveats

  • Evaluation benchmarks were conducted with temperature=0.6, and specific benchmarks used averaged scores over multiple repetitions. Compatibility with inference engines other than the recommended vLLM fork has not been verified.
Health Check
Last commit

1 month ago

Responsiveness

Inactive

Pull Requests (30d)
3
Issues (30d)
0
Star History
413 stars in the last 90 days

Explore Similar Projects

Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), George Hotz George Hotz(Author of tinygrad; Founder of the tiny corp, comma.ai), and
10 more.

TinyLlama by jzhang38

0.3%
9k
Tiny pretraining project for a 1.1B Llama model
created 1 year ago
updated 1 year ago
Feedback? Help us improve.