Lightweight deep learning framework for on-device inference and training
Top 4.0% on sourcepulse
MNN is a high-performance, lightweight deep learning inference and training framework designed for on-device deployment across mobile, embedded, and PC platforms. It targets developers and researchers needing efficient execution of diverse AI models, including LLMs and diffusion models, with a focus on minimizing resource footprint and maximizing speed.
How It Works
MNN leverages highly optimized assembly code for CPU execution and Metal, OpenCL, Vulkan, and CUDA for GPU acceleration. It supports advanced techniques like Winograd convolution and FP16/Int8 quantization to boost performance and reduce model size. The framework includes a converter for popular model formats (Tensorflow, ONNX, Caffe, Torchscripts) and supports complex model structures with dynamic inputs and control flow.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
MNN's support for certain architectures and precision modes (e.g., BF16 on CPU, NPU acceleration) is marked as 'B' (supported but not optimized or with bugs) or 'C' (not supported), indicating potential areas for improvement or requiring careful evaluation. Community support is predominantly in Chinese.
2 days ago
Inactive