Discover and explore top open-source AI tools and projects—updated daily.
airockchipSDK for deploying AI models on Rockchip chips
Top 36.0% on SourcePulse
This repository provides the RKLLM software stack, enabling users to deploy AI models, particularly Large Language Models (LLMs), on Rockchip NPUs. It targets developers and researchers working with Rockchip's RK3588, RK3576, and RK3562 series platforms, offering accelerated LLM inference and multimodal capabilities.
How It Works
The stack comprises RKLLM-Toolkit for PC-based model conversion and quantization, and RKLLM Runtime for on-device C/C++ API-based inference. Models are converted to an RKLLM format, then executed via a C API, leveraging the RKNPU kernel driver for hardware interaction. This approach optimizes LLM deployment on edge devices by providing a dedicated toolchain and runtime for Rockchip's NPU hardware.
Quick Start & Requirements
git clone https://github.com/airockchip/rknn-toolkit2.export BUILD_CUDA_EXT=0. Potential libomp.so issues may require manual placement from toolchains.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
libomp.so on certain platforms, requiring manual intervention. Python 3.12 requires a specific build flag. The full scope of supported hardware beyond the listed Rockchip series is not detailed.1 month ago
1+ week
samuel-vitorino
google
xorbitsai