Discover and explore top open-source AI tools and projects—updated daily.
inisisONNX model optimization toolkit
Top 68.2% on SourcePulse
OnnxSlim is a toolkit designed to optimize ONNX models by reducing operator count, aiming to improve inference speed while preserving accuracy. It targets developers and researchers seeking to enhance the performance of their ONNX models for deployment across various platforms.
How It Works
OnnxSlim employs graph simplification techniques to reduce the complexity of ONNX models. Its core approach involves identifying and eliminating redundant or equivalent operators within the model graph. This process results in a more streamlined model representation, which directly translates to faster inference times without sacrificing the model's predictive accuracy.
Quick Start & Requirements
pip install onnxslimpip install git+https://github.com/inisis/OnnxSlim@mainonnxslim <your_onnx_model> <slimmed_onnx_model>onnx.load, apply onnxslim.slim, and save using onnx.save.Highlighted Details
Maintenance & Community
https://discord.gg/nRw2Fd3VUS) and QQ Group (873569894).Licensing & Compatibility
Limitations & Caveats
5 days ago
Inactive
merrymercy
Shengjia Zhao(Chief Scientist at Meta Superintelligence Lab),
google
grahamjenson
ThilinaRajapakse
google-research
triton-inference-server
tensorflow
visenger
PaddlePaddle