OnnxSlim can help you slim your onnx model, with less operators, but same accuracy, better inference speed.
- 🚀 2025/05/17: OnnxSlim is merged into optimum 🤗🤗🤗
- 🚀 2025/04/30: Rank 1st in the AICAS 2025 LLM inference optimization challenge
- 🚀 2025/01/28: Achieved 1M downloads
- 🚀 2024/06/23: OnnxSlim is merged into transformers.js 🤗🤗🤗
- 🚀 2024/06/02: OnnxSlim is merged into ultralytics ❤️❤️❤️
- 🚀 2024/04/30: Rank 1st in the AICAS 2024 LLM inference optimization challenge held by Arm and T-head
- 🚀 2024/01/25: OnnxSlim is merged to mnn-llm, performance increased by 5%
pip install onnxslim
pip install git+https://github.com/inisis/OnnxSlim@main
git clone https://github.com/inisis/OnnxSlim && cd OnnxSlim/
pip install .
onnxslim your_onnx_model slimmed_onnx_model
For more usage, see onnxslim -h or refer to our examples
Mozilla/smart_autofill
alibaba/MNN
PaddlePaddle/PaddleOCR
huggingface/transformers.js
huggingface/optimum
THU-MIG/yolov10
ultralytics/ultralytics
ModelScope/FunASR
alibaba/MNN-LLM
deepghs/imgutils
sunsmarterjie/yolov12
nndeploy/nndeploy
Discord: https://discord.gg/nRw2Fd3VUS QQ Group: 873569894