ONNX Model Deployment ML Frameworks
Tools and frameworks for converting, optimizing, and deploying ONNX models across different runtime environments and hardware platforms. Does NOT include general model training, framework-specific inference (e.g., pure TensorFlow or PyTorch inference), or non-ONNX deployment solutions.
There are 115 onnx model deployment frameworks tracked. 6 score above 70 (verified tier). The highest-rated is microsoft/onnxruntime at 100/100 with 19,534 stars and 76,261,123 monthly downloads. 8 of the top 10 are actively maintained.
Get all 115 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=ml-frameworks&subcategory=onnx-model-deployment&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Framework | Score | Tier |
|---|---|---|---|
| 1 |
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training... |
|
Verified |
| 2 |
onnx/onnx
Open standard for machine learning interoperability |
|
Verified |
| 3 |
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on... |
|
Verified |
| 4 |
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras... |
|
Verified |
| 5 |
microsoft/onnxconverter-common
Common utilities for ONNX converters |
|
Verified |
| 6 |
NVIDIA/DALI
A GPU-accelerated library containing highly optimized building blocks and an... |
|
Verified |
| 7 |
pytorch/TensorRT
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT |
|
Established |
| 8 |
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX |
|
Established |
| 9 |
onnxsim/onnxsim
Simplify your onnx model |
|
Established |
| 10 |
onnx/onnx-tensorrt
ONNX-TensorRT: TensorRT backend for ONNX |
|
Established |
| 11 |
jolibrain/deepdetect
Deep Learning API and Server in C++14 support for PyTorch,TensorRT, Dlib,... |
|
Established |
| 12 |
NVIDIA/MinkowskiEngine
Minkowski Engine is an auto-diff neural network library for high-dimensional... |
|
Established |
| 13 |
nntrainer/nntrainer
NNtrainer is Software Framework for Training and Inferencing Neural Network... |
|
Established |
| 14 |
sdpython/onnx-extended
New operators for the ReferenceEvaluator, new kernels for onnxruntime, CPU, CUDA |
|
Established |
| 15 |
onnx/tensorflow-onnx
Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX |
|
Established |
| 16 |
aphrodite-engine/aphrodite-engine
Large-scale LLM inference engine |
|
Established |
| 17 |
triton-inference-server/dali_backend
The Triton backend that allows running GPU-accelerated data pre-processing... |
|
Established |
| 18 |
PINTO0309/simple-onnx-processing-tools
A set of simple tools for splitting, merging, OP deletion, size compression,... |
|
Established |
| 19 |
OAID/Tengine
Tengine is a lite, high performance, modular inference engine for embedded device |
|
Established |
| 20 |
facebookarchive/caffe2
Caffe2 is a lightweight, modular, and scalable deep learning framework. |
|
Established |
| 21 |
Quantco/spox
Pythonic framework for building ONNX graphs |
|
Established |
| 22 |
Tencent/TNN
TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep... |
|
Established |
| 23 |
kraiskil/onnx2c
Open Neural Network Exchange to C compiler. |
|
Established |
| 24 |
emmtrix/emx-onnx-cgen
Deterministic ONNX-to-C compiler for embedded and safety-critical systems,... |
|
Established |
| 25 |
VeriSilicon/tflite-vx-delegate
Tensorflow Lite external delegate based on TIM-VX |
|
Established |
| 26 |
zjhellofss/KuiperInfer
校招、秋招、春招、实习好项目!带你从零实现一个高性能的深度学习推理库,支持大模型 llama2... |
|
Emerging |
| 27 |
AlexanderLutsenko/nobuco
Pytorch to Keras/Tensorflow/TFLite conversion made intuitive |
|
Emerging |
| 28 |
dotnet/infer
Infer.NET is a framework for running Bayesian inference in graphical models |
|
Emerging |
| 29 |
xboot/libonnx
A lightweight, portable pure C99 onnx inference engine for embedded devices... |
|
Emerging |
| 30 |
onnx/onnx-tensorflow
Tensorflow Backend for ONNX |
|
Emerging |
| 31 |
hailo-ai/hailort
An open source light-weight and high performance inference framework for... |
|
Emerging |
| 32 |
kibae/onnxruntime-server
ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP... |
|
Emerging |
| 33 |
amir-abdi/keras_to_tensorflow
General code to convert a trained keras model into an inference tensorflow model |
|
Emerging |
| 34 |
justinchuby/onnx-shape-inference
Symbolic shape inference for ONNX |
|
Emerging |
| 35 |
wuba/dl_inference
通用深度学习推理工具,可在生产环境中快速上线由TensorFlow、PyTorch、Caffe框架训练出的深度学习模型。 |
|
Emerging |
| 36 |
solrex/caffe-mobile
Optimized (for size and speed) Caffe lib for iOS and Android with... |
|
Emerging |
| 37 |
OAID/Caffe-HRT
Heterogeneous Run Time version of Caffe. Added heterogeneous capabilities to... |
|
Emerging |
| 38 |
iwatake2222/InferenceHelper
C++ Helper Class for Deep Learning Inference Frameworks: TensorFlow Lite,... |
|
Emerging |
| 39 |
iwatake2222/play_with_tflite
Sample projects for TensorFlow Lite in C++ with delegates such as GPU,... |
|
Emerging |
| 40 |
NVIDIA/nvImageCodec
A nvImageCodec library of GPU- and CPU- accelerated codecs featuring a... |
|
Emerging |
| 41 |
gizatechxyz/orion
ONNX Runtime in Cairo 1.0 for verifiable ML inference using STARK |
|
Emerging |
| 42 |
gmalivenko/pytorch2keras
PyTorch to Keras model convertor |
|
Emerging |
| 43 |
NervanaSystems/ngraph
nGraph has moved to OpenVINO |
|
Emerging |
| 44 |
iree-org/onnxruntime-ep-iree
IREE based OnnxRuntime Execution Provider |
|
Emerging |
| 45 |
microsoft/nn-Meter
A DNN inference latency prediction toolkit for accurately modeling and... |
|
Emerging |
| 46 |
NVIDIA/tensorrt-laboratory
Explore the Capabilities of the TensorRT Platform |
|
Emerging |
| 47 |
luxonis/modelconverter
Model converter for Luxonis' cameras. Convert your model from ONNX, TF, ...... |
|
Emerging |
| 48 |
intel/npu-nn-cost-model
Library for modelling performance costs of different Neural Network... |
|
Emerging |
| 49 |
sdpython/onnx-array-api
Array APIs to write ONNX Graphs |
|
Emerging |
| 50 |
ysh329/deep-learning-model-convertor
The convertor/conversion of deep learning models for different deep learning... |
|
Emerging |
| 51 |
OAID/Tengine-Convert-Tools
Tengine Convert Tool supports converting multi framworks' models into tmfile... |
|
Emerging |
| 52 |
cyrusbehr/tensorrt-cpp-api
TensorRT C++ API Tutorial |
|
Emerging |
| 53 |
olilarkin/ort-builder
ONNX Runtime static library builder |
|
Emerging |
| 54 |
iwatake2222/play_with_tensorrt
Sample projects for TensorRT in C++ |
|
Emerging |
| 55 |
alrevuelta/cONNXr
Pure C ONNX runtime with zero dependancies for embedded devices |
|
Emerging |
| 56 |
AsuharietYgvar/AppleNeuralHash2ONNX
Convert Apple NeuralHash model for CSAM Detection to ONNX. |
|
Emerging |
| 57 |
kibae/pg_onnx
pg_onnx: ONNX Runtime integrated with PostgreSQL. Perform ML inference with... |
|
Emerging |
| 58 |
NobuoTsukamoto/tensorrt-examples
TensorRT Examples (TensorRT, Jetson Nano, Python, C++) |
|
Emerging |
| 59 |
xmba15/onnx_runtime_cpp
small c++ library to quickly deploy models using onnxruntime |
|
Emerging |
| 60 |
OAID/MXNet-HRT
Heterogeneous Run Time version of MXNet. Added heterogeneous capabilities to... |
|
Emerging |
| 61 |
Adlik/Adlik
Adlik: Toolkit for Accelerating Deep Learning Inference |
|
Emerging |
| 62 |
Tencent/Forward
A library for high performance deep learning inference on NVIDIA GPUs. |
|
Emerging |
| 63 |
Phoenix8215/A-White-Paper-on-Neural-Network-Deployment
模型部署白皮书(CUDA|ONNX|TensorRT|C++)🚀🚀🚀 |
|
Emerging |
| 64 |
merrymercy/tvm-mali
Optimizing Mobile Deep Learning on ARM GPU with TVM |
|
Emerging |
| 65 |
mshr-h/onnx2fx
Turn ONNX models into inspectable, transformable, and trainable PyTorch FX... |
|
Emerging |
| 66 |
MegEngine/MegFlow
Efficient ML solution for long-tailed demands. |
|
Emerging |
| 67 |
RobertBiehl/caffe2-ios
☕️ Caffe2Kit. A simple one step integration of Caffe2 for iOS. |
|
Emerging |
| 68 |
hshatti/TONNXRuntime
TOnnxRuntime is a Microsoft ONNXRuntime AI and Machine Learning Library for... |
|
Emerging |
| 69 |
vibeswithkk/ZENITH
An open-source bridge for faster ML inference. Supports PyTorch, JAX, and... |
|
Emerging |
| 70 |
MaitreChen/openvino-lenet-sample
本仓库包含了完整的深度学习应用开发流程,以经典的手写字符识别为例,基于LeNet网络构建。推理部分使用torch、onnxruntime以及openvino框架💖 |
|
Emerging |
| 71 |
qbxlvnf11/convert-pytorch-onnx-tensorrt
Converting weights of Pytorch models to ONNX & TensorRT engines |
|
Emerging |
| 72 |
yuanmu97/InFi
InFi is a library for building input filters for resource-efficient inference. |
|
Emerging |
| 73 |
yuzawa-san/onnxruntime-java
A type-safe, lightweight, modern, and performant binding Java binding of... |
|
Emerging |
| 74 |
RunEdgeAI/coreflow
Graph-based C++ runtime for building and executing AI, ML, and computer... |
|
Emerging |
| 75 |
RidgeRun/r2inference
RidgeRun Inference Framework |
|
Emerging |
| 76 |
athrva98/polyinfer
Unified deployment pipeline |
|
Emerging |
| 77 |
zerollzeng/tiny-tensorrt
Deploy your model with TensorRT quickly. |
|
Emerging |
| 78 |
fateshelled/OnnxGraphQt
ONNX model visualizer |
|
Emerging |
| 79 |
TeiaCare/TeiaCareInferenceClient
TeiaCareInferenceClient is a C++ inference client library that implements... |
|
Emerging |
| 80 |
haobosang/TinyTensor
TinyTensor is a tool for running already trained NN (Neural Network) models... |
|
Emerging |
| 81 |
smarter-project/armnn_tflite_backend
TensorFlow Lite backend with ArmNN delegate support for Nvidia Triton |
|
Experimental |
| 82 |
becauseofAI/caffe-plus-plus
Caffe++: assemble new features to enhance Caffe☕️ |
|
Experimental |
| 83 |
luchangli03/onnxsim_large_model
simplify >2GB large onnx model |
|
Experimental |
| 84 |
xiaochus/DeepModelDeploy
Deploy deep learning model on difference hardware and framework.... |
|
Experimental |
| 85 |
victorbadenas/SimpleOnnxInference
A simple c++ onnx application for inferencing images |
|
Experimental |
| 86 |
loong64/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training... |
|
Experimental |
| 87 |
aws-samples/sagemaker-cv-preprocessing-training-performance
SageMaker training implementation for computer vision to offload JPEG... |
|
Experimental |
| 88 |
zpye/SimpleInfer
A simple neural network inference framework |
|
Experimental |
| 89 |
domenicostefani/deep-classf-runtime-wrappers
Classification wrappers for 4 Deep Learning Embedded Runtimes for Music... |
|
Experimental |
| 90 |
Neumenon/cowrie
Multi-language binary JSON codec with tensors, compression, and streaming support. |
|
Experimental |
| 91 |
Orange20000922/Filerestore_CLI
一个 NTFS 文件恢复工具,支持 MFT 扫描、签名搜索、智能覆盖检测和高性能文件恢复 |
|
Experimental |
| 92 |
odeliyach/tiny-ml-runtime
Generic neural network inference engine in pure C. 258x faster than PyTorch... |
|
Experimental |
| 93 |
RyanCCC/Deployment
深度学习应用部署 |
|
Experimental |
| 94 |
brandonviaje/Infera
lightweight inference engine/runtime |
|
Experimental |
| 95 |
Dimitrios-Kafetzis/EdgeNN
Lightweight, zero-allocation C11 library for neural network inference on ARM... |
|
Experimental |
| 96 |
jazpeak/NanoNet
A lightweight, runtime C++ inference engine for running simple ONNX models. |
|
Experimental |
| 97 |
zhehaoxu/deep-deploy
deploy deep learning model with MNN |
|
Experimental |
| 98 |
veera-adithya-d/hardware-aware-algorithm
Inference module of Imagenet |
|
Experimental |
| 99 |
torinos-yt/NNOnnx
Using CUDA for Faster Machine Learning Inference on Unity |
|
Experimental |
| 100 |
DarkStarStrix/Nexa_Inference
A inference application to serve Scientific Models |
|
Experimental |
| 101 |
brandon-gong/siliconnn
Feedforward neural networks in pure ARM64 assembly for Apple Silicon. |
|
Experimental |
| 102 |
regel/onnxruntime_server
An easy way to start an ONNX inferencing server for prediction with GRPC endpoints |
|
Experimental |
| 103 |
MrSyabro/luaort
ONNX Runtime binding for Lua |
|
Experimental |
| 104 |
work-buddy/work-buddy-runtime
A custom Linux based OpenCV C++ runtime for powering the Work Buddy AI XR HUD. |
|
Experimental |
| 105 |
fancyovo/XingLing-cpp
🚀 High-performance C++/CUDA Inference Engine for XingLing-0.68B.... |
|
Experimental |
| 106 |
Konstantina155/InferONNX
Lightweight TEE-based system for secure ONNX model inference using Intel SGX... |
|
Experimental |
| 107 |
mlomb/onnx2code
Convert ONNX models to plain C++ code (without dependencies) |
|
Experimental |
| 108 |
glzbcrt/tf-onnx-sample
Sample Tensorflow model to classify images and infer using ONNX. |
|
Experimental |
| 109 |
HexaForge-1/highperf-ai-ml-inference
High-perf C++ AI/ML inference engine with ONNX Runtime & LibTorch. CPU... |
|
Experimental |
| 110 |
Natfii/onnxruntime-gpu-blackwell
Pre-built onnxruntime-gpu 1.24.1 with Blackwell sm_120 CUDA kernels (RTX... |
|
Experimental |
| 111 |
idsia-robotics/navground_onnx
Provides a C++ navigation behavior that evaluates a ONNX model. |
|
Experimental |
| 112 |
Darth-Kronos/trt-custom-plugins
TensorRT plugins for custom operators |
|
Experimental |
| 113 |
karthi0804/Pytorch-ResNet-CPP-Inference
ResNet based Image classification using Pytorch. |
|
Experimental |
| 114 |
selpoG/ort-sample
A Minimal Sample to use onnxruntime from C++. |
|
Experimental |
| 115 |
HichemMaiza/onnxruntime-extensions-java
Microsoft onnxruntime-extensions builds for Java |
|
Experimental |