NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Provides modular ONNX parser and plugin ecosystem for custom operator support, enabling developers to extend inference capabilities beyond built-in operations. The SDK performs graph optimization, kernel auto-tuning, and precision calibration (transitioning from implicit to explicit quantization) to compile trained models into optimized GPU engines. Integrates with TensorFlow, PyTorch, and ONNX through conversion tools and sample applications, with multi-device inference support via NCCL.
12,784 stars and 213,033 monthly downloads. Used by 2 other packages. Actively maintained with 2 commits in the last 30 days. Available on PyPI.
Stars
12,784
Forks
2,321
Language
C++
License
Apache-2.0
Category
Last pushed
Mar 09, 2026
Monthly downloads
213,033
Commits (30d)
2
Dependencies
1
Reverse dependents
2
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/NVIDIA/TensorRT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related frameworks
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnx/onnx
Open standard for machine learning interoperability
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The...
microsoft/onnxconverter-common
Common utilities for ONNX converters
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX