onnx/onnx-tensorrt
ONNX-TensorRT: TensorRT backend for ONNX
This project helps deep learning engineers and AI practitioners take ONNX neural network models and run them efficiently on NVIDIA GPUs using TensorRT. It takes an ONNX model as input and produces an optimized TensorRT engine that executes deep learning inferences at high speed. This tool is for those who need to deploy and accelerate AI models in production environments.
3,194 stars. Actively maintained with 1 commit in the last 30 days.
Use this if you need to convert and optimize your ONNX deep learning models to run with maximum performance on NVIDIA TensorRT-enabled hardware.
Not ideal if you are developing models in frameworks other than ONNX or do not have access to NVIDIA GPUs and TensorRT.
Stars
3,194
Forks
547
Language
C++
License
Apache-2.0
Category
Last pushed
Feb 03, 2026
Commits (30d)
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/onnx/onnx-tensorrt"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnx/onnx
Open standard for machine learning interoperability
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The...
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This...
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX