onnx/onnx-tensorrt

ONNX-TensorRT: TensorRT backend for ONNX

63
/ 100
Established

This project helps deep learning engineers and AI practitioners take ONNX neural network models and run them efficiently on NVIDIA GPUs using TensorRT. It takes an ONNX model as input and produces an optimized TensorRT engine that executes deep learning inferences at high speed. This tool is for those who need to deploy and accelerate AI models in production environments.

3,194 stars. Actively maintained with 1 commit in the last 30 days.

Use this if you need to convert and optimize your ONNX deep learning models to run with maximum performance on NVIDIA TensorRT-enabled hardware.

Not ideal if you are developing models in frameworks other than ONNX or do not have access to NVIDIA GPUs and TensorRT.

AI deployment deep learning inference model optimization GPU acceleration machine learning engineering
No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 24 / 25

How are scores calculated?

Stars

3,194

Forks

547

Language

C++

License

Apache-2.0

Last pushed

Feb 03, 2026

Commits (30d)

1

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/onnx/onnx-tensorrt"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.