TensorRT and onnx-tensorrt

TensorRT is the core inference engine, while ONNX-TensorRT is a plugin that enables TensorRT to execute ONNX model format directly—making them complements that work together.

TensorRT
74
Verified
onnx-tensorrt
63
Established
Maintenance 13/25
Adoption 12/25
Maturity 25/25
Community 24/25
Maintenance 13/25
Adoption 10/25
Maturity 16/25
Community 24/25
Stars: 12,784
Forks: 2,321
Downloads:
Commits (30d): 1
Language: C++
License: Apache-2.0
Stars: 3,194
Forks: 547
Downloads:
Commits (30d): 1
Language: C++
License: Apache-2.0
No risk flags
No Package No Dependents

About TensorRT

NVIDIA/TensorRT

NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.

NVIDIA TensorRT is a toolkit for developers who need to optimize and deploy deep learning models on NVIDIA GPUs for faster performance. It takes trained AI models, typically from frameworks like TensorFlow or ONNX, and processes them to run much more efficiently. This helps bring AI applications to users with minimal delay, making things like real-time image analysis or recommendation systems more responsive.

deep-learning-deployment AI-inference-optimization GPU-acceleration MLOps edge-AI

About onnx-tensorrt

onnx/onnx-tensorrt

ONNX-TensorRT: TensorRT backend for ONNX

This project helps deep learning engineers and AI practitioners take ONNX neural network models and run them efficiently on NVIDIA GPUs using TensorRT. It takes an ONNX model as input and produces an optimized TensorRT engine that executes deep learning inferences at high speed. This tool is for those who need to deploy and accelerate AI models in production environments.

AI deployment deep learning inference model optimization GPU acceleration machine learning engineering

Scores updated daily from GitHub, PyPI, and npm data. How scores work