TensorRT and onnx-tensorrt
TensorRT is the core inference engine, while ONNX-TensorRT is a plugin that enables TensorRT to execute ONNX model format directly—making them complements that work together.
About TensorRT
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
NVIDIA TensorRT is a toolkit for developers who need to optimize and deploy deep learning models on NVIDIA GPUs for faster performance. It takes trained AI models, typically from frameworks like TensorFlow or ONNX, and processes them to run much more efficiently. This helps bring AI applications to users with minimal delay, making things like real-time image analysis or recommendation systems more responsive.
About onnx-tensorrt
onnx/onnx-tensorrt
ONNX-TensorRT: TensorRT backend for ONNX
This project helps deep learning engineers and AI practitioners take ONNX neural network models and run them efficiently on NVIDIA GPUs using TensorRT. It takes an ONNX model as input and produces an optimized TensorRT engine that executes deep learning inferences at high speed. This tool is for those who need to deploy and accelerate AI models in production environments.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work