pytorch/TensorRT

PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

67
/ 100
Established

This project helps machine learning engineers and MLOps professionals accelerate the inference speed of their PyTorch deep learning models on NVIDIA GPUs. It takes an existing PyTorch model and optimizes it using TensorRT, allowing for significantly faster predictions. The output is an optimized model ready for deployment, either within Python or in a C++ environment.

2,955 stars. Actively maintained with 33 commits in the last 30 days.

Use this if you need to drastically reduce the time it takes for your PyTorch models to make predictions on NVIDIA hardware, especially for real-time applications or high-throughput systems.

Not ideal if your models are not deployed on NVIDIA GPUs, or if you are not working with PyTorch models.

deep-learning-deployment model-optimization GPU-acceleration machine-learning-inference MLOps
No Package No Dependents
Maintenance 20 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 21 / 25

How are scores calculated?

Stars

2,955

Forks

384

Language

Python

License

BSD-3-Clause

Last pushed

Mar 13, 2026

Commits (30d)

33

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/pytorch/TensorRT"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.