cyrusbehr/tensorrt-cpp-api

TensorRT C++ API Tutorial

46
/ 100
Emerging

This project helps machine learning engineers and researchers accelerate the performance of their GPU-based machine learning models. You can take an existing ONNX model and generate a highly optimized TensorRT engine file for faster inference. This allows you to deploy models like YOLOv8 or ArcFace on NVIDIA GPUs with significantly reduced inference times, transforming ONNX input into optimized GPU model output.

793 stars. No commits in the last 6 months.

Use this if you need to deploy your computer vision models for high-performance, real-time inference on NVIDIA GPUs using C++.

Not ideal if you are working with dynamic input shapes, require Windows support, or are not comfortable with C++ development and GPU-specific optimizations.

GPU-inference computer-vision real-time-AI model-deployment deep-learning-optimization
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 20 / 25

How are scores calculated?

Stars

793

Forks

103

Language

C++

License

MIT

Last pushed

Nov 15, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/cyrusbehr/tensorrt-cpp-api"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.