triton-inference-server/model_navigator
Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.
This tool helps machine learning engineers and MLOps specialists streamline the deployment of deep learning models and pipelines, especially for inference on NVIDIA GPUs. It takes models built in PyTorch, TensorFlow, or ONNX, optimizes them, and outputs highly performant models ready for serving on Triton Inference Server or PyTriton.
218 stars.
Use this if you need to optimize and deploy your deep learning models or entire inference pipelines for maximum performance on NVIDIA GPUs, ensuring correctness and efficiency.
Not ideal if you are not working with deep learning models, do not use NVIDIA GPUs for inference, or are not concerned with optimizing model performance for production deployment.
Stars
218
Forks
28
Language
Python
License
Apache-2.0
Category
Last pushed
Feb 03, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/triton-inference-server/model_navigator"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
gpu-mode/Triton-Puzzles
Puzzles for learning Triton
hailo-ai/hailo_model_zoo
The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment
open-mmlab/mmdeploy
OpenMMLab Model Deployment Framework
hyperai/tvm-cn
TVM Documentation in Chinese Simplified / TVM 中文文档