ai-dynamo/aitune
NVIDIA AITune is an inference toolkit designed for tuning and deploying Deep Learning models with a focus on NVIDIA GPUs.
This toolkit helps deep learning engineers and AI practitioners make their trained models run much faster and more efficiently on NVIDIA GPUs. It takes your existing PyTorch models or entire AI pipelines and processes them, producing optimized models ready for high-performance deployment across various AI applications like computer vision or generative AI.
Used by 1 other package. Available on PyPI.
Use this if you are a deep learning engineer looking to significantly speed up the inference of your PyTorch models on NVIDIA GPUs with minimal code changes.
Not ideal if your model is already supported by a specialized serving framework that offers advanced runtime optimizations like continuous batching or speculative decoding.
Stars
8
Forks
—
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Dependencies
15
Reverse dependents
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/ai-dynamo/aitune"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
gpu-mode/Triton-Puzzles
Puzzles for learning Triton
hailo-ai/hailo_model_zoo
The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment
open-mmlab/mmdeploy
OpenMMLab Model Deployment Framework
hyperai/tvm-cn
TVM Documentation in Chinese Simplified / TVM 中文文档