triton-inference-server/dali_backend
The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.
This tool helps machine learning engineers accelerate the data preparation stage for deep learning models, especially during inference. It takes raw input data, like images or sensor readings, and efficiently processes it using GPU-accelerated pipelines. The output is pre-processed data ready for your deep learning model, speeding up the overall application performance.
141 stars.
Use this if you need to significantly speed up the data pre-processing step for your deep learning inference applications, especially when dealing with large volumes of data like images or video.
Not ideal if your deep learning models do not require intensive data pre-processing or if your inference server does not use NVIDIA GPUs.
Stars
141
Forks
35
Language
C++
License
MIT
Category
Last pushed
Mar 10, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/triton-inference-server/dali_backend"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related frameworks
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnx/onnx
Open standard for machine learning interoperability
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The...
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This...
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX