wuba/dl_inference
通用深度学习推理工具,可在生产环境中快速上线由TensorFlow、PyTorch、Caffe框架训练出的深度学习模型。
This tool helps machine learning engineers or MLOps teams quickly deploy deep learning models trained with TensorFlow, PyTorch, or Caffe into production environments. You provide your trained model files, and it sets up a robust serving system that can handle many requests. It's designed for professionals who need to move their models from development to a live, scalable application.
415 stars. No commits in the last 6 months.
Use this if you need to rapidly deploy and manage deep learning models from various frameworks (TensorFlow, PyTorch, Caffe) for high-volume, real-time inference in a production setting.
Not ideal if you are still in the model development or experimentation phase and do not yet need a production-grade deployment solution.
Stars
415
Forks
121
Language
Java
License
—
Category
Last pushed
Dec 21, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/wuba/dl_inference"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnx/onnx
Open standard for machine learning interoperability
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The...
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This...
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX