kibae/onnxruntime-server
ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP and HTTP/HTTPS REST APIs for ONNX inference.
This tool helps machine learning engineers and MLOps professionals quickly deploy trained machine learning models to be accessed via standard web requests or TCP. You provide your models in ONNX format, and it creates a server that exposes APIs to send data for inference and receive predictions. This allows developers to integrate your models into applications without needing to write complex server-side code.
181 stars.
Use this if you need a high-performance, straightforward way to serve ONNX models as APIs for integration into web applications or other services.
Not ideal if your models are not in ONNX format or if you require advanced model serving features like A/B testing, canary deployments, or model ensembles built directly into the serving layer.
Stars
181
Forks
15
Language
C++
License
MIT
Category
Last pushed
Mar 06, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kibae/onnxruntime-server"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnx/onnx
Open standard for machine learning interoperability
PINTO0309/onnx2tf
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The...
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This...
onnx/onnxmltools
ONNXMLTools enables conversion of models to ONNX