PaddlePaddle/Serving

A flexible, high-performance carrier for machine learning models(『飞桨』服务化部署框架)

61
/ 100
Established

This project helps machine learning engineers and MLOps specialists deploy trained AI models into live applications. It takes your pre-trained PaddlePaddle, TensorFlow, or PyTorch models and packages them into a high-performance, scalable service. The output is a robust, always-on AI service ready to integrate with your products, supporting tasks like image classification, object detection, natural language processing, and recommendation systems.

925 stars.

Use this if you need to transform your deep learning models into production-ready, high-performance, and scalable inference services that can handle real-time requests from end-user applications.

Not ideal if you are looking for a tool to train new machine learning models or if your primary need is for local, offline model inference rather than a networked service.

MLOps model deployment AI inference production AI real-time predictions
No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 25 / 25

How are scores calculated?

Stars

925

Forks

250

Language

C++

License

Apache-2.0

Last pushed

Feb 20, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mlops/PaddlePaddle/Serving"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.