PaddlePaddle/FastDeploy

High-performance Inference and Deployment Toolkit for LLMs and VLMs based on PaddlePaddle

73
/ 100
Verified

This tool helps machine learning engineers and AI researchers deploy large language models (LLMs) and vision-language models (VLMs) efficiently. It takes trained PaddlePaddle-based models and optimizes them for high-performance inference, outputting a production-ready deployment solution. You would use this if you need to serve advanced AI models like ERNIE-4.5 or PaddleOCR-VL in real-world applications with speed and reliability.

3,659 stars. Actively maintained with 221 commits in the last 30 days.

Use this if you need to rapidly deploy and serve large language or vision-language AI models from the PaddlePaddle ecosystem, requiring high performance and compatibility with various hardware.

Not ideal if your primary focus is on training new models or if you are not working with PaddlePaddle-based LLMs or VLMs.

AI model deployment Large Language Models Vision-Language Models AI inference optimization Machine Learning Engineering
No Package No Dependents
Maintenance 22 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 25 / 25

How are scores calculated?

Stars

3,659

Forks

720

Language

Python

License

Apache-2.0

Last pushed

Mar 13, 2026

Commits (30d)

221

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/PaddlePaddle/FastDeploy"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.