quic/efficient-transformers

This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transformers library) into inference-ready formats that run efficiently on Qualcomm Cloud AI 100 accelerators.

58
/ 100
Established

This library helps AI developers and machine learning engineers take pre-trained AI models from the HuggingFace hub and make them run very efficiently on Qualcomm Cloud AI 100 hardware. It takes models like large language models, vision models, or audio models and converts them into an optimized format for high-performance inference. This is for AI practitioners deploying models in production on Qualcomm's cloud accelerators.

Use this if you need to deploy various types of large AI models, including text, image, and audio, for efficient and high-performance inference on Qualcomm Cloud AI 100 accelerators.

Not ideal if you are not working with Qualcomm Cloud AI 100 hardware or if you only need to train models rather than optimize them for deployment.

AI deployment machine learning inference model optimization cloud AI accelerators large language models
No Package No Dependents
Maintenance 10 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 23 / 25

How are scores calculated?

Stars

87

Forks

75

Language

Python

License

Last pushed

Mar 13, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/quic/efficient-transformers"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.