RBLN-SW/optimum-rbln

⚡ A seamless integration of HuggingFace Transformers & Diffusers with RBLN SDK for efficient inference on RBLN NPUs.

55
/ 100
Established

This is a tool for developers working with large language models and image generation models. It allows you to run existing Hugging Face models like Transformers and Diffusers on RBLN Neural Processing Units (NPUs) to achieve faster inference. You input your existing model code and get back a more performant execution on specialized hardware, enabling quicker AI application deployment.

Available on PyPI.

Use this if you are a machine learning engineer or MLOps specialist looking to accelerate the inference of your Hugging Face models on RBLN NPUs without major code overhauls.

Not ideal if you are not working with RBLN hardware or if your primary goal is model training rather than inference optimization.

AI-inference-optimization machine-learning-engineering model-deployment natural-language-processing computer-vision
Maintenance 10 / 25
Adoption 6 / 25
Maturity 25 / 25
Community 14 / 25

How are scores calculated?

Stars

15

Forks

3

Language

Python

License

Apache-2.0

Last pushed

Mar 13, 2026

Commits (30d)

0

Dependencies

7

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/RBLN-SW/optimum-rbln"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.