RBLN-SW/optimum-rbln
⚡ A seamless integration of HuggingFace Transformers & Diffusers with RBLN SDK for efficient inference on RBLN NPUs.
This is a tool for developers working with large language models and image generation models. It allows you to run existing Hugging Face models like Transformers and Diffusers on RBLN Neural Processing Units (NPUs) to achieve faster inference. You input your existing model code and get back a more performant execution on specialized hardware, enabling quicker AI application deployment.
Available on PyPI.
Use this if you are a machine learning engineer or MLOps specialist looking to accelerate the inference of your Hugging Face models on RBLN NPUs without major code overhauls.
Not ideal if you are not working with RBLN hardware or if your primary goal is model training rather than inference optimization.
Stars
15
Forks
3
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Dependencies
7
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/RBLN-SW/optimum-rbln"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
openvinotoolkit/nncf
Neural Network Compression Framework for enhanced OpenVINO™ inference
huggingface/optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers...
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
eole-nlp/eole
Open language modeling toolkit based on PyTorch