huggingface/optimum-habana
Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)
This project helps machine learning engineers accelerate the training and inference of large language models and diffusion models, such as those from the Hugging Face Transformers and Diffusers libraries. It takes existing model code and configuration and outputs significantly faster computations by leveraging Intel Gaudi AI Accelerators. This is for machine learning practitioners and researchers working with large-scale models who need to optimize performance on specific hardware.
207 stars.
Use this if you are a machine learning engineer working with Hugging Face models and have access to Intel Gaudi AI Accelerators, and you need to significantly speed up your model training or inference processes.
Not ideal if you do not use Intel Gaudi AI Accelerators or if your workflow does not involve models from the Hugging Face Transformers or Diffusers libraries.
Stars
207
Forks
270
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/huggingface/optimum-habana"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
openvinotoolkit/nncf
Neural Network Compression Framework for enhanced OpenVINOâ„¢ inference
huggingface/optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers...
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
eole-nlp/eole
Open language modeling toolkit based on PyTorch