aliemo/transfomers-silicon-research
Research and Materials on Hardware implementation of Transformer Model
This project offers a curated collection of research papers focused on the hardware implementation of Transformer models, particularly BERT. It provides researchers and engineers with a centralized resource to understand how these powerful AI models are optimized for physical silicon. You'll find papers detailing advancements in making these models run more efficiently on specialized hardware like FPGAs and GPUs.
299 stars. No commits in the last 6 months.
Use this if you are a hardware architect, a machine learning researcher specializing in model deployment, or an electrical engineer looking for papers on implementing AI models on custom silicon.
Not ideal if you are looking for ready-to-use software libraries, pre-trained models, or basic tutorials on using Transformer models for natural language processing tasks.
Stars
299
Forks
38
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Feb 28, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/aliemo/transfomers-silicon-research"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/transformers-bloom-inference
Fast Inference Solutions for BLOOM
Tencent/TurboTransformers
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc)...
mit-han-lab/lite-transformer
[ICLR 2020] Lite Transformer with Long-Short Range Attention
mit-han-lab/hardware-aware-transformers
[ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing
LibreTranslate/Locomotive
Toolkit for training/converting LibreTranslate compatible language models 🚂