sayakpaul/probing-vits
Probing the representations of Vision Transformers.
This project offers tools to understand how Vision Transformer (ViT) models 'see' and process images or videos. By taking your image or video input, it generates visualizations like attention maps and heatmaps, showing which parts of the input the model focused on. This helps researchers and AI practitioners analyze and debug the internal workings of various ViT architectures.
340 stars. No commits in the last 6 months.
Use this if you are a machine learning researcher or practitioner who wants to visualize and interpret the internal representations of different Vision Transformer models, particularly for understanding their attention mechanisms on images or videos.
Not ideal if you are looking for novel methods for probing neural networks or need to train and visualize ViTs with very small datasets, as these features are not the focus or are still under development.
Stars
340
Forks
22
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Oct 05, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/sayakpaul/probing-vits"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jaehyunnn/ViTPose_pytorch
An unofficial implementation of ViTPose [Y. Xu et al., 2022]
UdbhavPrasad072300/Transformer-Implementations
Library - Vanilla, ViT, DeiT, BERT, GPT
tintn/vision-transformer-from-scratch
A Simplified PyTorch Implementation of Vision Transformer (ViT)
icon-lab/ResViT
Official Implementation of ResViT: Residual Vision Transformers for Multi-modal Medical Image Synthesis
gupta-abhay/pytorch-vit
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale