all-things-vits/code-samples

Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.

37
/ 100
Emerging

This project provides practical code samples and notebooks to understand how Vision Transformers (ViTs) make decisions when classifying images. You input an image and a trained ViT model, and it outputs visual explanations showing which parts of the image the model focused on. This is for researchers and practitioners working with computer vision models, particularly those using attention-based architectures.

197 stars. No commits in the last 6 months.

Use this if you need to interpret why a Vision Transformer made a specific prediction on an image.

Not ideal if you are looking for a general-purpose image classification library or tools for traditional computer vision tasks.

computer-vision model-interpretability AI-explainability image-analysis deep-learning-research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

197

Forks

12

Language

Jupyter Notebook

License

Apache-2.0

Last pushed

Jun 20, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/all-things-vits/code-samples"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.