hila-chefer/Transformer-MM-Explainability
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Provides attention rollout and perturbation-based attribution methods to generate visual explanations across diverse transformer architectures including vision-language models (LXMERT, VisualBERT, CLIP), object detection (DETR), and pure vision transformers (ViT). The approach computes importance scores by analyzing attention weight propagation through transformer layers and measuring model sensitivity to input perturbations, supporting both unimodal and cross-modal reasoning tasks. Includes ready-to-use Colab notebooks and integrates with MMF, Hugging Face Transformers, and official model implementations for reproducible analysis.
903 stars. No commits in the last 6 months.
Stars
903
Forks
115
Language
Jupyter Notebook
License
MIT
Last pushed
Aug 24, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hila-chefer/Transformer-MM-Explainability"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jessevig/bertviz
BertViz: Visualize Attention in Transformer Models
inseq-team/inseq
Interpretability for sequence generation models 🐛 🔍
EleutherAI/knowledge-neurons
A library for finding knowledge neurons in pretrained transformer models.
cdpierse/transformers-interpret
Model explainability that works seamlessly with 🤗 transformers. Explain your transformers model...
taufeeque9/codebook-features
Sparse and discrete interpretability tool for neural networks