hila-chefer/Transformer-MM-Explainability

[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.

39
/ 100
Emerging

Provides attention rollout and perturbation-based attribution methods to generate visual explanations across diverse transformer architectures including vision-language models (LXMERT, VisualBERT, CLIP), object detection (DETR), and pure vision transformers (ViT). The approach computes importance scores by analyzing attention weight propagation through transformer layers and measuring model sensitivity to input perturbations, supporting both unimodal and cross-modal reasoning tasks. Includes ready-to-use Colab notebooks and integrates with MMF, Hugging Face Transformers, and official model implementations for reproducible analysis.

903 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 9 / 25
Community 20 / 25

How are scores calculated?

Stars

903

Forks

115

Language

Jupyter Notebook

License

MIT

Last pushed

Aug 24, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hila-chefer/Transformer-MM-Explainability"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.