Hila Chefer
Home
Publications
Talks
Blog
Contact
Self-Attention
Transformer Interpretability Beyond Attention Visualization
This paper presents an interpretability method for self-attention based models, and specifically for Transformer encoders. The method incorporates LRP and gradients, and achieves SOTA results for ViT, BERT, and DeiT.
Cite
×