Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/jacobgil/vit-explain

Explainability for Vision Transformers
https://github.com/jacobgil/vit-explain

deep-learning explainable-ai pytorch transformer vision-transformer

Last synced: about 6 hours ago
JSON representation

Explainability for Vision Transformers

Awesome Lists containing this project

README

        

# Explainability for Vision Transformers (in PyTorch)

This repository implements methods for explainability in Vision Transformers.

See also https://jacobgil.github.io/deeplearning/vision-transformer-explainability

## Currently implemented:
- Attention Rollout.
- Gradient Attention Rollout for class specific explainability.
*This is our attempt to further build upon and improve Attention Rollout.*

- TBD Attention flow is work in progress.

Includes some tweaks and tricks to get it working:
- Different Attention Head fusion methods,
- Removing the lowest attentions.

## Usage

- From code
``` python
from vit_grad_rollout import VITAttentionGradRollout

model = torch.hub.load('facebookresearch/deit:main',
'deit_tiny_patch16_224', pretrained=True)
grad_rollout = VITAttentionGradRollout(model, discard_ratio=0.9, head_fusion='max')
mask = grad_rollout(input_tensor, category_index=243)

```

- From the command line:

```
python vit_explain.py --image_path --head_fusion --discard_ratio --category_index
```
If category_index isn't specified, Attention Rollout will be used,
otherwise Gradient Attention Rollout will be used.

Notice that by default, this uses the 'Tiny' model from [Training data-efficient image transformers & distillation through attention](https://arxiv.org/abs/2012.12877)
hosted on torch hub.

## Where did the Transformer pay attention to in this image?

| Image | Vanilla Attention Rollout | With discard_ratio+max fusion |
| -------------------------|-------------------------|------------------------- |
| ![](examples/both.png) | ![](examples/both_attention_rollout_0.000_mean.png) | ![](examples/both_attention_rollout_0.990_max.png)
![](examples/plane.png) | ![](examples/plane_attention_rollout_0.000_mean.png) | ![](examples/plane_attention_rollout_0.900_max.png) |
![](examples/dogbird.png) | ![](examples/dogbird_attention_rollout_0.000_mean.png) | ![](examples/dogbird_attention_rollout_0.900_max.png) |
![](examples/plane2.png) | ![](examples/plane2_attention_rollout_0.000_mean.png) | ![](examples/plane2_attention_rollout_0.900_max.png) |

## Gradient Attention Rollout for class specific explainability

The Attention that flows in the transformer passes along information belonging to different classes.
Gradient roll out lets us see what locations the network paid attention too,
but it tells us nothing about if it ended up using those locations for the final classification.

We can multiply the attention with the gradient of the target class output, and take the average among the attention heads (while masking out negative attentions) to keep only attention that contributes to the target category (or categories).

### Where does the Transformer see a Dog (category 243), and a Cat (category 282)?
![](examples/both_grad_rollout_243_0.900_max.png) ![](examples/both_grad_rollout_282_0.900_max.png)

### Where does the Transformer see a Musket dog (category 161) and a Parrot (category 87):
![](examples/dogbird_grad_rollout_161_0.900_max.png) ![](examples/dogbird_grad_rollout_87_0.900_max.png)

## Tricks and Tweaks to get this working

### Filtering the lowest attentions in every layer

`--discard_ratio `

Removes noise by keeping the strongest attentions.

Results for dIfferent values:

![](examples/both_discard_ratio.gif) ![](examples/plane_discard_ratio.gif)

### Different Attention Head Fusions

The Attention Rollout method suggests taking the average attention accross the attention heads,

but emperically it looks like taking the Minimum value, Or the Maximum value combined with --discard_ratio, works better.

` --head_fusion `

| Image | Mean Fusion | Min Fusion |
| -------------------------|-------------------------|------------------------- |
![](examples/both.png) | ![](examples/both_attention_rollout_0.000_mean.png) | ![](examples/both_attention_rollout_0.000_min.png)

## References
- [Quantifying Attention Flow in Transformers](https://arxiv.org/abs/2005.00928)
- [timm: a great collection of models in PyTorch](https://github.com/rwightman/pytorch-image-models)
and especially [the vision transformer implementation](https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py)

- [An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale](https://arxiv.org/abs/2010.11929)
- Credit for https://github.com/jeonsworld/ViT-pytorch for being a good starting point.

## Requirements
`pip install timm`