[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
MIT License
801
stars
107
forks
source link
How to use these examples to visualize the Grounding Dino or GLIP? #41
How to use these examples to visualize the Grounding Dino or GLIP?