haoyanbin918 / Attention-in-Attention

Apache License 2.0
10 stars 0 forks source link

Attention in Attention: Modeling Context Correlation for Efficient Video Classification (IEEE TCVST 2022)

This is an official implementaion of paper "Attention in Attention: Modeling Context Correlation for Efficient Video Classification", which has been accepted by IEEE TCVST 2022. Paper link

Updates

Apr 20, 2022

Content

Prerequisites

The code is built with following libraries:

For video data pre-processing, you may need ffmpeg.

Data Preparation

Code

Pretrained Models

Here we provide some of the pretrained models.

Something-Something

Something-Something-V1

Model Frame view clip Top-1 Acc. Top-5 Acc. Checkpoint
AIA(TSN) ResNet50 8 1 1 48.5% 77.2% [link]()

Something-Something-V2

Model Frame view clip Top-1 Acc. Top-5 Acc. Checkpoint
AIA(TSN) ResNet50 8 1 1 60.3% 86.4% [link]()

Diving48

Model Frame view clip Top-1 Acc. Checkpoint
AIA(TSN) ResNet50 8 1 1 79.3% [link]()
AIA(TSM) ResNet50 8 1 1 79.4% [link]()

EGTEA Gaze

Model Frame view clip Split1 Split2 Split3
AIA(TSN) ResNet50 8 1 1 63.7% 62.1% 61.5%
AIA(TSN) ResNet50 8 1 1 64.7% 63.3% 62.2%

Train

Test

Contributors

GC codes are jointly written and owned by Dr. Yanbin Hao and [Dr. Shuo Wang].

Citing

Acknowledgement

Thanks for the following Github projects: