Awesome
Attention in Attention: Modeling Context Correlation for Efficient Video Classification (IEEE TCVST 2022)
This is an official implementaion of paper "Attention in Attention: Modeling Context Correlation for Efficient Video Classification", which has been accepted by IEEE TCVST 2022. Paper link
Updates
Apr 20, 2022
- Release this V1 version (the version used in paper) to public. Complete codes and models will be released soon.
Content
Prerequisites
The code is built with following libraries:
- PyTorch >= 1.7, torchvision
- tensorboardx
For video data pre-processing, you may need ffmpeg.
Data Preparation
Code
Pretrained Models
Here we provide some of the pretrained models.
Something-Something
Something-Something-V1
Model | Frame * view * clip | Top-1 Acc. | Top-5 Acc. | Checkpoint |
---|---|---|---|---|
AIA(TSN) ResNet50 | 8 * 1 * 1 | 48.5% | 77.2% | link |
Something-Something-V2
Model | Frame * view * clip | Top-1 Acc. | Top-5 Acc. | Checkpoint |
---|---|---|---|---|
AIA(TSN) ResNet50 | 8 * 1 * 1 | 60.3% | 86.4% | link |
Diving48
Model | Frame * view * clip | Top-1 Acc. | Checkpoint |
---|---|---|---|
AIA(TSN) ResNet50 | 8 * 1 * 1 | 79.3% | link |
AIA(TSM) ResNet50 | 8 * 1 * 1 | 79.4% | link |
EGTEA Gaze
Model | Frame * view * clip | Split1 | Split2 | Split3 |
---|---|---|---|---|
AIA(TSN) ResNet50 | 8 * 1 * 1 | 63.7% | 62.1% | 61.5% |
AIA(TSN) ResNet50 | 8 * 1 * 1 | 64.7% | 63.3% | 62.2% |
Train
Test
Contributors
GC codes are jointly written and owned by Dr. Yanbin Hao and [Dr. Shuo Wang].
Citing
Acknowledgement
Thanks for the following Github projects: