This repo holds the codes of paper: "G-TAD: Sub-Graph Localization for Temporal Action Detection", accepted in CVPR 2020.
15 Dec 2020: The configuration for HACS Segment dataset is in the hacs
branch. With the officail I3D pretrained features, G-TAD can reach 27.481 Average mAP without tuning the model architecture.
24 Nov 2020: to celebrate my 2nd anniversary with Sally,
I released the code for ActivityNet. :P Please checkout the branch anet
to see the details.
Feature: GooogleDrive, md5sum: 0ce54748883c4ce1cf6600f5ad04421b
.
30 Mar 2020: THUMOS14 feature is available! GooogleDrive, OneDrive
15 Apr 2020: THUMOS14 code is published! I update the post processing code so the experimental result is slightly better than the orignal paper!
29 Apr 2020: We updated our code based on @Phoenix1327's comment. The experimental result is slightly better. Please see details in this issue.
Temporal action detection is a fundamental yet challenging task in video understanding. Video context is a critical cue to effectively detect actions, but current works mainly focus on temporal context, while neglecting semantic context as well as other important context properties. In this work, we propose a graph convolutional network (GCN) model to adaptively incorporate multi-level semantic context into video features and cast temporal action detection as a sub-graph localization problem. Specifically, we formulate video snippets as graph nodes, snippet-snippet correlations as edges, and actions associated with context as target sub-graphs. With graph convolution as the basic operation, we design a GCN block called GCNeXt, which learns the features of each node by aggregating its context and dynamically updates the edges in the graph. To localize each sub-graph, we also design a SGAlign layer to embed each sub-graph into the Euclidean space. Extensive experiments show that G-TAD is capable of finding effective video context without extra supervision and achieves state-of-the-art performance on two detection benchmarks. On ActityNet-1.3, we obtain an average mAP of 34.09%; on THUMOS14, we obtain 40.16% in mAP@0.5, beating all the other one-stage methods.
Based on the idea of ROI Alignment from Mask-RCNN, we devoloped SGAlign layer in our implementation. You have to compile a short cuda code to run Algorithm 1 in our paper.
conda env create -f env.yml
source activate gtad
Align1D2.2.0
cd gtad_lib
python setup.py install
Align1D2.2.0
python align.py
To reproduce the results in THUMOS14 without further changes:
Download the data from GooogleDrive or OneDrive.
Place it into a folder named TSN_pretrain_avepool_allfrms_hdf5
inside data/thumos_feature
.
You could also pass the folder containing the HDF5 files if the script admits the following argument
--feature_path
.
gtad # this repo
├── data # feature and label
├── evaluation # evaluation code from offical API
├── gtad_lib # gtad library
└── ...
After downloading the dataset and setting up the envirionment, you can start from the following script.
python gtad_train.py
python gtad_inference.py
python gtad_postprocessing.py
or
bash gtad_thumos.sh | tee log.txt
If everything goes well, you can get the following result:
mAP at tIoU 0.3 is 0.5731204387052588
mAP at tIoU 0.4 is 0.5129888769308306
mAP at tIoU 0.5 is 0.43043083034478025
mAP at tIoU 0.6 is 0.32653130678508374
mAP at tIoU 0.7 is 0.22806267480976325
CVPR Version.
@InProceedings{xu2020gtad,
author = {Xu, Mengmeng and Zhao, Chen and Rojas, David S. and Thabet, Ali and Ghanem, Bernard},
title = {G-TAD: Sub-Graph Localization for Temporal Action Detection},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}
Those are very helpful and promising implementations for the temporal action localization task. My implementations borrow ideas from them.
BSN: Boundary Sensitive Network for Temporal Action Proposal Generation. Paper Code
BMN: BMN: Boundary-Matching Network for Temporal Action Proposal Generation. Paper Code - PaddlePaddle Code PyTorch
Graph Convolutional Networks for Temporal Action Localization. Paper Code
mengmeng.xu[at]kaust.edu.sa