jx-zhong-for-academic-purpose / GCN-Anomaly-Detection

Source codes of our paper in CVPR 2019: Graph Convolutional Label Noise Cleaner: Train a Plug-and-play Action Classifier for Anomaly Detection
234 stars 44 forks source link

About the number of video segment #18

Open linshuheng6 opened 5 years ago

linshuheng6 commented 5 years ago

Hi, Zhong, I am confused about the number of segments of training videos.

For C3D, the feature is extracted form every 16 frames and then compress these feature to 32 features for 32 segments, do I understand correctly?

For TSN, due to the complexity of installing caffe and configuring pycaffe, I tried to use the pytorch edition of TSN from tsn-pytorch. In this code, the feature of a short video is extracted from 7 or 9 frames in corresponding 7 or 9 snippets. In other words, only one frame is used from each snippets.

However, the number of feature for a video in this paper is 32. What confuse me is that in feature extracting stage, whether a video is divided into 32 segments first, and then each segments is divided into 7 or 9 snippets as the input of TSN? Is TSN used to extracted feature for each segment in a video?

Thank you for your time!