The PyTorch code of the TEA Module.
Please refer to TSN repo and TSM repo for the detailed guide of data pre-processing.
A list file is utilized to specify the video data information, including a tuple of extracted video frame folder path (absolute path), video frame number, and video label. A typical line in the file look like:
/data/xxx/xxx/something-something/video_frame_folder 100 12
Finally, the absolute path of your own generated list files should be added into ops/dataset_configs.py
We have provided several examples for training TEA models on different datasets. Please refer to the Appendix B of our paper for more training details.
bash ./scripts/train_tea_something_rgb_8f.sh
bash ./scripts/finetune_tea_hmdb_rgb_16f.sh
Two inference protocols are utilized in our paper: 1) efficient protocol and 2) accuracy protocol. For both protocols we provide the example scripts for testing TEA models:
bash ./scripts/single_test_tea_something_rgb_8f.sh
bash ./scripts/multi_test_tea_something_rgb_8f.sh
Currently, we do not provide the original pre-trained models on STHV1, STHV2, and Kinetics, since we have reorganized the structure of the codes and renamed the modules of TEA for public release. The old models cannot be loaded with new names. We plan to retrain the models with the new codes and release the models for evaluation.
The released codes are verified, and you will get a similar performance with our paper if you have followed the exact training settings of TEA (issue 2 and issue 4).