PingchuanMa / Temporal-Shift-Module

Unofficial implementation for paper `Temporal Shift Module for Efficient Video Understanding`
58 stars 8 forks source link

Have you reproduced the accuracy on Something dataset? #1

Open nemonameless opened 5 years ago

PingchuanMa commented 5 years ago

It is still on the way. I will publish my results as soon as I make it.

nemonameless commented 5 years ago

@Pika7ma Hi, have you first pre-train on Kinetics ? Otherwise, it seems that the accuracy cannot be reproduced. I just used imagenet-pretrained resnet50 and fixed the bug in https://github.com/Pika7ma/Temporal-Shift-Module/blob/master/tsm_util.py#L14 but only got 41 on something-v1. How about you?

PingchuanMa commented 5 years ago

@nemonameless Yes it should be pre-trained on Kinetics following the steps of original paper.

Haijunlv commented 5 years ago

@Pika7ma did you reproduce kinetics top1 72.5% at 16Frames? I followed tsn based source code(a little diffent at init weight). trained 45 epoch with tran param from paper. But only get 70.436% top1.

PingchuanMa commented 5 years ago

@Haijunlv Actually no. I am currently doing experiments using 8-frames. I'll try it soon thanku for the information.

tzzcl commented 5 years ago

@Pika7ma did you reproduce kinetics top1 72.5% at 16Frames? I followed tsn based source code(a little diffent at init weight). trained 45 epoch with tran param from paper. But only get 70.436% top1.

Hi, I'm interested in that do you use pre-trained resnet50 on ImageNet as init params? In the paper they don't claim for this but I guess it is important.

Haijunlv commented 5 years ago

@tzzcl yes. I use tsn codebase to reproduce. But still have 2% gap with the paper. Maybe pretrained weight is important. But lateset paper by kaiming said pretrained is not necessary. Anyway pretained model save time to converge.

tzzcl commented 5 years ago

@tzzcl yes. I use tsn codebase to reproduce. But still have 2% gap with the paper. Maybe pretrained weight is important. But lateset paper by kaiming said pretrained is not necessary. Anyway pretained model save time to converge.

Hi, I also use the tsn codebase and pretrained resnet50 to reimplement K=8 results on Kinetics. But there is also 2% gap (about 68.8% with my own implementation). Maybe we can only wait for the official code.

gongbudaizhe commented 5 years ago

@nemonameless May I ask what are your configurations to get 41% on something v1 ? e.g. Batch size, learning rate schedule

sKamiJ commented 5 years ago

hi guys, have you reproduced the accuracy on the Something-Something dataset based on the official code at https://github.com/mit-han-lab/temporal-shift-module ?