MCG-NJU / VideoMAE

[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
https://arxiv.org/abs/2203.12602
Other
1.32k stars 133 forks source link

RuntimeError: Unknown model (pretrain_videomae_base_patch16_224) #9

Closed bnbsking closed 2 years ago

bnbsking commented 2 years ago

Hello, thanks for your sharing again. I have downloaded the kinetic_400_vitl_epoch_1600 pretrained weight and trying to do visualization on videos. The vis.sh I run is same as the official except for the path. However, timm raise the error "RuntimeError: Unknown model (pretrain_videomae_base_patch16_224)" while loading the model.

I have tried to modify pretrain_videomae_base_patch16_224 to vit_base_patch16_224, then another error raised "TypeError: init() got an unexpected keyword argument 'decoder_depth'". Subsequently, I comment the keyword "deocder_path" in line80 of run_videomae_viz.py but the error occurs again "AttributeError: 'VisionTransformer' object has no attribute 'encoder'".

I really appreciate if you can help it. Thanks. Screenshot from 2022-05-09 11-43-33 Screenshot from 2022-05-09 11-43-48

yztongzhan commented 2 years ago

Hi @bnbsking! We have fixed this issue by adding import utils and import modeling_pretrain.