VIPL-SLP / VAC_CSLR

Visual Alignment Constraint for Continuous Sign Language Recognition. ( ICCV 2021)
https://openaccess.thecvf.com/content/ICCV2021/html/Min_Visual_Alignment_Constraint_for_Continuous_Sign_Language_Recognition_ICCV_2021_paper.html
Apache License 2.0
116 stars 19 forks source link

Finetuning and continue training #23

Closed khoapip closed 1 year ago

khoapip commented 2 years ago

Hello, Thank you for the awesome work. I am trying to use the model on another dataset, so I figure I should structure my data accordingly to the format of phoenix2014. Is there anything else I should worry about or just running the preprocessing with the same structure is gonna be alright?

Also, since I am training on google colab, I won't be able to train for 80 epochs consecutively and plan to split it into several different runs. Is there a built in function to load the previous model and continue training (or finetuning, if I want to finetune the pretrain) or how should I begin to tackle this problem? I am not sure if --load-weights tag is enough. Thank you so much.

ycmin95 commented 2 years ago

Thanks for your attention, If your resolusion of video data is pretty high, perhaps a human detection can preserve more useful information before resizing the whole image. Our recent version can achieve comparable results with 40 epochs, and --load-checkpoints can load the previous model and continue training. Details can be found in config and here.

Good luck~