ttengwang / PDVC

End-to-End Dense Video Captioning with Parallel Decoding (ICCV 2021)
MIT License
200 stars 23 forks source link

How to further train on our own dataset? #24

Closed pranavkamath14 closed 2 years ago

pranavkamath14 commented 2 years ago

Hello, Thank you so much for this amazing github repo. I wanted to use the pre-trained model and further train on my own dataset of videos and corresponding captions that I have. Do you have any suggestions on how I can do this?

qt2139 commented 2 years ago

same question

ttengwang commented 2 years ago

1) prepare the video features (.npy files) following TSP and convert_tsp_h5_to_npy.py. 2) construct the follow annoations of your dataset. train_caption_file, val_caption_file, visual_feature_folder, gt_file_for_eval, gt_file_for_para_eval, dict_file 3) To further finetune the pre-trained PDVC (TSP), run python train.py --cfg_path=anet_tsp_pdvc.yml --pretrain=full --pretrain_path=path/to/ckpt