Closed niladridutt closed 3 years ago
You need to get synchronized image, audio, and landmark data for your own dataset. Details to generate the .t7 file can be found in data/AnnScripts/script_AnnVI.py
.
Thank you so much for your reply and uploading the scripts! What do you mean by "synchronized audio, image, pose, landmark, and blink information" in the paper? And how do I get it (for example dlib 68 shape predictor?)? It'll be really helpful if you could explain what is synced images. Thanks!
https://github.com/davisking/dlib
or https://www.faceplusplus.com.cn/face-detection
Thanks! Do you have any script that generates the landmarks of the images and cropped images in the format needed by script_AnnVI.py ? It'll be very helpful if you could upload them as well. Essentially if you have a script that generates landmark files from the video it'll be great
I'm sorry, but you can refer to str_landmark to create the custom dataset that matches script_AnnVI.py
.
Right now the dataloader can only parse ANNVii dataset it seems. How can I use it on custom videos? How to get the .t7 file?