Closed chen8750 closed 1 year ago
Hi, for alignment of audio and meshes, please refer to https://github.com/facebookresearch/multiface/issues/31
To train Meshtalk with Multiface, get the Multiface tracked meshes and audio, write a dataloader, and use the provided code. There are some hints here: https://github.com/facebookresearch/meshtalk/issues/39
Note that Meshtalk has been trained on 250 identities but we were only able to publish 13. I'd recommend to initialize with the pretrained model or use the pretrained model to generate additional training data. Others have found the training to be unstable if you use the 13 identities from Multiface only.
The mouth_mask, eye_mask, and eye_keypoints have been manually selected from the template mesh.
Thanks for your excellent work! How to use the Multiface dataset with aligned audio and mesh? AND How are the parameters mouth_mask, eye_mask, eye_keypoints obtained?