Hello,Thank you for the excellent work and publicly available code.
I'm trying to use the mvlrs_v1 datasets to train the Syncnet,but the loss keep oscillating:
Since most video in the mvlrs_v1 is short,I randomly shift the audio up to 10-frame in order to generate synthetic false audio-video pairs.
1、Which datasets are you using?LRW/LRS2/LRS3?
2、Does the false pairs have to be shifted by up to 2s?
3、Can you show me your training log?
thank you!
Hello,Thank you for the excellent work and publicly available code. I'm trying to use the mvlrs_v1 datasets to train the Syncnet,but the loss keep oscillating:
Since most video in the mvlrs_v1 is short,I randomly shift the audio up to 10-frame in order to generate synthetic false audio-video pairs. 1、Which datasets are you using?LRW/LRS2/LRS3? 2、Does the false pairs have to be shifted by up to 2s? 3、Can you show me your training log? thank you!