Hangz-nju-cuhk / Talking-Face-Generation-DAVS

Code for Talking Face Generation by Adversarially Disentangled Audio-Visual Representation (AAAI 2019)
MIT License
817 stars 173 forks source link

Questions about pretraining process and small errors in train.py #37

Closed jixinya closed 4 years ago

jixinya commented 4 years ago

Hi, firstly I want to thank you for sharing such a great project. However, I noticed that you wrote 'The pretraining training code is temporarily not provided.' in README.md, so I was wondering if my understanding is right about the classification part. Here is my own assumption:

  1. Use the subset of the MS-Celeb-1M dataset to train the ID_encoder part.
  2. Use the optimize_parameters_no_generation() function in Gen_final_v1.py and LRW dataset to train the lip_feature_encoder, mfcc_encoder and model_fusion part. Moreover, when I read and try to train the model using train.py, I find some small errors. For example , opt.isTrain and opt.eval_freq are not defined in Options.py and pair in lip_reading_loader() should be (2,25), since there are only 24 files in /mfcc20. So I want to know if you will update the project later which will be of a great help to me.
Hangz-nju-cuhk commented 4 years ago

I figure I have answered this issue until I see the Github ID :) so I am just gonna close it.