zhanglonghao1992 / One-Shot_Free-View_Neural_Talking_Head_Synthesis

Pytorch implementation of paper "One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing"
Other
764 stars 143 forks source link

share log.txt #14

Closed romanvey closed 2 years ago

romanvey commented 2 years ago

@zhanglonghao1992 could you provide your log.txt from your training? I think it would be beneficial for people to understand are they going in the right direction or not.

Vijayue commented 2 years ago

ye! I hope so, the training process is really hard. 😢😢

zhanglonghao1992 commented 2 years ago

@romanvey @Vijayue I'll release the checkpoint and the training log file (if I can still find it - -!).

romanvey commented 2 years ago

Thank you!

vinayak015 commented 2 years ago

00000119-rec log.txt

I was able to train for an epoch taking @zhanglonghao1992 checkpoints and the losses and the visualisation looks like this. The expressions of driving is not matching with the source. I think, either expression network or generator needs more tuning.

I was able to train with the 2D flow and 2D features, but getting similar results

zhanglonghao1992 commented 2 years ago

log.txt @vinayak015 @romanvey @Vijayue

vinayak015 commented 2 years ago

Thank you @zhanglonghao1992, In FOMM issues they mention the perceptual should be around 80 (although you reached 92). I think we should just train for more epochs.

zhanglonghao1992 commented 2 years ago

@vinayak015 For now, my perceptual loss can only converge to about 90.

romanvey commented 2 years ago

Thank you @zhanglonghao1992 ! Yeah, I had the same perceptual loss and longer training helped me to increase quality.

vinayak015 commented 2 years ago

It's mentioned here: https://github.com/AliaksandrSiarohin/first-order-model/issues/460#issuecomment-897689413