p0p4k / pflowtts_pytorch

Unofficial implementation of NVIDIA P-Flow TTS paper
https://neurips.cc/virtual/2023/poster/69899
MIT License
214 stars 30 forks source link

Prior or encoder Loss #18

Closed epii2zero closed 9 months ago

epii2zero commented 9 months ago

Thank you for nice work!

I have a question about encoder loss In paper they used MSE for encoder loss (section 3.1) image

but in this implementation, use different loss https://github.com/p0p4k/pflowtts_pytorch/blob/297166f518263b2822f52918f86f7d27fa72fe8b/pflow/models/pflow_tts.py#L179C9-L180C74

I think this loss came from Matcha-TTS codes Is this intentional?

Thank you!

p0p4k commented 9 months ago

Essentially it is the same loss, scaled. (at a constant variance of 1, gaussian nle becomes mse)

epii2zero commented 9 months ago

Oh I see, thank you for reply