Closed bva1986 closed 5 years ago
yes, your loss function looks right
Should we backpropogate Encoder->Decoder->Encoder or just Encoder when minimize loss C1r-C1 ?
C1r = G(X1r[:, 0, :, :], S1, None)
L_content = criterion_content(C1r, C1)
or
C1r = G(X1r[:, 0, :, :].detach(), S1, None)
L_content = criterion_content(C1r, C1)
or even
C1r = G(X1r[:, 0, :, :].detach(), S1, None)
L_content = criterion_content(C1r, C1.detach())
The 1st one
According to AutoVC presentation on ICML-2019 loss function is: L = Lrecon + lambda Lcontent. But according to paper https://arxiv.org/abs/1905.05879 loss function is L = Lrecon + mu Lrecon0 + lambda*Lcontent. Which of them is more preferable?
The 2nd one
Ok, thanks
Hello
To reproduce results of AutoVC we used following loss function implementation on PyTorch
But we did not achieved comparable voice quality and loss value around 1e-3 according issue whats your final loss and final learning rate? One of supposed reason is that we used inapropriate loss function implementation. Is our loss function implementation compatible with that you used?
BR