Open XiaoHao-Chen opened 4 years ago
This is a WGAN, so no cross-entropy loss is used. You can find more details regarding the training scheme of WGAN at the paper "Improved Training of Wasserstein GANs" by Gulrajani et. al: https://arxiv.org/pdf/1704.00028.pdf (Generally, we follow this)
Why is the output of patch-d the average value of all patches, rather than the cross entropy of biclassification calculated by BCELoss, like the common patch-gan?