Thank you for providing your paper's code.
I'm quite curious about how I will use MINE in my code.
I add the output of the MINE torch model to the adversarial loss to train the generator (with Wasserstein loss).
But my issue is that I either receive -inf (if I don't utilize gradient normalization) or values that are so near to zero (1e-5), implying that the generator output and the real image are utterly misaligned (am I right?)
And this small loss will never differ from zero.
Do you have any ideas about why this is occurring? Should Increase the complexity of the MINE network or iterate more on MINE loss?
Hi,
Thank you for providing your paper's code. I'm quite curious about how I will use MINE in my code. I add the output of the MINE torch model to the adversarial loss to train the generator (with Wasserstein loss). But my issue is that I either receive -inf (if I don't utilize gradient normalization) or values that are so near to zero (1e-5), implying that the generator output and the real image are utterly misaligned (am I right?) And this small loss will never differ from zero. Do you have any ideas about why this is occurring? Should Increase the complexity of the MINE network or iterate more on MINE loss?