Open yhgon opened 6 years ago
I've modified the code to use BatchNormalization with batch(32) instead of Instance Normalization
below is my code the difference for generator.
d = BatchNormalization()(d) #d = InstanceNormalization()(d)
u = BatchNormalization()(u) # u = InstanceNormalization()(u)
for discriminator
if normalization:
d = BatchNormalization()(d) # d = InstanceNormalization()(d)
return d
I've modified the code to use BatchNormalization with batch(32) instead of Instance Normalization
below is my code the difference for generator.
d = BatchNormalization()(d) #d = InstanceNormalization()(d) u = BatchNormalization()(u) # u = InstanceNormalization()(u)
for discriminator
if normalization: d = BatchNormalization()(d) # d = InstanceNormalization()(d) return d
Please have you found the reason?
Did you see NaN in the training? configuration tensorflow : 1.11.0 keras : 2.1.6 GPU : K80 model : CycleGAN dataset : apple2orange
I've test multiple models. aggan, wgan, dcgan , pix2pix works well in 1,000 epochs in my environment. when I try to test cycleGAN, it train well.. but I got NaN in epoch 67.
I've modified the code to use BatchNormalization with batch(32) instead of Instance Normalization
gan.train(epochs=200, batch_size=64, sample_interval=15)
in Epoch 66, I got the result