AliaksandrSiarohin / pose-gan

383 stars 79 forks source link

improve training #30

Closed SheelaRaju closed 5 years ago

SheelaRaju commented 5 years ago

I have trained the network and see that the texture information is not generated properly and it is blurry. I would like to know for how many epochs did you train the model. What could be still changed to improve the generated images.

AliaksandrSiarohin commented 5 years ago

The number if epochs is 90, for fashion it is probably better to train for longer time. Are you using the this command:

CUDA_VISIBLE_DEVICES=0 python train.py --output_dir output/full --checkpoints_dir output/full --warp_skip mask --dataset fasion --l1_penalty_weight 0.01 --nn_loss_area_size 5 --batch_size 2 --content_loss_layer block1_conv2 --number_of_epochs 90
SheelaRaju commented 5 years ago

Yes I am using the same command. How could the training be improved other than training for longer time.

AliaksandrSiarohin commented 5 years ago

If I know how to significantly improve the performance I would probably do that. Some improvement can be probably obtained by using better HPE model (e.g Mask RCNN), better discriminator (for example discriminator with Spectral Normalization) and by incorporation perceptual and feature matchings losses (from pix2pixHD model).

lijiaqi commented 5 years ago

@AliaksandrSiarohin What is your opinion about the improvement by replacing vanilla GAN loss with WGAN loss ? From my experiences, may be it improves the results in noise-based image generation, but it seems it brings only limited improvements in image translation task, or even no. How do you think about this opinion?

AliaksandrSiarohin commented 5 years ago

I personally trust more in hinge loss with spectral normalization for generation from noise. But for image to image translation I do not think you observe some big difference by changing only the GAN loss. I believe that a main driving component in image2image is reconstruction loss. So the main benefits comes from better generator architecture and better design of this reconstruction loss.

lijiaqi commented 5 years ago

@AliaksandrSiarohin yep, it's the reconstruction loss that dominates img2img generation results.