Open ajaykmr2292 opened 5 years ago
train
and test
directory. I also added a new script scripts/test_colorization.sh
. resnet_9blocks
and unet_256
architecture. I am new to GANs. I have this query: If we cannot interpret anything from loss functions and suppose the generated images are not good, how can we understand what parameters (or model architecture) to change and check?
Loss function can be used to identify failure mode as suggested by ganhacks. For example, If D loss is always 0, maybe D is too strong. You can increase the capacity of G.
When the binary input image used for training is sent to the generator (after training for 500 epochs), the result is like this: Input Binary Image Real colored image Generated color image
And for test set, the output is like this: Input binary image Generated color image
As it can be seen, the generated image for both train and test set are worse. The loss seems to be normal but the generator still didnt learn anything useful (and it didnt overfit the training data even after 500 epochs as can be seen from the generated image for the training input image). Can you suggest anything for this purpose? I feel a single input image has too many features for the generator to learn while training. So, is it better to use a much complex architecture like inception module (GoogLeNet architecture)? Please suggest some other alternatives to try if any.
I feel this task is an alternative to edges2shoes conversion except the fact that the input image has too many features to learn at a time. Please tell me whether I am right and also suggest something to try.
D loss goes to 0: failure mode
Is it failure mode for my loss curve for the discriminator?Also, inorder to understand the concept indepth, I tried to train the pix2pix model on edges2handbags dataset - 138k images (downloaded using your code)..after training for 15 epochs with a batch size of 32, the generated image is like this: Can you tell me whats going wrong? How did the paper get so good images? I am just running the same code. Thanks in advance.
Thanks for your patient reply. But then, there was a case where the model was trained to watercolor the black and white image (a church kind of an input image..an issue was raised here about that. Issue) which used just 2 training images but after training, it was found to work. Can I know why then its not working in my case?
Also, for me, 256x256 input images are not overfitting whereas for 512x512 images are found to overfit (they are found to be colored exactly like the training image)..can I conclude that this is because there are too many small features in the image which the 3x3 filters used in the generator cannot recognize when 256x256 image is used while they can recognize the features when 512x512 images are used?
Also, I am learning to train GANS (training deep networks for that matter) via DIY. I would like to know how to go about training GANs. Are these tweaking parameters, loss functions and then checking the output completely based on trial-and-error or does there exist any logic behind trying every change? Please help me with some tips on how to go about training GANs (or any deep networks).
--load_size 600 --crop_size 512
. )@ajaykmr2292 Can you tell me where you get the one piece dataset? I've been looking all over the manga sites but cannot find paired color/uncolor version
请问可以提供损失曲线图的代码吗
colorization
dataset-mode. I followed the steps inNotes for colorization
intips.md
and trained the model. Can you tell me how to test the model? If I run thetest.py
code using the commandpython test.py --dataroot ./datasets/MY_DATASET_DIR/ --name color_pix2pix --model pix2pix_colorization --phase test
, its throwing errorPlease tell me how to run the test.py file for colorization dataset mode.
Also, I assume the coloring dataset directory must be inside the
dataset
directory and must take the form./datasets/MY_DATASET_DIR_NAME/train
containing train images (just the colored images) and./datasets/MY_DATASET_DIR_NAME/test
containing test images (again just the colored images). I hope thats correct. If its correctm when I trained the model, the number of training images is shown as the sum of the images in thetrain
andtest
directory. Can you please tell me what is the issue?I trained the model for colorization using the earlier code (before pushing colorization dataset mode). For 256x256 images, the model was not overfitting while the model was overfitting for 512x512 images for the same model architecture. Can you explain the possible reasons please?