imlixinyang / HiSD

Code for "Image-to-image Translation via Hierarchical Style Disentanglement" (CVPR 2021 Oral).
Other
391 stars 49 forks source link

About celeba-hq.yaml #13

Closed Hpjhpjhs closed 3 years ago

Hpjhpjhs commented 3 years ago

Thanks for your great works. I have some questions about the training configuration. There are differences between celeba-hq.yaml and celeba-hq_256.yaml in configs folder, such as the normalization in different channels. What's the reason here? If I try to train the model based on the size of 512 or 1024, how should I set? It is appreciated if I can receive your reple

imlixinyang commented 3 years ago

Actually, the normalization methods are the same among different configs, while for different modules, they are different (AdaIN for translator, None for discriminator and extractor, IN for encoder and decoder). Non-normalization is discriminator is from most recent GANs like stylegan or starganv2. AdaIN is to make the output condition on the style, you can change this to demodulation following stylegan2. Using IN in encoder and decoder is the most uncertain one, you can try different normalization methods here (but I found that non-normalization will make the training unstable).

The difference between these two configs is the different channel numbers in the discriminator and extractor, which I have to decrease since the limited GPU memory (2x 1080Ti).

For high-resolution training, please kindly refer to this issue (https://github.com/imlixinyang/HiSD/issues/12), where I clarify why I haven't released a 512 or higher resolution checkpoint and some possible solutions.

Hpjhpjhs commented 3 years ago

@imlixinyang Thanks for your detailed explanation. I also tried some other methods like pixel2style2pixel to do image transfer, which can manipulate 1024 resolution of images, but most of them are strongly dependent on the encoded latent code。 So the inspiration of this work is amazing. Thanks for your great work again. By the way, I have a relative research background and strong interest in image translation research, So I want to ask whether the research of image-to-image translation is led by Pr. Ji or any other people in MAC lab?

imlixinyang commented 3 years ago

Thanks for your nice words. All co-authors did a great help for this paper. If you have any further interest for my lab or advisor, would you mind contacting me by my email or my zhihu account 李新阳.