podgorskiy / ALAE

[CVPR2020] Adversarial Latent Autoencoders
3.51k stars 560 forks source link

question about w #35

Open zhang-zx opened 4 years ago

zhang-zx commented 4 years ago

Thanks for your excellent work. As stated in your paper, your encoder generates styles code in w space. However, in your code, I think you use z space instead. Maybe there's some setting I overlooked. I hope you can help me understand it better.

huangzh13 commented 4 years ago

I have the same question.

csxmli2016 commented 4 years ago

same question.

5agado commented 4 years ago

I am also very interested regarding this aspect. Looking at the make_figures scripts, seems indeed you are just passing z to the model.generate function. At the beginning I thought that the low fidelity of projected images was because of w tiling, but it would be a total different story if your encoder works only to z.

florian-hoenicke commented 4 years ago

I have the same question. Since w contains more information, using w should improve the reconstructions.

youngAt19 commented 4 years ago

I think actually the author is using w space. The discriminator is implemented as a component in encoder. And the two returning terms of model.encode is w and d's output 图片

podgorskiy commented 3 years ago

@youngAt19 is right. Naming is misleading, that was a result of trying many configurations and things got their names after the writing started. I'll try to fix whatever I can without breaking compatibility with existing models