Closed markduon closed 3 years ago
Hi @duongquangvinh!
Another option to obtain a representation in the style space is to first obtain a (18, 512) "W+" code using any pretrained encoder (ReStyle. e4e, pSp etc.) or optimization, and then extracting the StyleSpace representation during a forward pass of the generator with the obtained "W+" code (which are the relevant activations as described in the StyleSpace paper).
Hope it helps, Best, Omer
Thank you so much!
I will read more about this before the implementation.
Cool! After further looking into the appendix I saw that:
In total, for a 1024x1024 generator, there are 6048 style channels that control feature maps, and 3040 additional channels that control the tRGB blocks.
So in total the size of the StyleSpace S is 6048 + 3040 = 9088
, which is slightly less then the 18 x 512 = 9216
parameters of the "W+" space.
This means that dividing some codes obtained from the current network architecture might work, although modifying the amount and the shape of the "map2style" layers output is also possible :).
The layer codes division should match the #channels described in the following table from the StyleSpace paper (Meaning there are 26 codes instead of 18, and that the first code is of size 512, while the last is of size 32 etc.):
Best, Omer
Hi sir,
I recently have read a paper called StyleSpace. 1) Can we apply the GANs inversion technique to invert the image into latent space of StyleSpace? It seems that this is S space (not W or W+) 2) Should I retrain the psp model with StyleSpace Generator because StyleSpace generator seems to be a little different compared to StyleGan2 generator?