dorarad / gansformer

Generative Adversarial Transformers
MIT License
1.32k stars 149 forks source link

Do we still need bidirectional interaction attention? #24

Closed arieling closed 2 years ago

arieling commented 3 years ago

Hi Drew, Thank you a lot for sharing the code!

I notice that all checkpoints you released don't have _AttLayer_n2l._ So they are simplex + kmeans models? I also notice that number reported in github has already a bit better than paper. Does it mean we can safely ignore bidirectional attention for now? Does it further help and boost cityscpaes FID better than 5.23?

dorarad commented 2 years ago

Hi, thanks for reaching out and apologies for the large delay in the response! The checkpoints indeed are of the simplex model and reach a bit better scores than the paper since kept training them further! I expect bidirectional attention to give further improvement but I still actively explore the architecture structure and ways to optimize it further so currently released pretrained models for the simplex version. Note that you can also train the duplex model using the --g-img2ltnt flag, hope it helps!