energy-based-model / Compositional-Visual-Generation-with-Composable-Diffusion-Models-PyTorch

[ECCV 2022] Compositional Generation using Diffusion Models
https://energy-based-model.github.io/Compositional-Visual-Generation-with-Composable-Diffusion-Models/
Other
456 stars 41 forks source link

Are the composed models separately trained? #15

Closed ziqihuangg closed 2 years ago

ziqihuangg commented 2 years ago

Hi, thanks for open sourcing this exciting work.

I wonder whether the composed diffusion models are separately trained? For example, for composing facial attributes, is there one model for synthesising faces with Smiling, and another separately trained model for synthesising faces with Glasses? Or we only need to train on FFHQ once to obtain P(x|c), where "c" can be Smiling or Glasses or Male?

Thank you very much!

nanlliu commented 2 years ago

In this paper, we used the same model trained on all possible attributes for a specific dataset for composing attributes. But it should also work if you train separate models on each attribute individually.

hope that helps!

ziqihuangg commented 2 years ago

Got it. Thank you very much!