menyifang / DCT-Net

Official implementation of "DCT-Net: Domain-Calibrated Translation for Portrait Stylization", SIGGRAPH 2022 (TOG); Multi-style cartoonization
Apache License 2.0
752 stars 75 forks source link

Questions about the separation of the head model and the background model #7

Open neuralchen opened 2 years ago

neuralchen commented 2 years ago

Thank you for your impressive work. Also thank you for sharing the model and code. We believe this work is one of the most important in the field of style transfer this year.

I studied the code you posted in detail. The separation of the head transition model and the background transition model is very confusing to me. In the paper, the details we get are that we only need a Texture translation network to uniformly transfer the style of the whole image. It is this feature that impresses me!

Looking forward to your reply.

Best,

menyifang commented 2 years ago

Good question. Here we provide a more practical infer version with head & background processed separately for time efficiency (especially for images with small faces & image with multi-faces in different size). The model also supports directly full image translation with face in suitable scale. With resized images, directly use "cartoon_animebg.pb" for inference. Actually, cartoon[style]bg.pb and cartoon[style]_h.pb are the same model with same params. For anime style, we give an additonal optim for head model, but bg model also works well.