Tengfei-Wang / HFGI

CVPR 2022 HFGI: High-Fidelity GAN Inversion for Image Attribute Editing
https://tengfei-wang.github.io/HFGI/
457 stars 50 forks source link

Could you provide a FFHQ 256x256 model? #12

Closed SleyAI closed 1 year ago

SleyAI commented 1 year ago

Hello, I'm struggling to train a model on the FFHQ 256x256 dataset. I trained an Encoder4Editing model on the entire FFHQ dataset (66k images for training, 4k for validation) and the results look comparable to the ones in the Encoder4Editing paper. Then I trained a HFGI model based on that e4e checkpoint with good results aswell. But when I try to project an image the inversion looks noticeably different than the input image. This problem doesn't appear when I use your pretrained FFHQ 1024x1024 model. I'm assuming that it should be possible to train a 256x256 model with comparable quality.

Could you share a FFHQ 256x256 checkpoint so that I can validate my results? Thank you!

Tengfei-Wang commented 1 year ago

Hi, You can find the checkpoint on faces in our repo. What do your failure results look like? Could you share them?

获取 Outlook for iOShttps://aka.ms/o0ukef


发件人: Henning @.> 发送时间: Tuesday, March 7, 2023 11:50:13 PM 收件人: Tengfei-Wang/HFGI @.> 抄送: Subscribed @.***> 主题: [Tengfei-Wang/HFGI] Could you provide a FFHQ 256x256 model? (Issue #12)

Hello, I'm struggling to train a model on the FFHQ 256x256 dataset. I trained an Encoder4Editing model on the entire FFHQ dataset (66k images for training, 4k for validation) and the results look comparable to the ones in the Encoder4Editing paper. Then I trained a HFGI model based on that e4e checkpoint with good results aswell. But when I try to project an image the inversion looks noticeably different than the input image.

Could you share a FFHQ 256x256 checkpoint so that I can validate my results? Thank you!

― Reply to this email directly, view it on GitHubhttps://apc01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2FTengfei-Wang%2FHFGI%2Fissues%2F12&data=05%7C01%7Ctfwang%40connect.ust.hk%7Ca1de75d977a8480bf7e208db1f23aacd%7C6c1d415239d044ca88d9b8d6ddca0708%7C1%7C0%7C638138010267121108%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=5jut40j6x4zz4fMgF2yM7xxgYfLpkgCtGxoNn067Cio%3D&reserved=0, or unsubscribehttps://apc01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFNHPI2YUX2XIAH3O3QKOODW25KLLANCNFSM6AAAAAAVSUXWJY&data=05%7C01%7Ctfwang%40connect.ust.hk%7Ca1de75d977a8480bf7e208db1f23aacd%7C6c1d415239d044ca88d9b8d6ddca0708%7C1%7C0%7C638138010267277322%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=SH4tqi5iZj1mheBTgDU5Qcx0PcP6nb8pIHYY2obmBQY%3D&reserved=0. You are receiving this because you are subscribed to this thread.Message ID: @.***>

SleyAI commented 1 year ago

Hi, thanks. I already found your checkpoint but I believe it's trained to generate 1024x1024 images. Currently I'm trying to train a model which generates 256x256 images. Image downscaling from 1024x1024 -> 256x256 is not an option since I want to compare my method (based on HFGI) to another paper which also uses 256x256 images. The 256x256 inversions from the model I trained myself have a noticeable distortion compared to your 1024x1024 model. I attached samples below. First images is input, second is with your checkpoint (1024x1024 downscaled to 256x256) and third is the model trained by me (256x256).

Generator is the one from Rosinality you provided and e4e model was trained by myself with good sim values (~0.9). With both I trained HFGI by following your description until LPIPS converged against ~0.04.

223659163-5927d32a-94da-4292-a4fb-0d830180c1da 223659206-1b638f25-bc86-40e2-a680-ed93b51c51d9 223659221-69ddb0a1-3179-46e4-b234-e96a980ceefb

zpshs commented 1 year ago

Hello, may I ask where the FFHQ 256x256 checkpoint is? I can't find it. thank you.

Tengfei-Wang commented 1 year ago

As the pretrained styleGAN we used is of 1024x1024, the resolution of output image is 1024 by default, but the input image can be 256x256.

获取 Outlook for iOShttps://aka.ms/o0ukef


发件人: zpshs @.> 发送时间: Tuesday, April 18, 2023 10:49:06 AM 收件人: Tengfei-Wang/HFGI @.> 抄送: Tengfei Wang @.>; Comment @.> 主题: Re: [Tengfei-Wang/HFGI] Could you provide a FFHQ 256x256 model? (Issue #12)

Hello, may I ask where the FFHQ 256x256 checkpoint is? I can't find it. thank you.

― Reply to this email directly, view it on GitHubhttps://github.com/Tengfei-Wang/HFGI/issues/12#issuecomment-1512357955, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AFNHPI5N2Q7GU7ZT7DBMTPLXBX6KFANCNFSM6AAAAAAVSUXWJY. You are receiving this because you commented.Message ID: @.***>