Closed garbe-github-support closed 6 months ago
These results look abnormal. This has not happened in all my test examples including wild images. test_images/unaligned_img/test.jpg
is a random wild image I downloaded from www.pexels.com and it works fine. Here are some issues you need to be aware of:
python scripts/align_face.py
.Perhaps there may be some other cases that I haven't mentioned. Can you send me your test image? I will help you to reproduce it after January 2, 2024 (right now I'm on a new year vacation). To protect the privacy of your images, you can send them to my email bestwty@mail.ustc.edu.cn.
Also, as far as speed is concerned, the main time overhead of our approach is in the inversion phase. You can choose to speed it up by reducing W_steps
inside utils/options.py,
at the cost of some degree of reconstruction loss, of course.
Compared to hair editing based on stable diffusion inpainting, our approach is a multi-modal hair editing system with additional support for hair transfer, sketch, mask, RGB and other interaction modes. Also, for text-based editing, I consider our approach to be comparable to stable diffusion in terms of effectiveness. If the StyleGAN is trained on a dataset with more diverse hairstyles, I think our method will perform even better.
The process is to run 200 rounds, how long does it take? 3090 can run?
Maybe 1 minute. You can reduce the number to 100 for most cases. As shown below, 60 rounds are enough for bowl cut hairstyle. Besides, 3090 is suitable for our project.
I referred to readme for many configurations and dealt with various bugs, The image has been changed to 1024 * 1024。
Does the provided image need to be centered on the front face in order to be reconstructed?
Using your provide image(Generate NPZ simultaneously instead of using preset ones) can reconstruct the bald image of the original image very well, like this![image](https://github.com/wty-ustc/HairCLIPv2/assets/131808107/bede3f6d-675e-4e9a-83b4-aa874d3c48b0)
but if I use a new image, the newly created image is completely incorrect and very ugly。Like this
![image](https://github.com/wty-ustc/HairCLIPv2/assets/131808107/a6e99a75-eb30-412a-aaa9-bc68ee055edd)
If My configuration is correct, I just want to say, the overly boastful paper, it wastes my time。The effect and speed of using stable diffusion are much better than yours