bcmi / DCI-VTON-Virtual-Try-On

[ACM Multimedia 2023] Taming the Power of Diffusion Models for High-Quality Virtual Try-On with Appearance Flow.
https://arxiv.org/abs/2308.06101
MIT License
387 stars 56 forks source link

Regarding Texture #29

Open Deepak2405 opened 10 months ago

Deepak2405 commented 10 months ago

How can I improve texture quality for custom images. I did retrain them, but texture quality does not improve much.

Also, the sleeve size is not according to the cloth image, but the sleeve gets adjusted to user input image. Could you suggest how I can make results better? 1_3 E347008_cloth_3 E342223_cloth_2

SalehAhmad1 commented 10 months ago

@Deepak2405 Can you share with me how you managed to execute the dci-vton script? It will be very helpful. I have tried following the readme file but it's a headache.

Deepak2405 commented 10 months ago

Actually, I just followed HR-VITON (ReadMe instructions) and applied it here.

hoangtdhdhcn commented 10 months ago

@Deepak2405 How many images did you use to train the DCI-VTON with your own data? and How many GPUs did you use? I met a problem "CUDA out of memory" although I used 4, then 5 GPUs. Did you meet this problem? Hope you can respond me. Thank you so much.

Deepak2405 commented 10 months ago

@hoangtdhdhcn we used 2 GPUs (each GPU has 40 GB memory). I did not face any issues as such.

hoangtdhdhcn commented 10 months ago

@Deepak2405 yes. Can I ask one more question? How many images did you train?

Deepak2405 commented 10 months ago

Around 2000 I guess. But we started training from VITON-HD weights.

hoangtdhdhcn commented 10 months ago

Around 2000 I guess. But we started training from VITON-HD weights.

OK, thank you so much

Limbor commented 10 months ago

@Deepak2405 In our subsequent experiments, we found that inpainting mask will have a great impact on the results, so we tried to use HR-VTON's mask selection strategy or a more aggressive strategy, which may help you improve performance. You can refer to https://github.com/bcmi/DCI-VTON-Virtual-Try-On/issues/21#issuecomment-1770821828

Deepak2405 commented 10 months ago

Thank you. I will check it.

wenhao728 commented 9 months ago

@Deepak2405 Here are a few insights from my experiences that might help enhance the current outcomes:

  1. Consider altering the masking strategy. The one used in this repository is quite reserved, with fewer masked regions. You might want to look into more assertive approaches, similar to those found in VITON-HD, HR-VITON, and TryOnDiffusion. However, be aware that TryOnDiffusion hasn't published its code, so you would need to develop an implementation on your own.

image

  1. Prompt the model by fine-tuning it with your dataset. The released weights of DCI-VTON were trained on a dataset (VTON-HD) comprising just 10k pairs for training, which means there's considerable potential to expand and adapt it to a wider variety of scenarios.
BJQ123456 commented 9 months ago

Hello, I have a question I'd like to ask. If training on my own dataset, how should I preprocess the data? For example, with DensePose images, should I retain the IUV format? However, I noticed that in the provided dataset, DensePose images seem to have only the I values. Thank you very much for your response

sachintha443 commented 2 months ago

@Deepak2405 how you managed to execute the dci-vton script? i try several attempts but not success full