Open Laidawang opened 1 year ago
Hi, when you fine-tune with multiple ground-truth images, be careful about the mapping during training, i.e. images should be first divided into different groups, X = {x_0, x_1, ..., x_n}, Y ={y_0, y_1, ... , y_m}, Z = {z_0, z_1, ..., z_k} ... each group represents different views for the same person. Then images from one group should only be mapped/correspond to images from the same group. This is not implemented in test.ipynb, but it is implemented in train.py, you can find some details there.
For the celebrity images, you can find it here, which contains some images that I manually collected (thus the data size is not very large).
Thank you very much for your help, as you said, I need to ensure that each one has a subfolder, such as A/01.jpg, A/02.jpg and B/01.jpg, B/02.jpg(A and B are two people), and then call train.py for fine-tuning?
You can either use train.py or test.ipynb, but both need to be revised.
If you want to use train.py for this experiment:
If you want to use test.ipynb:
Hello, Thanks a little for the interruption,what is the structure of your training data?
When I test multiple images, I first do a data augmentation with these two images at random, and then fine-tune my model on these data. But when I try to use different cfg combinations of these two pictures, I get pictures with almost the same kind of faces (combination of faces from both pictures) It seems that the cfg only changes their clothing and background. I noticed that you provided an identity_small model in your paper, it seems to change the face (more like A or more like B) according to different cfg. I notice that this models is fine-tuned on some celebraties images. Can you provide this data? Or simply tell me how to generate my fine-tuning dataset so that the model does not fuse faces.