Closed KyonP closed 1 year ago
Just like you. I also encountered this problem.
Just like you. I also encountered this problem.
I am not sure I made the right move, however, I managed to achieve "looks-okay" images.
In my case, I changed the size of the input images to as same as the LambdaLabs' pokemon examples (512*512).
and it worked "okay," but the reason why I keep using quotes is that the input images were only a small portion of them for fast development.
as soon as I inputted the full batches, it collapsed, not like the noise images but cloudy ones. (maybe it is because the dataset had to be upscaled to 512 since it has low resolution)
it seems like hyperparameter searching and tuning are essential.
I hope this small article helps.
LambdaLabsML/examples#33
Hi, can I ask that how do you find other repos showing finetuning text2image diffusion model on customized datasets? I want to fine tune text2image diffusion model on my own dataset using this latent-diffusion repo, but I do not know how I could create the dataset file and config, and also the script to train, do you know if there are some instructions or guidance about it? Thanks a lot!
@LilyDaytoy Hi,Do you have a solution now?
I am trying to train a custom dataset from the cartoon domain with text captions.
Tried out some other repo such as fine-tuning examples and optimizedSD, but I haven't been able to achieve results - they only outputted 2 types of results - partial characters from my dataset appeared all over an image or brown foggy noise images.
my config file is as follows:
while undergoing several hyperparameters tunning - lowering learning rate, reducing UNet dimensions etc., every time the validation outputs this kind of image:
I am not sure where did I make mistake. any suggestions would be greatly appreciated.