Closed drimeF0 closed 10 months ago
Today I checked the main code of kohya_ss, the results are the same, so I'm closing the issue here
The problem is probably the datasets, also don't rely on training sample.
The problem is probably the datasets, also don't rely on training sample.
I've tried it 10 times already with different datasets and it always ends up with a terrible result. Today I’ll try to manually select images for training and see what happens
or use lower learning rate
The problem is probably the datasets, also don't rely on training sample. 100 learning steps, 48 pictures from safebooru with the tag "one_eye_closed", I already used my notebook, and apparently I made everything even worse than before. I also noticed this: Could this be a problem with such terrible image quality?
or use lower learning rate
by 0.00002, the deterioration of the quality of the generated images seems to be much slower. I'll see what happens after 1000 steps of learning.
or use lower learning rate
by 0.00002, the deterioration of the quality of the generated images seems to be much slower. I'll see what happens after 1000 steps of learning.
the image style has changed, but LoRA has not yet learned the concept of the one_eye_closed tag
The problem is probably the datasets, also don't rely on training sample. 100 learning steps, 48 pictures from safebooru with the tag "one_eye_closed", I already used my notebook, and apparently I made everything even worse than before. I also noticed this: Could this be a problem with such terrible image quality?
yes, 512x512 is unusable for sdxl
or use lower learning rate
I changed the regular lora to lora_fa from https://github.com/bmaltais/kohya_ss and it worked immediately after just 50 learning steps Also, I set lr to 0.00005
0.0002 or 2e-4 is higher than 1e-4, maybe 1e-5 or 5e-5
The problem is probably the datasets, also don't rely on training sample. 100 learning steps, 48 pictures from safebooru with the tag "one_eye_closed", I already used my notebook, and apparently I made everything even worse than before. I also noticed this: Could this be a problem with such terrible image quality?
yes, 512x512 is unusable for sdxl
how can this be fixed? resize images in the dataset, or add an argument to the command?
edit: I found a way, just add --max_resolution 1024,1024
argument to prepare_buckets_latents.py
I use the repository from qaneel
config:
The problem is that after each epoch, the quality of generation breaks down dramatically, and at about 3-4 epochs, a terrible mess begins.
Here is an example from the first training epoch:
And here is an example already at the 2nd epoch of training:
And finally, the 8th epoch:
What am I doing wrong, maybe I need to increase the number of epochs?