Closed Linaqruf closed 1 year ago
It would be helpful if you can write on there a little about the function of those arguments.
I get an error when I try to use 「bucket_reso_steps」and 「bucket_no_upscale」
train_network.py: error: unrecognized arguments: --bucket_no_upscale Traceback (most recent call last): File "/usr/local/bin/accelerate", line 8, in
sys.exit(main()) File "/usr/local/lib/python3.8/dist-packages/accelerate/commands/accelerate_cli.py", line 45, in main args.func(args) File "/usr/local/lib/python3.8/dist-packages/accelerate/commands/launch.py", line 1104, in launch_command simple_launcher(args) File "/usr/local/lib/python3.8/dist-packages/accelerate/commands/launch.py", line 567, in simple_launcher raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
you need to input dev
to branch field
It would be helpful if you can write on there a little about the function of those arguments.
you can read it here, https://github.com/kohya-ss/sd-scripts
Can you add unet scale and text encoder scale? According to https://github.com/cloneofsimo/lora/discussions/37 , in original repo both items can be changed separately, and not having them at 1.0 can sometimes be better.
Thanks for your hard work.
What Changes?
Login to Huggingface Hub
to Deployment
section, in the same cell with defining repo.Install Kohya Trainer
, Install Dependencies
, and Mount Drive
cellsDataset Cleaning
and Convert RGB to RGBA
cellsImage Upscaler
cell, because bucketing automatically upscale your dataset (converted to image latents) to min_bucket_reso
value.Colab Ram Patch
because now you can set --lowram
in the training script.Unzip dataset
cell to make it look simplerA100
Pretrained Model
section
Anything V3.3
, Chilloutmix
, and Counterfeit V2.5
as new pretrained model for SD V1.x based modelReplicant V1.0
, WD 1.5 Beta
and Illuminati Diffusion V1
as new pretrained model for SD V2.x 768v based modelStable Diffusion 1.5
pretrained model to pruned one.beam_search
enabled by defaultsafebooru
to booru listcustom_url
option, so you can copy and paste the url instead of specify which booru sites and tags to scrapeuser_agent
field, because you can't access some image board with default user_agentlimit_rate
field to limit your countwith_aria2c
to scrape your dataset, not a wrapper, just a simple trick to extract urls with gallery-dl
and download them with aria2c instead. Fast but seems igonoring --write-tags
..txt
format instead of .jpg.txt
additional_arguments
to make it more flexible if you want to try other argsAppend Custom Tag
cell
.txt/.caption
) if you didn't want to use BLIP or WD Tagger--keep_tokens
args to the cellTraining Model
section.
prettytable
for easier maintenance and bug fixingv2
, v2_parameterization
and all important folder and project_nameOptimizer Config
for notebook outside LoRA traininglearning_rate
and lr_scheduler
goes hereDAdaptation
if you're in Colab free tier because it need more VRAM--optimizer_args
for custom args, useful if you want to try adjusting weight decay, betas etccaption dropout
, you can drop your caption or tags by adjusting dropout rates.--bucket_reso_steps
and --bucket_no_upscale
--noise_offset
, read Diffusion With Offset Noise--lowram
to load the model in VRAM instead of CPUConvert Diffusers to Checkpoint
cell, now it's more readable.output_dir
located in google drive, it assert an error because of something like /content/drive/dreambooth_cmd.yaml
which is forbidden, now instead of saved to {output_dir}
, now training args history are saved to {training_dir}
News
experimental
section.
Please let me know if there is bugs, and error.
pretty confusing fr
Ikr, the trainer is more advanced than before, you can use old commit notebook or just train with default value if you find it too hard.
All new optimizer types, args, offset noise, caption dropout, new bucketing option, etc, too much new args to try, but not much time to try all of them.
I didn't change anything about default value nor backend script (except tag reading) compared to the previous version. You need to provide a screenshot, training logs, or copied notebook with the same hyperparams and dataset so I can figure out what is the error.
And yes, max_train_epochs
is not that accurate.
You can complain about backend script here instead : https://github.com/kohya-ss/sd-scripts
something not passed to the accelerate, 1680 was the default maxsteps
i think this is because the linebreak \
in the training config
can you check the train.sh to see all hyperparams passed?
i need to change this to yaml asap
I am working on updating Kohya's script to the latest version. The current version is from February 3rd, but the latest version in sd-scripts is from February 11th. This update will bring many new arguments and changes. I am also restructuring the notebooks to make the code more readable, maintainable, and user-friendly.
Here are some of the changes I have made:
5.I have chosen not to use these and let advanced users register new arguments in additional_arguments.
Here's a sneak peek: https://colab.research.google.com/github/Linaqruf/kohya-trainer/blob/8c52bc700f8cbb68a6ee58e4f73e27067d885e5d/kohya-LoRA-dreambooth.ipynb