Is it possible to set the lora [sdxl_train_network.py] text encoder learning rate separately?
similar to Flux,
In Clip_L there are 4.45 times fewer parameters than Clip_G.
As a result, the Text-Encoder adjustment may not match the steps very well.
Clip_L has a smaller capacity. Lower flexibility, higher TE-LR required.
Clip_G, high capacity, more flexible. But there is a higher risk of overfitting than L due to different parameters.
which in SDXL_train.py Instead, there are separate settings.
sdxl_train.py
parser.add_argument(
"--learning_rate_te1",
type=float,
default=None,
help="learning rate for text encoder 1 (ViT-L) / text encoder 1 (ViT-L)の学習率",
)
parser.add_argument(
"--learning_rate_te2",
type=float,
default=None,
help="learning rate for text encoder 2 (BiG-G) / text encoder 2 (BiG-G)の学習率",
Is it possible to set the lora [sdxl_train_network.py] text encoder learning rate separately? similar to Flux, In Clip_L there are 4.45 times fewer parameters than Clip_G. As a result, the Text-Encoder adjustment may not match the steps very well.
which in SDXL_train.py Instead, there are separate settings.
sdxl_train.py