Hi,
Thanks for sharing this amazing work. I have been using this hugging face-based repo for fine-tuning - https://github.com/huggingface/instruction-tuned-sd but I have some queries regarding the learning rate scheduler.
In the paper, it was mentioned - We use a learning rate of 10−4 (without any learning rate warm up).
Is it constant or some scheduler is used, I do see some config in train.yaml but can't understand
Hi, Thanks for sharing this amazing work. I have been using this hugging face-based repo for fine-tuning - https://github.com/huggingface/instruction-tuned-sd but I have some queries regarding the learning rate scheduler. In the paper, it was mentioned -
We use a learning rate of 10−4 (without any learning rate warm up).
Is it constant or some scheduler is used, I do see some config intrain.yaml
but can't understandIf it's being used, what would be the equivalent of this scheduler on diffusers.