Open KaiyueSun98 opened 1 year ago
The samples_scaled images use a guidance scale of 5.0 and you are using 10.0. If the poor quality you are seeing is oversaturation and similar effects, this is likely the reason. Try to drop the guidance scale a bit (stable diffusion typically uses 7.5).
@KaiyueSun98 Which weight file did you provide to --ckpt_path
to load stable diffusion?
Hi @wonjunior, I downloaded the weight file "sd-v1-4.ckpt" from hugging face https://huggingface.co/CompVis/stable-diffusion-v-1-4-original
@KaiyueSun98 Hello, I've downloaded the weight as what you mentioned and used the configs/stable-diffusion/v1-finetune.yaml
for training. There is an error code like Some weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPTextModel:
and followed by a lot of layer name. Did you have the same situation?
@KaiyueSun98 Hello, I've downloaded the weight as what you mentioned and used the
configs/stable-diffusion/v1-finetune.yaml
for training. There is an error code likeSome weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPTextModel:
and followed by a lot of layer name. Did you have the same situation?
Hi, did you resolve this issue? Maybe I need some help. Thanks.
@KaiyueSun98 Hello, I've downloaded the weight as what you mentioned and used the
configs/stable-diffusion/v1-finetune.yaml
for training. There is an error code likeSome weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPTextModel:
and followed by a lot of layer name. Did you have the same situation?Hi, did you resolve this issue? Maybe I need some help. Thanks.
I found this issue will not affect the pipeline, so I just ignored it lol.
Thanks : )
Hi,
I use configs/stable-diffusion/v1-finetune.yaml for inversion and configs/stable-diffusion/v1-inference.yaml for inference (by replacing the original
config
in script/txt2img.py withconfigs/stable-diffusion/v1-inference.yaml
). other parameters remain unchanged.when I use stable diffusion 1.4 to generate images (with the same SD model doing the inversion),
the output images looks worse than the
samples_scaled
images in log. Are the waysamples_scaled
images generated consistent with the way output images generated? or Are there any other parameters for inference I need to change to fit the SD model?