Open randompaga opened 1 year ago
Please specify the commit ID @randompaga .
git clone https://github.com/takuma104/diffusers.git
cd diffusers
git checkout 9a37409663a53f775fa380db332d37d7ea75c915
Please specify the commit ID @randompaga .
git clone https://github.com/takuma104/diffusers.git cd diffusers git checkout 9a37409663a53f775fa380db332d37d7ea75c915
Yea,it's worked.
But another erro occured.
Please help!
How do you install diffusers? You should run following command to install it, instead of pip install diffusers
. @randompaga
pip install .
Surely, diffusers recently has supported this officially, you can refer to their codebase if you need the newest version of diffusers, I will also update this tutorial soon. Anyway, if you just want to convert, this still works.
How do you install diffusers? You should run following command to install it, instead of
pip install diffusers
. @randompaga
pip install .
Surely, diffusers recently has supported this officially, you can refer to their codebase if you need the newest version of diffusers, I will also update this tutorial soon. Anyway, if you just want to convert, this still works.
Very thanks!I can generate image.But result seems get some wrong. What i input :
Prompt:1girl
What output:
I used the file https://huggingface.co/toyxyz/Control_any3/resolve/main/control_any3_openpose.pth which convert to diffusers.
root@VM-0-7-ubuntu:/home/ubuntu/diffusers# python ./scripts/convert_controlnet_to_diffusers.py --checkpoint_path /home/ubuntu/ControlNet/models/control_any3_openpose.pth --dump_path control_any3_openpose --device cpu global_step key not found in model Some weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPTextModel: ['vision_model.encoder.layers.16.mlp.fc1.weight',
root@VM-0-7-ubuntu:/home/ubuntu/diffusers# python test2.py /usr/local/lib/python3.8/dist-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead. warnings.warn( 16%|ββββββββββ
Thanks for your help.
@randompaga How do you infer? Can you attach your code?
python ./scripts/convert_controlnet_to_diffusers.py --checkpoint_path control_any3_openpose.pth --dump_path control_any3_openpose --device cpu
Did not find file convert_controlnet_to_diffusers.py