frank-xwang / InstanceDiffusion

[CVPR 2024] Code release for "InstanceDiffusion: Instance-level Control for Image Generation"
https://people.eecs.berkeley.edu/~xdwang/projects/InstDiff/
Apache License 2.0
484 stars 25 forks source link

Other ways to training the model #17

Open 66ling66 opened 6 months ago

66ling66 commented 6 months ago

屏幕截图 2024-03-18 171430 Could you provide a new instruct? or some example of the training data

66ling66 commented 6 months ago

屏幕截图 2024-03-18 173953

frank-xwang commented 6 months ago

Can you provide the full command you used for running the experiments and the error message? Thanks.

66ling66 commented 6 months ago

Thanks you,my gpu memory is not encough,and I just change my training strategy.

66ling66 commented 6 months ago

https://github.com/frank-xwang/InstanceDiffusion/blob/dadf0e3b09c2de82bf35b24e3424a14197a29906/trainer.py#L167 if I don't want to add this when training the unet model,what should I do?Is there a very simple way to do this?

frank-xwang commented 5 months ago

Sorry for the late reply. I am a little bit confused, you mean you want to train the model without any grounding inputs? Our model training needs to have instance/part-level location and captioning inputs, otherwise, it should be equivalent to directly fine-tuning the Stable Diffusion model.

66ling66 commented 5 months ago

Yes I want to train the model without any grounding inputs,I just train a lora model ,when sampling image,add the grounding inputs,it seems work a litle bit.

frank-xwang commented 5 months ago

Oh, I see. Maybe you can provide zero tensors as a placeholder for bbox, masks and instance caption embeddings. The easies way might be calling 'self.grounding_tokenizer_input.get_null_input()'. You should manually set 'self.set=True', and provide the 'self.device', 'self.dtype', 'self.max_box', etc.

frank-xwang commented 5 months ago

You can find more details on this function at 'grounding_input/text_grounding_tokinzer_input.py'

Chaoriticle commented 1 month ago

Yes I want to train the model without any grounding inputs,I just train a lora model ,when sampling image,add the grounding inputs,it seems work a litle bit.

Hi, I am a beginner in deep learning and I am very interested in the process you used to train the lora model, can you please share the whole process in detail? It will be very helpful for me, thank you very much!

66ling66 commented 1 month ago

Yes I want to train the model without any grounding inputs,I just train a lora model ,when sampling image,add the grounding inputs,it seems work a litle bit.

Hi, I am a beginner in deep learning and I am very interested in the process you used to train the lora model, can you please share the whole process in detail? It will be very helpful for me, thank you very
First you should do that https://github.com/frank-xwang/InstanceDiffusion/issues/17#issuecomment-2073965484 make sure the grounding inputs is null then you can train the lora model.Maybe you can search what is lora and how to train it.Then all the things solved.