Closed BJQ123456 closed 2 months ago
Hello, thanks for your feedback!
Yes, the first arXiv paper was a bit rushed and lacks some refinement. Since this project is still under development, it might not fully align with the paper at this stage.
We will continue to refine the paper until the final version is ready.
We would greatly appreciate any help you can offer in correcting it!
Thanks for your reply. I have a few questions:
1.The text mentions "plug-and-play," but the base model was trained. Isn't this contradictory?
2.Is there no comparison with methods like ControlNet?
3.Formulas (8) and (9) seem incorrect.
4.Figure 5 does not explain what the three results represent.
5.Figure 8 is unclear.
6.Figure 9 does not explain the conditions of the experiment on the left side.
Hello, thanks for your question, and I think that they are all good questions! I will share more details.
Since the paper has limited space, I would like to share additional experiences. Can I change the title to 'Training Experiments and Insights'? @BJQ123456
SVD-related
We spent a lot of time to find these tips, now share all with all of you. May there will help you!
2.We have compared the efficiency and training convergence. More detailed results will be added later. 3.Have been corrected. 4.5.6.: Thanks, we will refine these details
Thank you for your response; it was very helpful.
There's one point I still don't understand. If it's plug-and-play, does it mean that after training, I can directly use it on odels? But during training, a part of the base model was trained, so if I insert it into a new model, that part hasn't been trained and could lead to a performance drop, right? thanks for your reply
Yes, as demonstrated in our experiments with SD1.5 and SDXL, we trained on a single backbone and then conducted experiments across various backbones. The results show that our method effectively performs control on different backbones.
We also considered this approach and initially attempted to store the weight increments, same as LoRA only without low-rank compression. However, we eventually found that this step was unnecessary.
I get it, thanks
hi could the authors also share light on the combining this with IP-ADAPTERS? will it cause any issues making it work with IP-adapters? does it work well with pretrained IP-Adapters?
感觉论文里面错误好多啊