Open yasserben opened 1 year ago
Does anyone solve this issue? I think generating one image in each time (2 times iteration on unet) would be the solution. It would be really appreciated if someone can share the modifications of the code.
You can use the ddim latents that are stored while doing the null text optimization and initialize the first image with those latents while editing.
Hi ! :smiley:
I have a problem regarding the prompt-to-prompt notebook. The image of the squirrel changes a little bit between the Cross-Attention Visualization :
and Replacement edit cells :
sections:
The one on the left was generated from the cell of the Cross-Attention Visualization, and the right one from **the Replacement edit** cell. If you look closely at the two black circles on the left, you’ll see a difference between the two squirrels, and this is not supposed to happen I guess. I think you can reproduce the same errors, if you run the following code :
The single squirrel in
image_1
would be different from the two squirrels generated inimage_2
After going a little bit through the code, I suspect it comes from the size of the prompt. Because it contains two sentences, so we have
batch_size = 2
in the Replacement cell. I think that’s why it doesn’t generate the exact same picture as if we used only one sentence, so maybe the total size the batch influences the generation of an image from a text even if the prompt remains the same.The same problem arises when we work with the notebook of Null-text inversion :
Thank you for your help !! :blush: