openai / glide-text2im

GLIDE: a diffusion-based text-conditional image synthesis model
MIT License
3.53k stars 500 forks source link

How to get the results closer to what is shown in the paper? #5

Closed kaloyan-chernev closed 2 years ago

kaloyan-chernev commented 2 years ago

Really inspirational work guys!

But the results from the published code and models are not even remotely comparable to the shown results in the paper. Is there anything we can do to get closer to the original work?

Image from the paper for: "a surrealist dream-like oil painting by salvador dalı́ of a cat playing checkers"

image

Image from the code for the same text prompt "a surrealist dream-like oil painting by salvador..."

image

It's almost like that meme: " Your vs. The guy she told you not to worry about" :rofl:

Anyway, if you can give us some advice on this matter it would be greatly appreciated! :+1:

woctezuma commented 2 years ago

See https://github.com/openai/glide-text2im/issues/2#issuecomment-999037054.

unixpickle commented 2 years ago

We have not released the full GLIDE model--only GLIDE (filtered) which is 10x smaller than the original model and trained on a much more restricted dataset. We hope this model is still useful for future research, but it won't be able to reproduce the best images in the paper because of these limitations.