Closed Artanic30 closed 5 months ago
Hi~ This is because we originally planned to train 3 text-conditioned image generation models from 1B to 7B. But limited to resources, we didn’t achieve it.
Thanks for your pointing out. We will fix it soon.
Thanks for the response.
Hi, I'm currently testing the official checkpoints. I found the model name and config in
autoregressive/models/gpt.py
In Readme.md, I can successfully load c2i_3B_384.pt as GPT_3B in
LlamaGen-3B | 3.1B | FSDP | 24x24 | 2.18 c2i_3B_384.pt
. However, the GPT_3B is marked as text-conditional in above code.