Closed ikuinen closed 1 year ago
Thanks for your interest in our work! Indeed, the #hack line should be removed which allows the gradient to flow back to the transformer as we aim to update the global sampler. I've updated the code. https://github.com/FrozenBurning/Text2Light/blob/af8ec40412777c13d7f1739da0d9ca1de00bcc1f/taming/models/global_sampler.py#L167
Thanks for your feedback! 🍻
Thanks for your interest in our work! Indeed, the #hack line should be removed which allows the gradient to flow back to the transformer as we aim to update the global sampler. I've updated the code.
Thanks for your feedback! 🍻
Thanks for your reply! But the transformer is auto-regressive model which generates index step by step. The generated index will stop the gradient to the transformer. Does the loss item allow the gradient flow back to transformer?
Looking forward to your reply.
I got your point. And I've dug into our original implementation and updated the code.
Thanks for your feedback! 🍻
Closed due to inactivity. Feel free to reopen it 🙌
It seems that the contrastive loss have no gradient on the network. The "gen_img_emb" is generated by fixed CLIP while the " psed_emb" is pre-computed in dataloader:
looking forward to your reply.