Closed katrina433 closed 2 years ago
Hello!
I have a question about the mlm loss computation in model_vqa.py. It seems like alpha is never used in forward, so the answer_output is independent of logits_m. Is this the intended behavior?
alpha
forward
answer_output
logits_m
Hi, thanks for pointing this out. The code has been updated now to use alpha when self.distill=True.
Hello!
I have a question about the mlm loss computation in model_vqa.py. It seems like
alpha
is never used inforward
, so theanswer_output
is independent oflogits_m
. Is this the intended behavior?