mlpc-ucsd / BLIVA

(AAAI 2024) BLIVA: A Simple Multimodal LLM for Better Handling of Text-rich Visual Questions
https://arxiv.org/abs/2308.09936
BSD 3-Clause "New" or "Revised" License
257 stars 26 forks source link

final loss of pretrain stage #11

Open weidong2018 opened 11 months ago

weidong2018 commented 11 months ago

Hi, nice work! 1)Could you please provide the final loss (the loss log during the training process) of bliva in the pre-training stage? 2)I find that in epoch>0, cosine_lr_schedule is performed according to epoch. Would it be better to follow the steps here to conduct cosine_lr_schedule?