Open AndrewBens opened 1 week ago
hi, do you load with the same tokenizer and config? examples/GFSLT-VLP/configs/config_gloss_free_csl.yaml examples/scripts.sh
I didn't meet with this
@JinhuiYE Thank you a lot. I am using the mbart_char that I trimmed myself. There may be some bugs. I plan to try the pretrained model you shared in other issues.🙂
@JinhuiYE May I ask you, how long will it take to pre-train and fine-tune CSL-Daily on eight A800s?🤝
Hi, we are happy to share. In our case, we use 8*A800 GPUs for the two training stages. The pretraining takes about 12 hours for 80 epochs. The finetuning stage takes approximately one day.
@JinhuiYE Get it!Thank you!
Hello🙂! Thanks for your excellent works!😊
Have you ever encountered this problem? I am trying to reproduce GFSLT. But when I try to run train_slt.py after completing 80 epochs of pre-training(VLP_2), I found that a problem occurs when loading the visual encoder on both the Phoenix and CSL-Daily datasets:
Load parameters for Visual Encoder...
Missing keys: mbart.final_logits_bias mbart.model.shared.weight mbart.lm_head.weight sign_emb.src_emb.weight sign_emb.src_emb.bias sign_emb.bn_ac.0.weight sign_emb.bn_ac.0.bias sign_emb.bn_ac.0.running_mean sign_emb.bn_ac.0.running_var
Do you have any suggestions? Thank you!