Great repo! It's of great help to me. I just had 2 questions:
1) How do you do evaluation for pre-training?
2) Does the pre-training involve both mask infill and sentence permutation? If it does both can I just do mask infill? My main motive is to fine-tune a pre-trained BART with mask infill on an English corpus.
I don't assume evaluation during pretraining. Pretraining for a fixed number of iterations is what I follow usually. However if you want to do some validation then you can use the flag to save intermediate checkpoints and validate them separately via the decoding script in the repo.
By default it's only mask infilling. Sentence permutation hasn't led to major benifits in my experience.
Hello,
Great repo! It's of great help to me. I just had 2 questions:
1) How do you do evaluation for pre-training? 2) Does the pre-training involve both mask infill and sentence permutation? If it does both can I just do mask infill? My main motive is to fine-tune a pre-trained BART with mask infill on an English corpus.
Thank You so much!