ict-bigdatalab / CorpusBrain

CIKM 2022: CorpusBrain: Pre-train a Generative Retrieval Model for Knowledge-Intensive Language Tasks
Apache License 2.0
32 stars 3 forks source link

The Details of Fine-tuning Process #3

Open ma787639046 opened 1 year ago

ma787639046 commented 1 year ago

Hi, thanks for your awesome work!

Could you please share a more detailed hypermeter setting for your Fine-tuning Process, especially for the CorpusBrain(π‘šπ‘‘+𝐡𝐿𝐼𝑁𝐾) trail? Like max-tokens/GPU, the num of GPUs you used, update-freq, seed, and max steps you use. And how did you make the train and dev corpus for CorpusBrain(π‘šπ‘‘+𝐡𝐿𝐼𝑁𝐾)? Did you just mix and shuffle the BLINK train corpus + KILT train for training, and mix all KILT devs for the development set?

Thanks so much! @Chriskuei