Closed MichalPitr closed 3 years ago
Hi, We did not try to convert Huggingface's checkpoints into fairseq checkpoints (used in KEPLER pre-training) since all our KEPLER pre-training are based on native fairseq framework. For your question, my suggestion is to read this code, which is used to convert fairseq RoBERTa checkpoints into Huggingface's and you may inversely convert it in similar way.
Best regards, Xiaozhi
Hi Xiozhi, Yea, I slightly feared that would be necessary. Thanks for the confirmation! If I get to do this, I'll share the code.
Kind regards, Michal
Hi, I'm curious, do you happen to know if there's a way to use a pre-trained model from Huggingface models and use that to initialize KEPLER training? I was hoping to initialize KEPLER with a RoBERTa pretrained on medical data and use KEPLER pretraining with medical knowledge graphs and medical MLM.
Many thanks, Michal