SimiaoZuo / MoEBERT

This PyTorch package implements MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation (NAACL 2022).
Apache License 2.0
97 stars 13 forks source link

What is the bash script of finetune without MoE #5

Open CaffreyR opened 2 years ago

CaffreyR commented 2 years ago

Hi @SimiaoZuo , as you mentioned that we need to finetune first. But how to get the finetune model and translate into bert_base_mnli_example.sh! Many thanks!