microsoft / ProphetNet

A research project for natural language generation, containing the official implementations by MSRA NLC team.
MIT License
654 stars 105 forks source link

How to pretrain about other language #6

Open wanaDoctor opened 4 years ago

wanaDoctor commented 4 years ago

Hello, thanks for sharing this awesome paper. I want to pretrain this model in Korean. But this project is not provide pretraining code. if you don't mind me asking, would you provide pretraining code?

thanks :)

qiweizhen commented 4 years ago

Multi-language pretrained model will be released in near future.

Pretraining code is not schedulred to release. But you can simply use Fairseq build-in pretraining tasks or format your data in input-output format to pretrain a Kerean model.

nickcom007 commented 4 years ago

Hello, thanks for sharing this awesome paper. I want to pretrain this model in Korean. But this project is not provide pretraining code. if you don't mind me asking, would you provide pretraining code?

thanks :)

Hi, have you successfully pre-train in another language?

yuyan2do commented 4 years ago

@nickcom007 See cross-lingual version at this page

rkoyanagi commented 4 years ago

Thank you for sharing great code. I also want to pretrain ProphetNet in other language and trying, but I cound not success even in English yet. (no cross-lingual, because I want to use only our vocabulary of our language and domainf without noise.)

But you can simply use Fairseq build-in pretraining tasks

Does it means fairseq-train --task masked_lm or another one? Could you show me a sample command to reproduct?

Or

format your data in input-output format to pretrain a Kerean model.

Could you show me an example of input-output format of pretrain (in English)?

rkoyanagi commented 3 years ago

Using the mass_s2s task from MASS , it worked.

ghost commented 3 years ago

@rkoyanagi Can you show me how to train new model for other languages? I am using it for Vietnamese Summarization. Thanks.