bigscience-workshop / multilingual-modeling

BLOOM+1: Adapting BLOOM model to support a new unseen language
https://arxiv.org/abs/2212.09535
Apache License 2.0
69 stars 15 forks source link

Incrementally adding new languages to multilingual checkpoint #7

Closed vnikouliNLE closed 2 years ago

vnikouliNLE commented 2 years ago

Reproduce the same experiments as done by @yongzx on English checkpoint with multilingual checkpoint: https://huggingface.co/bigscience/tr5a-1B3-multilingual-mt5tok-checkpoints/tree/global_step30000 Related issues:

yongzx commented 2 years ago

Thanks @vnikouliNLE! For the adapter layer training, I think I should work on this instead of GPT-2, since now we have this model.

yongzx commented 2 years ago

Here's the link to the repo for incremental-addition-of-languages: https://github.com/bigscience-workshop/multilingual-modeling/tree/+lang