Unbabel / OpenKiwi

Open-Source Machine Translation Quality Estimation in PyTorch
https://unbabel.github.io/OpenKiwi/
GNU Affero General Public License v3.0
229 stars 48 forks source link

Plan to release other pre-trained models? #42

Closed sorami closed 5 years ago

sorami commented 5 years ago

Hi!

Do you have any plan to release pre-trained models other than En-De NMT?

captainvera commented 5 years ago

Hello @sorami,

Yes, in the future we intend to release the models of our participation on WMT19.

That is: En-De/En-Ru/En-Fr NMT

Did you have anything specific in mind?

sorami commented 5 years ago

Hi @captainvera , thank you very much for your reply!

Yes, in the future we intend to release the models of our participation on WMT19.

Got it!

I was looking for De-En language pair (WMT18 De-En SMT).

captainvera commented 5 years ago

Hi @sorami,

Let us discuss this, and I'll get back to you early next week 🙂

captainvera commented 5 years ago

Hello, I know that my colleague has already answered your email, but in an effort to be as open as possible to anyone with future interest in this:

Thank you for your interest in OpenKiwi.

We don't have concrete plans for releasing more pre-trained models. On the other hand, the same parallel corpus used for pre-training the en-de Predictor model could also be used for pre-training it in the reverse direction, so it should be somewhat easy to do.

However, its usefulness really depends on the domain of the data your going to use on top of it.

We do plan on future updates to OpenKiwi later this year so stay tuned!