neulab / awesome-align

A neural word aligner based on multilingual BERT
https://arxiv.org/abs/2101.08231
BSD 3-Clause "New" or "Revised" License
321 stars 46 forks source link

Statistics between models #46

Closed b3ade closed 2 years ago

b3ade commented 2 years ago

Do you have any statistics or model performance between mBERT, model_without_co and model_with_co? I see that you have alignment error rates on your models between aligners. So I'm wondering if you have any information on how your model is different from mBERT.

zdou0830 commented 2 years ago

Hi, we essentially propose a way to extract alignments from mBERT and propose objectives to fine-tune mBERT on parallel data for word alignment. For the table in README, Ours (w/o fine-tuning, softmax) is the mBERT model performance using our alignment extraction method.