Open pfliu-nlp opened 3 years ago
As first steps to working on multilingual evaluation, one should:
Read the tutorials on implementing new tasks, features, and formats.
Get system outputs for different multilingual systems. Here are some potential sources:
Run the ExplainaBoard SDK over these tasks and generate reports.
Compare the reports across languages. See if we can extract any interesting insights about the cross-lingual variations of trends.
More systematically, we might also try correlating being good or bad at particular fine-grained analysis categories with a few things:
As first steps to working on multilingual evaluation, one should:
Read the tutorials on implementing new tasks, features, and formats.
Get system outputs for different multilingual systems. Here are some potential sources:
Run the ExplainaBoard SDK over these tasks and generate reports.
Compare the reports across languages. See if we can extract any interesting insights about the cross-lingual variations of trends.
More systematically, we might also try correlating being good or bad at particular fine-grained analysis categories with a few things: