Open malteos opened 1 year ago
List of cross-lingual tasks and evaluation benchmarks:
XTREME: Cross-lingual Transfer Evaluation of Multilingual Encoders (XTREME) benchmark is a benchmark for the evaluation of the cross-lingual generalization ability of pre-trained multilingual models. URL: https://huggingface.co/datasets/xtreme
XQUAD: Benchmark dataset for evaluating cross-lingual question-answering performance. URL: https://huggingface.co/datasets/xquad
TyDiQA: Question-answering dataset covering 11 typologically diverse languages with 204K question-answer pairs. URL: https://huggingface.co/datasets/tydiqa
XNLI: Subset of a few thousand examples from MNLI which has been translated into 14 different languages (some low-ish resource). URL: https://huggingface.co/datasets/xnli
Wiki-Lingua Cross-Lingual Summarization: Multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. URL: https://huggingface.co/datasets/GEM/wiki_lingua
MLQA: (MultiLingual Question Answering) is a benchmark dataset for evaluating cross-lingual question-answering performance. URL: https://huggingface.co/datasets/mlqa
PAWS-X: Cross-lingual Adversarial Dataset for Paraphrase Identification. This dataset contains 23,659 human-translated PAWS evaluation pairs and 296,406 machine-translated training pairs in six typologically distinct languages. URL: https://huggingface.co/datasets/paws-x
OPUS: English-centric, meaning that all training pairs include English on either the source or target side. The corpus covers 100 languages (including English). Selected the languages based on the volume of parallel data available in OPUS. URL: https://huggingface.co/datasets/opus100
List of Monolingual Benchmarks:
German
French
Slavic and Other Languages
Thanks for the list. Can you also look for monolingual benchmarks in the respective languages?
TODO Languages:
Top languages with at least three tasks per language:
Other languages with at least one task per language:
Multilingual
Done: German, English