fajri91 / discourse_probing

Discourse Probing of Pretrained Language Models. In Proceedings of NAACL 2021.
10 stars 1 forks source link
discourse naacl2021 nlp probing

Discourse Probing

About

In this paper, we introduce document-level discourse probing to evaluate the ability of pretrained LMs to capture document-level relations. We experiment with 7 pretrained LMs, 4 languages, and 7 discourse probing tasks, and find BART to be overall the best model at capturing discourse — but only in its encoder.

Paper

Fajri Koto, Jey Han Lau, and Timothy Baldwin. Discourse Probing of Pretrained Language Models. In Proceedings of the 20th Conference of the North American Chapter of the Association for Computational Linguistics (NAACL 2021), Mexico (virtual).

Discourse Probing Tasks

1. Next sentence prediction

2. Sentence ordering

3. Discourse connective prediction

4-5. RST nuclearity and relation prediction.

6. RST elementary discourse unit (EDU) segmentation.

7. Cloze story test.

Post Experiments

After running all the experiments, we provide some post-processing codes: