Open howardyclo opened 6 years ago
This paper shows that it is possible to infer token-level label based on attention mechanism, despite that the model is only trained on sentence-level classification (binary classification). Evaluation on text uncertainty detection, grammatical error detection and sentiment classification along with several alternative methods shows that their attention-based method achieves best performance and are competitive to fully-supervised method. Also, the paper gives interesting attention visualization that interprets model's predictions.
L = L1 + r(L2 + L3), where L1 is L2-loss from sentence classification, L2 and L3 are regularization losses designed for the above two constrains respectively (see paper for details).
Instead of using attentions, three alternative methods can also infer token-level labels:
Metadata
Authors: Marek Rei and Anders Søgaard Organization: University of Cambridge & University of Copenhagen Conference: NAACL 2018 Paper: https://arxiv.org/pdf/1805.02214.pdf Code: https://github.com/marekrei/mltagger