When I did the experiment on RCV1, while textcnn reach the similar result in your paper. The performance of the BERT model got a big gap, about 10% lower, comparing with another unpublished paper's result.
For the file lyrl2004_tokens_train.dat.gz, do I need some additional preprocessing since the texts in the file are unreadable.
When I did the experiment on RCV1, while textcnn reach the similar result in your paper. The performance of the BERT model got a big gap, about 10% lower, comparing with another unpublished paper's result.
For the file lyrl2004_tokens_train.dat.gz, do I need some additional preprocessing since the texts in the file are unreadable.
Thanks~