Closed QipengGuo closed 2 years ago
The command looks right to me.
The ontonotes model should not work well on the LitBank data (see our CRAC paper). Change the experiment=litbank
to experiment=ontonotes
.
If you want to experiment with LitBank, I would recommend using model/doc_encoder/transformer=longformer_joint
and downloading the joint model parameters from Google Drive.
The Joint model is trained on OntoNotes, LitBank, and PreCo and performs quite well on LitBank.
Note I have updated the codebase a bit and the new models are slightly better than the results reported earlier. I'm going to release them by this weekend.
Great thanks, it works now. I recommend giving more example commands in the README :).
Thanks for your great work, and I have a few questions about reproducing the results. I followed your steps on "Install Requirements", download your pre-trained models, and processed data from your google drive. The code ran smoothly, but pre-trained models work somehow weirdly, hope to get your help.
I tested the downloaded pre-trained model from your google drive on OntoNotes and Litbank, but the latter get strange results.
My run command is
The F1 score (58.7) is pretty low, but the more interesting thing is that the Oracle F-score is 0.825. If I understand correctly, this is the upper bound of the F1-score with the mention detection results. I wonder if there are some changes in huggingface models, but hard to track them.
BTW, I also try the command without override_encoder, but the result is less than 10 points.