Closed bjascob closed 3 years ago
Looks like you ran on the entire amr 1.0 corpus. I was only running the dev set. When I run the full corpus I get similar results to what's published.
Is there a way to save the model after training on the entire corpus and then "infer" the dev set with it later?
I'm trying to get the old ISI aligner to work as it still represents near SoTA performance and I have been using your code here, but I'm not able to replicate the results and I'm wondering if you have any idea why that might be.
When I run the scripts against amr-release-1.0-dev-consensus.txt I get, Precision: 78.76 Recall: 74.83 F1: 76.75. From the original paper, and your thesis, I was expecting an F1 of 86.5.
I've split the amr1 data using your scripts/mt_scripts/split_en_amr.py and I'm running against the ISI gold dev alignments. I'm using mgiza from https://github.com/moses-smt/mgiza/tree/master/mgizapp as the link in your README is not working. Given that the scores are reasonable, I think I must be doing something subtly wrong.
Questions: