goodbai-nlp / AMRBART

Code for our paper "Graph Pre-training for AMR Parsing and Generation" in ACL2022
MIT License
92 stars 28 forks source link

Difference in hyper-parameters #4

Closed PhMeier closed 2 years ago

PhMeier commented 2 years ago

Hello,

thank you very much for you work and providing the code!

While comparing the fine-tuning scripts to your reported hyper-parameters in your paper I have seen some differences:

I guess the parameters from the scripts are more recent?

goodbai-nlp commented 2 years ago

Hi, thanks for pointing this out. For learning rate, that's a typo and we will update the paper accordingly. We provide a bigger value for sequence length and early stopping patience in scripts because they are helpful for learning longer (> 512) AMR sequences (although only ~0.03% in corpus) and selecting better models. We believe these changes are beneficial for other researchers to achieve comparable or better results than our paper, especially under a different experimental environment.

PhMeier commented 2 years ago

Hi muyeby, thank you very much for explaining!