wtangdev / UniRel

released code for our EMNLP22 paper: UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction
Apache License 2.0
82 stars 17 forks source link

Experiment evaluation #2

Closed TornadoXuRocket closed 1 year ago

TornadoXuRocket commented 1 year ago

Hello, thank you very much for the code release. I have used your code to train and test on nyt_star. Why is the performance much worse than that provided in your paper? d07472ac4f3319ce169220461f8adeb

wtangdev commented 1 year ago

Hi Jing, I can not directly give any suggestion according to the information you pasted. And I also noticed that you are not using the complete codes we provided. So please update more information about your reimplementation (like learning rate, optimizer type, etc.). I am happy to help you.

TornadoXuRocket commented 1 year ago

Thank you very much for your response. I just used your complete codes and run the script run_nyt.sh which is provided by you. My only change is to close the wandb service. Could you give me some suggestions, please? Thanks again.

wtangdev commented 1 year ago

Sorry, can you give me more details? I cannot figure out what problem caused this from the information you provided. But from the results, especially the loss, it looks like the model have not be well convergent.

wtangdev commented 1 year ago

Sorry again. I didn't notice the learning rate is provided in your picture, also forgot the runtime summary is the output from transformers trainer (which I have thought is generated by yourself).

The learning rate looks not correct in your picture, please double check it.

TornadoXuRocket commented 1 year ago

Thank you very much. I will check them carefully. Best regards.

TornadoXuRocket commented 1 year ago

I run the script run_webnlg.sh and the performance is also much worse than the results reported in the paper. I carefully have checked the learning rate and the training losses. At the beginning of training, the training information is shown as follows: image When the epoch is about 20, the training information is shown as follows: image As for the end of the training, the learning rate has decreased to a very small value due to the weight decay and the training information is shown as follows: image May I ask if this training process is normal? Thank you again.

Go0day commented 1 year ago

This is strange, I tried bash run_nyt.sh and got a fine result at 18 epochs, maybe there is something wrong with your experimental setup. 302771676469112_ pic

wtangdev commented 1 year ago

I run the script run_webnlg.sh and the performance is also much worse than the results reported in the paper. I carefully have checked the learning rate and the training losses. At the beginning of training, the training information is shown as follows: image When the epoch is about 20, the training information is shown as follows: image As for the end of the training, the learning rate has decreased to a very small value due to the weight decay and the training information is shown as follows: image May I ask if this training process is normal? Thank you again.

Hi, The training procedure looks good. Do you use the dataset directly downloaded from TPLinker?

TornadoXuRocket commented 1 year ago

This is strange, I tried bash run_nyt.sh and got a fine result at 18 epochs, maybe there is something wrong with your experimental setup. 302771676469112_ pic

Thank you. I will check my experimental setup again.

TornadoXuRocket commented 1 year ago

I run the script run_webnlg.sh and the performance is also much worse than the results reported in the paper. I carefully have checked the learning rate and the training losses. At the beginning of training, the training information is shown as follows: image When the epoch is about 20, the training information is shown as follows: image As for the end of the training, the learning rate has decreased to a very small value due to the weight decay and the training information is shown as follows: image May I ask if this training process is normal? Thank you again.

Hi, The training procedure looks good. Do you use the dataset directly downloaded from TPLinker?

Yes, I followed TPLinker to download and preprocess the dataset which is provided by CasRel. Maybe there is something wrong with my data preprocessing. Would it be convenient for you to provide me with your preprocessed data set by email? Thank you very much.

wtangdev commented 1 year ago

Of course. Here is the link. Actually the uploaded data is directly obtained from the preprocessed data of TPLinker.

TornadoXuRocket commented 1 year ago

Thank you again. Best regards!