Open Spico197 opened 1 year ago
Thank you for your interest in our work. The bert-serving is your understanding, and we updated the datasets.py to include the code generating repeat.json. Also, in experiments, we have used the development set and conducted experiments. The corresponding average effect (Avg) is 94.24%. Since the experimental hyperparameters were set according to manual experience and not adjusted, following the source code of the paper "Relational Graph Attention Network for Aspect-based Sentiment Analysis", we used the test set to evaluate the performance of the model. Regarding metrics, we employ token classification due to tokens as the prediction objects, which we have reported in the paper.
Thank you very much for the response and the updated code.
I'm still confusing about the metrics. I've understood the token classification metrics. However, the other baseline systems for comparison use event-instance-based metrics, maybe it is not fair to compare different metrics and claim a new SOTA result ? Have you ever tried to evaluate the model via the F1-score metric introduced in Doc2EDAG?
It is unfair to compare with other works using different metrics and claim that you are the new SOTA. Please @hawisdom 😂
Hi there. Thanks for the excellent work. We are shocked by such a huge performance improvement. When reproducing the results, I encountered the following problems. I'd be very appreciative if you could help me solve them:
bert-serving
setting. Is it justbert-serving-start -model_dir /tmp/english_L-12_H-768_A-12/ -num_worker=4
?repeat.json
file in https://github.com/hawisdom/EDEE/blob/8cfd4f2e8128e13db2dab5e99876571d634651bb/datasets.py#L98 ?company.txt
and add those names into the user dict to avoid unexpected segmentation. But it brings boundary leakage.Thanks for your kindness and looking forward to your reply.