Open KarstenLasse opened 2 years ago
Hi, @KarstenLasse The evaluation of ATEPC inference is not available now, I will work on it later.
Because the evaluation of APC is based on the results of ATE, so it is different. However there is a Dutch dataset here to evalute the peroformance in training.
Hi, @KarstenLasse The evaluation of ATEPC inference is not available now, I will work on it later.
Really appreciate it, I also look forward the evaluation :)
is there any update on this problem? thank you.
Dear @yangheng95,
Thanks for making and maintaining this repo, it's great!
I have some trouble to get the accuracy and F1 scores for the Restaurant Test data Gold. (Ideally I want to make a confusion matrix). What is the easiest way to get F1 scores for APC & ATE after running a checkpoint model on test data? Does the model store these metrics somewhere?
Alternatively, how do you compare your predictions to the TRUE test data (Restaurant Test data Gold annotated)? I can easily transform the models' predictions ('atepc_inference.result_json') to a pandas dataframe. But it is very hard to transform the test data stored in integrated datasets (from ABSAdatasets) (it is in IOB format) to that exact same format (pandas dataframe) in order to test performance. Do you have a script for that, or a certain function? I was not able to find it.
Btw: I used the multilingual checkpoint model (FAST-LCF-ATEPC) on the Restaurant14 Test data Gold (But, ultimately I want to use this model on Dutch data. That is why I want to know how to test performance).
Thanks a lot,
Karsten
Code: