Closed BrandonFair closed 2 years ago
First, I recommend you to set learning_rate=1e-5, l2reg=1e-8 as you use the deberta-ve3-bert.
Then, if the problem is still, please test if other datasets work fine
The final attempt to fix this problem will be sharing a balanced cut of your dataset for debugging
First, I recommend you to set learning_rate=1e-5, l2reg=1e-8 as you use the deberta-ve3-bert.
Then, if the problem is still, please test if other datasets work fine
It works, thank you :). Its also performing quite well.
My dataset is quite small(5000) and very imbalanced(90% neutral). Are there any other parameters that might be worth investigating? Is there another ABSA model that might have better performance? Is it worth trying to first fine-tune the model on another dataset (such as SemEval) and then fine-tune it on my dataset?
Sorry about all the questions, but any answers will be invaluable.
You can try your last hypothesis, and I have no more recommendations. However, you can find more research about ABSA on GitHub, good luck!
Hey, Brandon, can I ask you for some help with the process?
Is it worth trying to first fine-tune the model on another dataset (such as SemEval) and then fine-tune it on my dataset?
@BrandonFair did you ever try your last question? If so, what were the results?
Hey, Brandon, can I ask you for some help with the process?
Sorry for the late reply @pepi99. Yes I can help, but it seems like you might have found a solution in another issue.
h trying to first fine-tune the model on another dataset (such as SemEval) and then fine-tune it on my dataset?
From what I recall, it worked best when I only fine tuned on the target corpus @enzo-ca. However, my target corpus and SemEval were from very different domains.
I'm having a bit of trouble when fine tuning any of the APC models to my custom APC dataset. The models seems to classify each instance as Neutral. The dataset is named and located in
citation/apc_citation.test.txt
andcitation/apc_citation.train.txt
.Snippet of the data:
Code I'm using to train:
Full stack:
Please let me know if I should supply any more information. Any help will be greatly appreciated.