Closed lshowway closed 2 years ago
As we used the results reported in the original papers, please refer to the corresponding papers.
@ikuyamada LUKE reports the results based on ROBERTA-large
, while the mentioned baselines are not based on ROBERTA-large
, e.g., KnowBert is based on bert-base-uncased
. Therefore, I have a question, are LUKE and mentioned baselines compared fairly?
LUKE is based on RoBERTa-large because at the time of writing the paper, the state-of-the-art model of our entity-related tasks was the K-Adapter model which is based on RoBERTa-large. Although comparison with a model based on a smaller PLM (e.g., KnowBERT) may not be a fair comparison, it is also difficult to run expensive pretraining multiple times due to our limited computational budget.
The reported results of
ERNIE, KEPLER, KnowBERT, K-Adapter
are based onROBERTA-large
?