BrambleXu / knowledge-graph-learning

A curated list of awesome knowledge graph tutorials, projects and communities.
MIT License
736 stars 120 forks source link

EMNLP-2019/11-Commonsense Knowledge Mining from Pretrained Models #259

Open BrambleXu opened 5 years ago

BrambleXu commented 5 years ago

Summary:

Resource:

Paper information:

Notes:

Knowledge bases are represented as a graph, with nodes consisting of conceptual entities (i.e. dog , running away, excited, etc.) and the pre-defined edges representing the nature of the relations between concepts (IsA, UsedFor, CapableOf, etc.). Commonsense knowledge base completion (CKBC) is a machine learning task motivated by the need to improve the coverage of these resources.

In this formulation of the problem, one is supplied with a list of candidate entityrelation-entity triples, and the task is to distinguish which of the triples express valid commonsense knowledge and which are fictitious (Li et al., 2016). 严格来说,这其实是个分类问题。输入时triple,输出是二分类结果。

现在的一些方法在held-out test set上效果好,但是在新数据上效果差。因此我们提出了不要在某个特定的数据集上训练模型,而是直接利用 the world knowledge of large language models to identify commonsense facts directly。用language model来模拟真实性。

For our masked model, we use BERT-large (Devlin et al., 2018). For sentence ranking, we use the GPT-2 117M LM (Radford et al., 2019).

Model Graph:

Result:

image

Though our approach performs worse on a held-out test set developed by Li et al. (2016), it does so without any previous exposure to the ConceptNet database, ensuring that this performance is not biased.

Thoughts:

Next Reading: