Closed riedelcastro closed 7 years ago
Experiments:
X-Shot Relational Learning
Debug an Embedding Model using Inferbeddings
(Optional) Reinforcing Rules
(EMNLP Idea) Get External Rules
Setting up the X-Shot Relational Learning
experiment right now.
On a side note - I'm wondering if the current approach might be useful also to "mine" rules, e.g. by trying out different rules and checking whether they are violated in the embedding space
Suggestions for the part on NYT experiments:
By the way, @pminervini, It appears that forcing the dummy embeddings to ones has a negative impact on the results. By not doing that, we have a somewhat 'extended' model F with an additional global weighting of all dimensions - all right if we do that? It's logical from the point of view of DistMult, and I think it will suffice if we shortly mention how we mimic model F based on DistMult.
For the part on the synthetic data experiments: (is there going to be space for that?)
For NYT, I think one of the main experiments that would be good to have and that we discussed briefly yesterday is to
This should increase the margin we see for using rules vs. not using rules quite dramatically as it is closer to the NAACL and EMNLP zero and x-shot experiments.
Not sure I understand exactly what you mean.
- extract the clauses
do you mean: take those 36 NAACL simple implications (and as model: the Model F simplification of DistMult)?
Or: use Amie+ to extract a wider set of rules of the forms q(X0, X1) :- p(X0, X1)
or even
r(X0, X1) :- p(X0, X1), q(X0, X1)
(to be manually pruned?) for which the Model F simplification of DistMult can still be used?
Or: use Amie+ to extract a new rule set with more general rules (to be manually pruned) from the training data with Amie+, involving both FB and NYT heads and body's, and don't model embeddings for (subj, obj)
together as in NAACL/EMNLP, but use TransE, ComplEx, DistMult (possibly with worse results compared to learning entity pair embeddings).
Completely agree with the setting of dropping head facts, that's what I head in mind for the experiment with the 36 simple implication clauses with FreeBase head predicates.
Can’t we use the same rules and clauses we used for the NAACL/EMNLP experiments?
S
On 21 Feb 2017, at 19:27, Tim Rocktäschel notifications@github.com wrote:
For NYT, I think one of the main experiments that would be good to have and that we discussed briefly yesterday is to
take the entire dataset extract the clauses gather the set of head predicates H of all clauses subsample the dataset, but only for facts with predicates in H, i.e., for every predicate appearing as head in one of the clauses, drop 10%, 20%, ..., 100% of the facts This should increase the margin we see for using rules vs. not using rules quite dramatically as it is closer to the NAACL and EMNLP zero and x-shot experiments.
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/uclmr/inferbeddings/issues/15#issuecomment-281452980, or mute the thread https://github.com/notifications/unsubscribe-auth/AAKMzCuAJX32Uw09jxs8YbKB_J2ZAmLzks5rezq1gaJpZM4L-LJc.
sure! Only disadvantage: they're very simple. I would prefer the 36 pruned NAACL clauses, because for these there are results as a function of fraction of head facts in both papers.
@riedelcastro How about adding synthetic data experiments to for analyzing the more complex rules (see suggestions above)?
Yes, I think they would be too simple and I don't know what we would expect to see. The EMNLP approach is probably very efficient for these simple clauses.
For the paper and the experiment section, it would be good to be precise about the hypotheses we like to test, and how to test them. Here is a start:
Feel free to comment, edit and add more...