uclnlp / inferbeddings

Injecting Background Knowledge in Neural Models via Adversarial Set Regularisation
MIT License
59 stars 12 forks source link

Collect Hypotheses to Test #15

Closed riedelcastro closed 7 years ago

riedelcastro commented 7 years ago

For the paper and the experiment section, it would be good to be precise about the hypotheses we like to test, and how to test them. Here is a start:

Feel free to comment, edit and add more...

rockt commented 7 years ago

Experiments:

pminervini commented 7 years ago

Setting up the X-Shot Relational Learning experiment right now.

On a side note - I'm wondering if the current approach might be useful also to "mine" rules, e.g. by trying out different rules and checking whether they are violated in the embedding space

tdmeeste commented 7 years ago

Suggestions for the part on NYT experiments:

By the way, @pminervini, It appears that forcing the dummy embeddings to ones has a negative impact on the results. By not doing that, we have a somewhat 'extended' model F with an additional global weighting of all dimensions - all right if we do that? It's logical from the point of view of DistMult, and I think it will suffice if we shortly mention how we mimic model F based on DistMult.

For the part on the synthetic data experiments: (is there going to be space for that?)

rockt commented 7 years ago

For NYT, I think one of the main experiments that would be good to have and that we discussed briefly yesterday is to

This should increase the margin we see for using rules vs. not using rules quite dramatically as it is closer to the NAACL and EMNLP zero and x-shot experiments.

tdmeeste commented 7 years ago

Not sure I understand exactly what you mean.

  • extract the clauses

do you mean: take those 36 NAACL simple implications (and as model: the Model F simplification of DistMult)?

Or: use Amie+ to extract a wider set of rules of the forms q(X0, X1) :- p(X0, X1) or even r(X0, X1) :- p(X0, X1), q(X0, X1) (to be manually pruned?) for which the Model F simplification of DistMult can still be used?

Or: use Amie+ to extract a new rule set with more general rules (to be manually pruned) from the training data with Amie+, involving both FB and NYT heads and body's, and don't model embeddings for (subj, obj) together as in NAACL/EMNLP, but use TransE, ComplEx, DistMult (possibly with worse results compared to learning entity pair embeddings).

Completely agree with the setting of dropping head facts, that's what I head in mind for the experiment with the 36 simple implication clauses with FreeBase head predicates.

riedelcastro commented 7 years ago

Can’t we use the same rules and clauses we used for the NAACL/EMNLP experiments?

S

On 21 Feb 2017, at 19:27, Tim Rocktäschel notifications@github.com wrote:

For NYT, I think one of the main experiments that would be good to have and that we discussed briefly yesterday is to

take the entire dataset extract the clauses gather the set of head predicates H of all clauses subsample the dataset, but only for facts with predicates in H, i.e., for every predicate appearing as head in one of the clauses, drop 10%, 20%, ..., 100% of the facts This should increase the margin we see for using rules vs. not using rules quite dramatically as it is closer to the NAACL and EMNLP zero and x-shot experiments.

— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/uclmr/inferbeddings/issues/15#issuecomment-281452980, or mute the thread https://github.com/notifications/unsubscribe-auth/AAKMzCuAJX32Uw09jxs8YbKB_J2ZAmLzks5rezq1gaJpZM4L-LJc.

tdmeeste commented 7 years ago

sure! Only disadvantage: they're very simple. I would prefer the 36 pruned NAACL clauses, because for these there are results as a function of fraction of head facts in both papers.

@riedelcastro How about adding synthetic data experiments to for analyzing the more complex rules (see suggestions above)?

rockt commented 7 years ago

Yes, I think they would be too simple and I don't know what we would expect to see. The EMNLP approach is probably very efficient for these simple clauses.