Open lvilnis opened 10 years ago
Or is it that patterns aren't features and we have patterns for the test data?
The problem is that we're working in the inductive setting, so we don't have access to test data when we're learning the model. Thus, for "test entities", we don't have an embedding, and therefore, cannot predict for them directly.
What we do have are the features and patterns at test time. So, just so that we can compare against a fair universal schema baseline, we need to predict test relations. Computing an embedding for the test entities somehow is one option. Options:
Once we data loaded in and train embeddings learned, implementing the above approaches will be straightforward. We can then stream in the test sentences/entities, and evaluate which of these works best
This should be pretty much the same as getting embeddings for train data - is the issue that train data is distant supervision-y and aggregated across a bunch of instances?