Closed liangeric closed 3 years ago
On second thought, probably best to still have a threshold, maybe top 5 or 10 (maybe 10 is too much?) sentences to loop through to get possibly have BERT find an answer. If we keep going until an answer is found then we might be returning an answer when in reality an answer can not be found.
I think its reasonable to do a top 5? We just need a way to know if the BERT returned answer is the true answer or as close as we can get to it. Also, still need to keep the potential of unanswerable questions.
Solution we decided upon was an ensemble method to pick better sentences, and upgrading our BERT model to train on a larger wikipedia dataset, so that answers produced were more likely to appear in top 3.
Instead of choosing top 3, maybe keep looping until BERT finds an answer? Seems like from our test cases if a sentence does not have a good answer it returns [CLS], so we can keep looping until something is found, although more testing is needed with extra questions to see if this is a good enhancement.