DistilBERT can only take 512 tokens which is not enough for most passages.
So... maybe using a summarizer to give context outside of the sentence with a mask token to improve results.
Huggingface has a DistilBART model that I could use, but that would only increase the (already considerable) runtime.
Maybe make it optional?
DistilBERT can only take 512 tokens which is not enough for most passages. So... maybe using a summarizer to give context outside of the sentence with a mask token to improve results. Huggingface has a DistilBART model that I could use, but that would only increase the (already considerable) runtime. Maybe make it optional?