etherlabsio / ai-engine

Core AI services and functions powering the ETHER Platform
MIT License
0 stars 0 forks source link

Inconsistent relevance scores while comparing sentences. #81

Open reaganrewop opened 5 years ago

reaganrewop commented 5 years ago

metric_comparison

Based on the above results, the relevance between Out of domain and Out of the domain sentences gives Inconsistent results even with our Metric similarity (NSP + Cosine) and the model/similarity is not able to differentiate between subtopics.

While the above issues are taken care in the current PIMs approach (because the comparison always involves one In-domain sentence and, ranking within topics hasn't yet been our focused goal), it is not guaranteed to work on other tasks like community detection.

shashankpr commented 5 years ago

Maybe some references for text/passage relevance:

  1. https://arxiv.org/abs/1904.07531
  2. https://paperswithcode.com/paper/passage-re-ranking-with-bert
  3. http://docs.deeppavlov.ai/en/master/components/bert.html
arXiv.org
Understanding the Behaviors of BERT in Ranking
This paper studies the performances and behaviors of BERT in ranking tasks. We explore several different ways to leverage the pre-trained BERT and fine-tune it on two ranking tasks: MS MARCO passage reranking and TREC Web Track ad hoc document ranking. Experimental results on MS MARCO demonstrate the strong effectiveness of BERT in question-answering focused passage ranking tasks, as well as the fact that BERT is a strong interaction-based seq2seq matching model. Experimental results on TREC show the gaps between the BERT pre-trained on surrounding contexts and the needs of ad hoc document ranking. Analyses illustrate how BERT allocates its attentions between query-document tokens in its Transformer layers, how it prefers semantic matches between paraphrase tokens, and how that differs with the soft match patterns learned by a click-trained neural ranker.
Papers With Code : Passage Re-ranking with BERT
🏆 SOTA for Passage Re-Ranking on MS MARCO(MRR metric)
BERT in DeepPavlov — DeepPavlov 0.4.0 documentation