Open we1559 opened 1 year ago
[1] A Large Scale Search Dataset for Unbiased Learning to Rank. https://arxiv.org/abs/2207.03051 [2] Can clicks be both labels and features? Unbiased Behavior Feature Collection and Uncertainty-aware Learning to Rank. [3] Approximated doubly robust search relevance estimation.
2023年3月27日 19:11,we1559 @.***> 写道:
Hi, I find we only use query, title, abstract to train the model. But there is a lot of other features in dataset, including Continuous Value and Discrete Number.
How could I use these features to finetune an unbiased LTR model such as dla?
It seems very strange to concat the fc3 output of Transformer4Ranking/model.py and the original feature values.
Hope to get your reply.
— Reply to this email directly, view it on GitHub https://github.com/ChuXiaokai/baidu_ultr_dataset/issues/14, or unsubscribe https://github.com/notifications/unsubscribe-auth/AHPSGYBFBHUXSKT7XAYCQFTW6FYULANCNFSM6AAAAAAWJBF2WA. You are receiving this because you are subscribed to this thread.
Thanks for your reply, I will try the THUIR_WSDM_Cup.
Actually, I'm confused about how to use other features such as tf-idf with query-title-abstract transformer model.
I could only design a model combine the emdedding of the transformer and the original value of other features. But it seems strange.
Could you give me some advice?
I think you may refer to "hybrid retrieval model", which combines both sparse retrieval, such as bm25, and dense retrieval, as the feature output from this repo.
Hi, I find we only use query, title, abstract to train the model. But there is a lot of other features in dataset, including Continuous Value and Discrete Number.
How could I use these features to finetune an unbiased LTR model such as dla?
It seems very strange to concat the fc3 output of Transformer4Ranking/model.py and the original feature values.
Hope to get your reply.