Open Aniketto16 opened 1 year ago
Hi, I'm doing the same,
I'm trying to use the https://github.com/UKPLab/sentence-transformers/blob/master/examples/training/ms_marco/train_bi-encoder_margin-mse.py script by changing queries and corpus to the target mmarco starting from a multilingual pretraiend model.
I think MarginMSE Loss should be better than MultipleNegativeRankingLoss for a single small GPU, as stated by @nreimers. Do you think this could work by using the same hard negatives and cross-encoder scores from the original MSMARCO script?
Hello @buoi, I completely agree that biencoder with MarginMSE will work the best for the use case.
But I wanted to clarify few things :
If you are not busy with some work, can you help me ready the training script! Thank you so much for the reply, looking forward to work with you.
Hii everyone!
I wanted to know the exact training procedure/script for training a Japanese bi-encoder for asymmetric search. I am planning to use the translated version of ms-marco : https://github.com/unicamp-dl/mMARCO
I am farily new to sentence transformers I don't know much about training my own model, from what I know :
I want to know the exact procedure and a sample training script would be really helpful too, but please guide me what should be my approach and thank you so much for your awesome work out here!