UKPLab / sentence-transformers

Multilingual Sentence & Image Embeddings with BERT
https://www.SBERT.net
Apache License 2.0
14.69k stars 2.42k forks source link

Requesting checkpoints for cross-encoder/ms-marco-* for Bert-base or Bert-large models #2307

Open cramraj8 opened 11 months ago

cramraj8 commented 11 months ago

Hi @nreimers

I would like to use Bert-base and Bert-large versions cross-encoder trained on ms-marco. I tried to fine-tune "cross-encoder/ms-marco-MiniLM-L-12-v2" on NQ and other standard datasets. But the fine-tuned performance drops. I suspect it's because of the actual objective of MiniLM is the distillation regression loss thus a cross-entropy based fine-tuning doesn't help with improved performance. Therefore, I am planning on fine-tuning the source bert-base/ bert-large/ albert-large that was initially used to distill MiniLM. Do you have those checkpoints ? And if so would you mind sharing them ? Thanks in advance!

tomaarsen commented 9 months ago

Hello!

I don't have access to the checkpoints, but I do want to point you to these places:

Perhaps there's some interesting information there for you.