I am trying to reproduce (with some differences), the results obtained for the e5 models family.
The second stage fine-tuning is perfectly reproducible with this repo (I am using the v2)
However, for the contrastive pre-training, I wanted to change the cross-entropy loss of the EncoderModel to the InfoNCE loss. This should be enough, right?
Also, it is not clear to me about how perform only in-batch negatives tuning with this repo. I did not found any option on the trainer, and when I have examples without negative_passages I obtain an error.
I am trying to reproduce (with some differences), the results obtained for the e5 models family.
The second stage fine-tuning is perfectly reproducible with this repo (I am using the v2)
However, for the contrastive pre-training, I wanted to change the cross-entropy loss of the
EncoderModel
to theInfoNCE
loss. This should be enough, right?Also, it is not clear to me about how perform only in-batch negatives tuning with this repo. I did not found any option on the trainer, and when I have examples without
negative_passages
I obtain an error.Thank you.