using the official open-source weights hkunlp/instructor-large · Hugging Faceto test the retrieval task ArguAna, and the ndcg_at_10 metric can reach 0.57.
The training parameters are the same as described in the paper, with a batch_size=4 and training for 20k steps. The resulting model performance is as follows:
step:1k
step:20k
step:54K
0.52
0.4957
0.47
I have not been able to reproduce the target result: ndcg_at_10=0.57
using the official open-source weights hkunlp/instructor-large · Hugging Faceto test the retrieval task ArguAna, and the ndcg_at_10 metric can reach 0.57.
Next, I used the following command for training:
I have not been able to reproduce the target result: ndcg_at_10=0.57
Other people have also encountered similar issues: https://github.com/xlang-ai/instructor-embedding/issues/42
How can I train the model to reproduce the results of the paper? Can you help me with this, please?