McGill-NLP / llm2vec

Code for 'LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders'
https://mcgill-nlp.github.io/llm2vec/
MIT License
816 stars 59 forks source link

Does performance increase when learning more steps? #106

Open daegonYu opened 1 week ago

daegonYu commented 1 week ago

hello!

In the LLM2Vec paper, only 1000 steps were trained when learning MNTP and SimCSE. Will learning more steps result in higher performance? Or is the performance improvement minimal? Also, since it is an English-based model when learning other languages, do you expect it to help performance when learning more steps than 1000 steps? Your answer would be of great help!