Closed daegonYu closed 1 week ago
Hello, @daegonYu. We performed pretraining for 1 epoch. For the finetuning process, we performed 2 epochs for warm-up (1 epoch for dense embedding, 1 epoch for sparse embedding and multi-vectors), and 1 epoch for unified-finetuning.
You said you warmed up each epoch with 1 epoch for dense embedding and 1 epoch for sparse embedding and multi-vectors. Do you provide code to learn dense embedding and sparse embedding and multi-vectors separately?
We provide the corresponding arguments to achieve this:
For more details, you can refer to our code.
Oh thank you!
It says that pretraining is 25,000 steps and finetuning is 6000 steps for warm up only. Can I know the number of learning epochs for pretraining and finetuning including warm up? I have taken part of the paper below.
M3-Embedding: Multi-Linguality, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation