-
First, awesome project!
How did you train your model at https://huggingface.co/raphaelsty/neural-cherche-sparse-embed? Did you train it from scratch? I found an old copy of your sparsembed library…
-
When I used the pre-trained model 'raphaelsty/neural-cherche-sparse-embed' to evaluate the dataset, specifically, the arguana dataset, with a retrieval k value of 100, the result was very poor
{'map'…
-
Pre-training:
1. Hyperlink-induced Pre-training for Passage Retrieval in Open-domain Question Answering(ACL2022)
2. RetroMAE v2: Duplex Masked Auto-Encoder For Pre-Training Retrieval-Oriented Langua…
-
Batch size in indexing should either be on the pipeline only and inherited by the transformer, or set on the transformer only and used by the pipeline.
EmbeddingModels currently have a double batch s…
-
Hi @raphaelsty, could you please let me know what is the licence of the below splade/sparsembed models hosted in huggingface?
https://huggingface.co/raphaelsty/splade-max
https://huggingface.co/ra…