UKPLab / sentence-transformers

State-of-the-Art Text Embeddings
https://www.sbert.net
Apache License 2.0
15.28k stars 2.48k forks source link

what's the difference between USE and then SBERT? #64

Closed Cumberbatch08 closed 4 years ago

Cumberbatch08 commented 4 years ago

First, many thanks to your paper and code. But I read the universal sentence encoder(USE) paper, the architecture is like simaese network, they also used the SNLI dataset. But your result is well performed. So I'm very interested in your work.

nreimers commented 4 years ago

Hi @Cumberbatch08 sadly the USE papers (at least the ones I know) are extremely high-level, not going really into the details. So it is unclear which architecture they exactly used and how the training was done (exact datasets, exact loss function etc.)

Differences:

Cumberbatch08 commented 4 years ago

haha, yes, absolutely agreed what you said. The USE don't public much more details, such as the layers, dataset, loss etc. I get some information about the architecture: image Just as you said, maybe the pretraining is important.

Gurutva commented 2 years ago

what would be best USE (https://tfhub.dev/google/universal-sentence-encoder/4) or SBERT models (https://huggingface.co/sentence-transformers) for good semantic search results ?

nreimers commented 2 years ago

@Gurutva SBERT works much better: https://arxiv.org/pdf/2104.08663v1.pdf