-
Hi -
could you provide a code snippet for how to load the model weights from
https://transformer-models.s3.amazonaws.com/2019n2c2_tack1_roberta_pt_stsc_6b_16b_3c_8c.zip
into the Roberta mod…
-
I have more information about each data point such as language and contextual data that could potentially help (maybe) for our task. The task is to generate sentence similarity embedding and labels.
…
-
Hi,
I'm trying training sentence embedding, but I only have an imbalance data, say, there are 3500 sentence pairs with a similar score 0.0, 2000 pairs with a score 0.5, while only 250 pairs with a …
-
def prepare_data(lang1_name, lang2_name, reverse=False):
input_lang, output_lang, pairs = read_langs(lang1_name, lang2_name, reverse)
print("Read %s sentence pairs" % len(pairs))
…
-
Hi there,
I'm a little confused with the terminology. I am using the model named 'all-mpnet-base-v2'.
When speaking to my team is it fair to say:
1. We are using S-BERT, a sentence transformer …
-
Trying to use `generate_qa_embedding_pairs` method to create synthetic data.
`from llama_index.finetuning import generate_qa_embedding_pairs`
I run into an error:
```
---------------------…
-
#### Link: [Semantic Scholar](https://www.semanticscholar.org/paper/A-Framework-for-Neural-Machine-Translation-by-Fuzzy-Wang-Wloka/1943b69f17e3a8ee6db041a4664795bb12951a5b?utm_source=direct_link)
###…
-
- the dataset of (prototype, revision) sentence pairs
- the sentence analogy evaluation
-
Hey, @nreimers . I have finetuned distilbert-nli-mean-tokens on my custom data. It is giving embedding for every input sentence pairs. But it is giving high cosine for irrelevant sentences. I am not g…
-
HI , amazing work , highly inspirational. Thanks a lot for make it opensource .
which datasets did you use for pre-training english only model? , it is mentioned that you used the below for pretrai…