-
Thanks for your works!
Now I want to evaluation the model by test data, but I failed!
In examples i found this:
#3) Create a sentence transformer model to glue both models together
model = Sente…
-
Hi, I am working on scoring subjective answers. And now I'm wondering if the textual similarity score should be closed to the label normalized value?
I try this work on the `STS` dataset and here i…
-
hello, is there a script to run the evaluation as mentioned on the website:
[https://www.sbert.net/docs/pretrained_models.html#sentence-embedding-models](https://www.sbert.net/docs/pretrained_models.…
-
hi @nreimers , it's a nice repo. When I read your code in training_stsbenchmark_bilstm.py. I want to test the performance of bert + bilstm, but maybe there is a bug in lstm. I have read all issues abo…
-
Hello,
I have a question regarding the specific settings used for the Llama-7B model in Figure 4 of the BeLLM paper.
Following the steps in the README, I was able to successfully reproduce the r…
-
Hi, I'd like to evaluate the accuracy of AllNLI task during training. Could you kindly provide an example for that? Thank you!
-
I'm fine-tuning my model (neuralmind/bert-large-portuguese-cased) for STS task using assin dataset, which contains pairs in Portuguese for RTE and STS tasks, I'm facing a problem, when I try 1 epoch, …
-
Can you please point me towards the standard datasets being used in 10.1, 10.2, 10.3 and 10.4.
-
Hi @vaibhavad,
I’ve been trying to reproduce your work using the recently published MTEB evaluation script. However, the results of the tested subtasks using the Mistral-7b-Instruct-v2-nmtp-unsuper…
-
Hi All,
I'm trying to fine-tune an existing sentence-transformer model (all-MiniLM-L6-v2) to get better scores in my sentence similarity problem. Test data shows ~70% accuracy and I'd like to impro…