Open navneet1v opened 2 years ago
This would be a great step into "productifying" this plugin. Is there already an idea about what criteria the benchmarking corpora should fulfill? E.g. size, domain, nature of queries (keywords, passages, questions, etc.). One good starting point would be the BEIR dataset. And what about the models used? It probably makes sense to have a fixed baseline model and to make it easy to extend the benchmarking by adding new models to it.
@br3no We have used the BEIR datasets and one specific model only. All the details that you have requested is not updated right now on the issue. Will try to update them ASAP for more visibility.
Added the initial Commit here: https://github.com/navneet1v/neural-search/tree/perf-testing/benchmarks/osb
I have opened a github issue to add the benchmarks workload in OpenSearch benchmarks workload repo. Also, I have started the work on the same.
Description
Perform the benchmarks for the Query via the new "neural" query type.
Benchmarking Search API
This will provide insights around performance of the new Query type (neural) that we are adding in OpenSearch. We will be using Open Search Benchmarks to perform this.
Metrics Identified:
Relevance of Search
This benchmark will concrete results which we have got from experiments that we have done by combining the BM-25 and K-NN score separately. This will also provide insights around if we need to boost scores for one query type or not and when to boost it.
Metrics Identified:
Appendix
Science Experiment Metrics