castorini / rank_llm

Repository for prompt-decoding using LLMs (GPT3.5, GPT4, Vicuna, and Zephyr)
http://rankllm.ai
Apache License 2.0
273 stars 35 forks source link

Time profiling of RankVicunna or RankZephyr zero-shot evaluation/ inference on BEIR datasets #108

Open cramraj8 opened 2 months ago

cramraj8 commented 2 months ago

Hi, I wonder the time profiling of each LLMs to run across queries for re-ranking. I am running RankVicunna and RankZephyr on the zero-shot setting across BEIR datasets. For FiQA (648 queries) to conduct re-ranking of BM25 top-100 documents, RankVicunna takes ~4.5 hrs on a powerful machine (H100 GPU). The calculation leads to 30-40 seconds per query to re-rank. I wonder if this is the ideal time profiling anyone observed, or the code can be optimized with different window size or strides. Thanks in advance!