Open joslefaure opened 2 weeks ago
We recently rerun the evaluation and the score is actually a bit higher than what is reported in the paper because some bugs was fixed for the videomme data in lmms-eval.
Can you check the log output by lmms-eval and see if there is any thing unusual?
Thanks for your reply. Upon inspecting the results, I found an alarming number of !!!!!!!!!!!!!!!!
in the generated results for the following key resps
: Ex: "resps": [["!!!!!!!!!!!!!!!!"]]
. Do you have an idea what the issue might be?
I first installed lmms-eval and then installed longva following the official instructions for both projects.
I use the same script as in the Readme:
With the latest in the latest commit of lmms_eval (main branch):
bcbdc493
I get the following results:
Could you please advise on what I am doing wrong? Thanks