Closed sheryc closed 8 months ago
Thanks so much for your valuable suggestions !!! I will fix this bug as soon as possible.
I have updated the code to fix the typo and unify the loyalty and actuality tests! Thank you again!
As for the first issue, changing the prompt will have an influence on the results and we are currently unable to re-run these baselines again and update the results in the paper. So I will keep the prompt unchanged in this repo. To my experience, this won't have a significant impact on the results. If it has in your case, you can also change fix it, and re-run the baselines to compare with your model.
Hi, thank you for the great work! I've found some problems with the sci_fi dataset.
"2.\tThere exists a planet in Sirius star system that has gravity similar to Earth and signs of life but without a breathable atmosphere, True or False? Answer this question based on the world described in the document."
). The "2.\t" is sometimes "Question2:", where the prompt would contain two "Question"s. Should the question number be removed from the prompt?Also, there's a typo in the prompt "Please directly give answer without any additonal output or explanation." in https://github.com/OpenLMLab/LEval/blob/745aa8c5f0e3ef37010f9d7634f20ec30f017c01/Baselines/llama2-chat-test.py#L121. additonal -> additional.