FreedomIntelligence / LLMZoo

⚡LLM Zoo is a project that provides data, models, and evaluation benchmark for large language models.⚡
Apache License 2.0
2.93k stars 201 forks source link

How to eval and rank model outputs? #30

Open LuciusMos opened 1 year ago

LuciusMos commented 1 year ago

In https://github.com/FreedomIntelligence/LLMZoo/tree/main/llmzoo/eval, the readme lacks the 2nd section "How to evaluate?".

I am a bit confused by so many scripts and python files. Could you please write a guideline on how to run those files? Thank a lot:)