zeno-ml / zeno-evals-hub

https://zeno-ml-openai-evals.hf.space/
MIT License
6 stars 3 forks source link

Zeno + OpenAI Evals

Github Actions CI tests MIT license Discord

OpenAI's Evals library is a great resource providing evaluation sets for LLMS.

This repo provides a hub for exploring these results using the Zeno evaluation tool.

Add New Evals

To add new evals, add a new entry to evals/evals.yaml with the following fields:

Make sure you test your evals locally before submitting a PR!

Running

poetry install

python -m zeno-evals-hub evals/evals.yaml