openai / evals

Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Other
14.76k stars 2.58k forks source link

Release 3.0.0 #1520

Closed etr2460 closed 5 months ago

etr2460 commented 5 months ago

Release 3.0.0 of evals. This is a major version bump because https://github.com/openai/evals/pull/1514 was a breaking change to how we handle zstd files