bigcode-project / bigcode-evaluation-harness

A framework for the evaluation of autoregressive code generation language models.
Apache License 2.0
771 stars 201 forks source link

Add metric name to outfile #30

Closed loubnabnl closed 1 year ago

loubnabnl commented 1 year ago

Issue https://github.com/bigcode-project/bigcode-evaluation-harness/issues/24 Changed docstring of postprocess_results to require output to be a dict indicating metric name This is needed for BLEU tasks, but code_eval already outputs pass@k which is the correct metric name.

pass@1 0 0 Pasted Graphic 1