DevQualityEval: An evaluation benchmark 📈 and framework to compare and evolve the quality of code generation of LLMs.
57
stars
3
forks
source link
Log model responses directly to file and reuse them for debugging #181
Open
bauersimon opened 2 weeks ago
Goal, be able to use exactly 1:1 responses from a previous run to debug the evaluation logic.