haesleinhuepf / human-eval-bia

Benchmarking Large Language Models for Bio-Image Analysis Code Generation
MIT License
19 stars 11 forks source link

Guarding against LLMs that would learn our repo "by heart" #119

Open tischi opened 1 week ago

tischi commented 1 week ago

Even though I think LLMs generally do not work like this, I still wonder whether we could guard against some - otherwise super dumb - LLM to just learn our repo by heart and then achieve great results.

Given the discussions in https://github.com/haesleinhuepf/human-eval-bia/pull/118 I wonder whether we could somehow maintain a separate secret branch where we ask the conceptually same questions but just with a slightly modifications?

Maybe:

It would be a bit of work...but maybe worth it?

What do you think?

haesleinhuepf commented 1 week ago

I'm currently working on training such an LLM, because I wanted to know how to achieve this:

and then achieve great results.

When I'm done, I'll share it (+ training scripts) and we can develop a strategy against it.