Closed MarkRx closed 1 month ago
The correct way to address this is via prompt adjustments (maybe dedicated for PRs with small changes). Not by a threshold to ignore PRs
The PR that broke the world was small. PR Agent should review all code changes, no matter if they are small or big.
I think that the main cause for this problem (which I am not debating; I am also aware of it) is when the model doesn't have suggestions to give, it fallback to the silliest option, which is to "hallucinate" the PR content.
Also interesting is that it happens both to GPT4 and Claude (I guess you saw it with GPT4. i now have an example where I see it with Claude)
This change should prevent/improve those problems
We've found that the LLM is more likely to hallucinate or provide "noise" suggestions on small changes (such as one line changes). We suspect it is because the bot has less context to work with so is more likely to "reach for straws".
I suggest configuration option(s) to set the "minimum diff/context/PR" size.
It could be:
When the bot skips processing it should provide some kind of feedback to that effect so the end user knows it ran (and didn't crash).
I could also see:
The defaults would be set to 0 for backwards compatibility.