deadbits / vigil-llm

⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
https://vigil.deadbits.ai/
Apache License 2.0
304 stars 35 forks source link

Dataset eval helper #44

Open deadbits opened 11 months ago

deadbits commented 11 months ago

Vigil should have a helper utility to assist users in evaluating datasets against the detection mechanisms and return true positives, false positives, false negatives, accuracy score, etc.

I'm thinking a simple CSV of prompts, classifications, and categories.