deadbits / vigil-llm

⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
https://vigil.deadbits.ai/
Apache License 2.0
270 stars 32 forks source link

Add paraphrasing #33

Open deadbits opened 9 months ago

deadbits commented 9 months ago

Update API so that users can submit prompts with the "paraphrase" flag. The submitted prompt will be paraphrased by an LLM and returned with the normal scan results.

Idea is that if the prompt contains an injection, paraphrasing while retaining the overall meaning may break injection strings.

Probably something like this POST /analyze/prompt?paraphrase=true