GodXuxilie / PromptAttack

An LLM can Fool Itself: A Prompt-Based Adversarial Attack (ICLR 2024)
56 stars 11 forks source link