Open brandon-dubbs opened 8 months ago
There are a few examples of impersonation prompts that have been used to make LLMs say whatever the user wants. This is often toxic. An example is included here: ChatGPT-Unleashed. This probably falls under the broader DAN attack umbrella.
oh, thanks. this could fit into the DANs, agree
There are a few examples of impersonation prompts that have been used to make LLMs say whatever the user wants. This is often toxic. An example is included here: ChatGPT-Unleashed. This probably falls under the broader DAN attack umbrella.