-
Appreciate your excellent works!
Recently, we use autodan turbo whose strategies trained on gpt4o-gpt4o adversarial attack to jailbreak the gpt-o1-preview and got 74.6 asr finally, which is little…
-
### What's the name of your attack?
AIR
### What's the title of the paper where you present your attack?
You Know What I'm Saying - Jailbreak Attack via Implicit Reference
### What's the URL of th…
-
-
Hello,
When using OpenAI models as attack model in TAP&PAIR, an error happens:
```
Traceback (most recent call last):
File "PATH/easyjailbreak/run_PAIR.py", line 43, in
attacker.attack(s…
-
from easyjailbreak.models.huggingface_model import (HuggingfaceModel,
from_pretrained)
from easyjailbreak.models.openai_model import OpenaiModel
…
-
I tried to follow your paper but it did not work in any case. The setting described in the paper is vague and hard to follow and reproduce your results. No way to benchmark with other research or deve…
-
Hi,
I mentioned about this on discord:
https://discord.com/channels/1156626270772269217/1272542948848042086
**Describe the bug**
I'm using AzureOpenAiChatModel to connect to Azure OpenAI and …
-
This seems like a very important finding mentioned in your [blog](https://huggingface.co/blog/leaderboard-decodingtrust) and something deserving of further exposition.
Submitting your paper to Gemi…
-
### Did you check docs and existing issues?
- [x] I have read all the NeMo-Guardrails docs
- [x] I have updated the package to the latest version before submitting this issue
- [x] I have searched th…
-
✅ I confirm that I have searched existing issues and couldn't find a similar bug.
✅ I have read the documentation carefully and it couldn't resolved my problem.
### Describe the bug
Apparently th…