Open LuckMonkeys opened 4 days ago
Hi, thank you for your interest in our work.
Achieving consistently successful attacks remains a challenging task, so you might need to manually select successful examples. If you're looking for clear examples of successful attacks, Figures 6, 7, 8, 9, 10, and 11 in the paper might be what you're looking for.
To quickly improve the transferability of adversarial images, increasing the diversity of auxiliary models could be an effective approach. This might help generate more transferable adversarial examples across different tasks.
Hello, congratulations on your outstanding work! I have been exploring the transferability of the adversarial images generated using your method, but I encountered some performance issues.
Setting
Clean Image: ImageNet Target Image: COCO image Encoder: ViT-B/32 Decoder: checkpoints/coco_cos.pt
I followed Steps 1, 2, and 4 to generate adversarial images. My goal is to verify the transferability of these generated adversarial images in the image captioning task using the InstructBLIP and LLaVA1.5 models.
Result
Despite the adversarial images being generated, the captions produced by both models still closely resemble those of the clean images. For instance, the target image's caption is:
"A man with a red helmet on a small moped on a dirt road."
However, the captions generated by InstructBLIP and LLaVA1.5 are:
Question
Is there another configuration or method I should use to enhance the transferability of the adversarial images? Alternatively, could you provide some adversarial images generated in your experiments that demonstrate successful transferability?
Thank you for your time and assistance.