ebagdasa / backdoors101

Backdoors Framework for Deep Learning and Federated Learning. A light-weight tool to conduct your research on backdoors.
MIT License
335 stars 82 forks source link

Questions about the low benign accuracy on CIFAR-10 and GTSRB dataset of Blind Backdoor #17

Closed THUYimingLi closed 2 years ago

THUYimingLi commented 2 years ago

Hi, Eugene Bagdasaryan,

Congratulations on the acceptance of your paper `Blind Backdoors in Deep Learning Models' and thanks for the sharing of its codes.

However, when we run your code on CIFAR-10 dataset and GTSRB dataset, we get a very low benign accuracy (CIFAR: BA: 18.24, ASR: 98.64; GTSRB: BA: 5.7, ASR: 100) with the default settings in your codes. (PS: we get satisified results on MNIST (BA: 98.86, ASR: 99.99)). We are not for sure where the problems are or whether you used different settings in the experiments of your paper. Can you kindly help us for this problem?

Besides, we also reproduce your codes in our open-sourced toolbox (https://github.com/THUYimingLi/BackdoorBox/blob/main/core/attacks/Blind.py) based on your codes and we meet the same problem. I would be very grateful if you can also help us to check our reproduced codes.

Best Regard, Yiming Li

ebagdasa commented 2 years ago

Hey Yiming, let me double check, I tried to simplify the API and might have broken something.

THUYimingLi commented 2 years ago

Hey Yiming, let me double check, I tried to simplify the API and might have broken something.

Thank you for your helps and looking forward to your results :)

ebagdasa commented 2 years ago

I think my config for cifar confused pre-trained models, but I think simply using a fixed scale and not MGDA will work better.

THUYimingLi commented 2 years ago

I think my config for cifar confused pre-trained models, but I think simply using a fixed scale and not MGDA will work better.

Thanks for your response. I just want to make sure there is no misunderstanding. The model is trained from the scratch instead of fine-tuning on a pre-trained benign model and we adopt the same settings used in your codes. In this case, why we can not reproduce your results?

ebagdasa commented 2 years ago

Hey, I actually never tested CIFAR as we did Imagenet in the paper so I didn't check it thoroughly, sorry.

I believe it's fixed now. I added a threshold to start poisoning after normal loss reaches 1.0 (we have a fancier idea in the paper on Figure 10 but I think this one works well). I also updated the parameters for the CIFAR run. Can you please let me know if it fixed your issue.

CHR-ray commented 2 years ago

I think my config for cifar confused pre-trained models, but I think simply using a fixed scale and not MGDA will work better.

hello, I am also try to reproduce your experiment, can you provide your fixed scale that you get? Thank you!

ebagdasa commented 2 years ago

I guess you can just set it to 0.9

CHR-ray commented 2 years ago

I guess you can just set it to 0.9

Thank you for your fast reply!

Here you mean clean:backdoor = 1 : 0.9 or clean:backdoor = 0.1 : 0.9 ? Thank you!