Closed zhaitongqing233 closed 2 years ago
Hi, thanks for your interest in our work. The response to your questions are as follows:
pruned model
(described in your question), but it is a backdoored model after fine-tuning (See the Figure 1 in our paper). Actually, the cause of effectiveness for NAD is mainly due to the regularization and integration of attention maps. We have provided both intuitive analysis (see section 4.3) and experimental results that compare the defense effect on feature maps and attention maps (see Table 8), as well as the comparison of feature visualization between the different functions of attention operations (see Figure 11). We also believe a depth-reading to the whole content of our paper would benefit to your understanding of NAD.Hope this response are helpful to your research.
Hello. I am looking for the possible solution for backdoor attack. I've read the interesting and promising research, but still in confusion.
Looking for your reply.