Closed kartikgupta-at-anu closed 4 years ago
From my experience of usage, for undefended small models, the complete verification approaches such as MILP verifier and AI2 are good choices. Because their tight bounds exactly show the distance of the closest adversarial samples. For undefended large models, these certification tools may not be useful. Because the complete verification approaches cannot finish in a reasonable time, and incomplete verification approaches are too conservative to be used as comparison criteria. In this situation, I would suggest some heuristic such as CLEVER score (https://arxiv.org/pdf/1801.10578.pdf).
Thanks for the prompt response and suggestion. I think CLEVER being dependent on input gradients of network, suffers from gradient masking issues. Correct me if I'm wrong.
Yes, you are right - CLEVER is dependent heavily on the input gradients. But for undefended classifiers, they may not use gradient obfuscation or masking so I guess it should be fine.
Can you suggest any of these certification methods be used for comparing robustness of two undefended (not trained on smooth/noisy images) networks (like ResNet), ideally without backpropagating through the network (treating n/w as blackbox), to avoid suffering from gradient masking?