Closed kamathhrishi closed 3 years ago
Hi,
Differential privacy is an effective defense against membership inference. For example if you have a prior that there is a 50% chance an image is present in the dataset (before looking at the model), then after looking at the model the probability is no more than 50% + epsilon / 4 (see Property 1 in Section 3.4 of [1]).
Of course the actual mitigation might be better than that, but this is at least an upper-bound.
[1] White-box vs Black-box: Bayes Optimal Strategies for Membership Inference, ICML'2019
Hello Team Opacus , I would like to understand if there are any examples anywhere which demonstrates how training with DP could mitigate model Inversion attacks , membership inference and other privacy attacks. I wanted it for RGB image classifiers. Thank you.