TR2024-009

Why Does Differential Privacy with Large ε Defend Against Practical Membership Inference Attacks?


    •  Lowy, A., Li, Z., Liu, J., Koike-Akino, T., Parsons, K., Wang, Y., "Why Does Differential Privacy with Large ε Defend Against Practical Membership Inference Attacks?", AAAI Workshop on Privacy-Preserving Artificial Intelligence, February 2024.
      BibTeX TR2024-009 PDF
      • @inproceedings{Lowy2024feb2,
      • author = {Lowy, Andrew and Li, Zhuohang and Liu, Jing and Koike-Akino, Toshiaki and Parsons, Kieran and Wang, Ye},
      • title = {Why Does Differential Privacy with Large ε Defend Against Practical Membership Inference Attacks?},
      • booktitle = {AAAI Workshop on Privacy-Preserving Artificial Intelligence},
      • year = 2024,
      • month = feb,
      • url = {https://www.merl.com/publications/TR2024-009}
      • }
  • MERL Contacts:
  • Research Areas:

    Artificial Intelligence, Machine Learning

Abstract:

For “small” privacy parameter e (e.g. e ă 1), e-differential privacy (DP) provides a strong worst-case guarantee that no membership inference attack (MIA) can succeed at determining whether a person’s data was used to train a machine learning model. The guarantee of DP is worst-case because: a) it holds even if the attacker already knows the records of all but one person in the data set; and b) it holds uniformly over all data sets. In practical applications, such a worst-case guarantee may be overkill: practical attackers may lack exact knowledge of (nearly all of) the private data, and our data set might be easier to defend, in some sense, than the worst-case data set. Such considerations have motivated the industrial deployment of DP models with large privacy parameter (e.g. ε ě 7), and it has been observed empirically that DP with large ε can successfully defend against state-of-the-art MIAs. Existing DP theory cannot explain these empirical findings: e.g., the theoretical privacy guarantees of ε ě 7 are essentially vacuous. In this paper, we aim to close this gap between the- ory and practice and understand why a large DP parameter can prevent practical MIAs. To tackle this problem, we pro- pose a new privacy notion called practical membership pri- vacy (PMP). PMP models a practical attacker’s uncertainty about the contents of the private data. The PMP parameter has a natural interpretation in terms of the success rate of a practical MIA on a given data set. We quantitatively analyze the PMP parameter of two fundamental DP mechanisms: the exponential mechanism and Gaussian mechanism. Our analysis reveals that a large DP parameter often translates into a much smaller PMP parameter, which guarantees strong privacy against practical MIAs. Using our findings, we offer principled guidance for practitioners in choosing the DP pa- rameter.

 

  • Related Publication

  •  Lowy, A., Li, Z., Liu, J., Koike-Akino, T., Parsons, K., Wang, Y., "Why Does Differential Privacy with Large Epsilon Defend Against Practical Membership Inference Attacks?", arXiv, February 2024.
    BibTeX arXiv
    • @article{Lowy2024feb,
    • author = {Lowy, Andrew and Li, Zhuohang and Liu, Jing and Koike-Akino, Toshiaki and Parsons, Kieran and Wang, Ye},
    • title = {Why Does Differential Privacy with Large Epsilon Defend Against Practical Membership Inference Attacks?},
    • journal = {arXiv},
    • year = 2024,
    • month = feb,
    • url = {https://arxiv.org/abs/2402.09540}
    • }