Bounding Information Leakage in Machine Learning - Inria - Institut national de recherche en sciences et technologies du numérique Access content directly
Journal Articles Neurocomputing Year : 2023

Bounding Information Leakage in Machine Learning

Abstract

Recently, it has been shown that Machine Learning models can leak sensitive information about their training data. This information leakage is exposed through membership and attribute inference attacks. Although many attack strategies have been proposed, little effort has been made to formalize these problems. We present a novel formalism, generalizing membership and attribute inference attack setups previously studied in the literature and connecting them to memorization and generalization. First, we derive a universal bound on the success rate of inference attacks and connect it to the generalization gap of the target model. Second, we study the question of how much sensitive information is stored by the algorithm about its training set and we derive bounds on the mutual information between the sensitive attributes and model parameters. Experimentally, we illustrate the potential of our approach by applying it to both synthetic data and classification tasks on natural images. Finally, we apply our formalism to different attribute inference strategies, with which an adversary is able to recover the identity of writers in the PenDigits dataset.
Fichier principal
Vignette du fichier
2105.03875.pdf (608.62 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-04349219 , version 1 (17-12-2023)

Licence

Attribution

Identifiers

Cite

Ganesh Del Grosso, George Pichler, Catuscia Palamidessi, Pablo Piantanida. Bounding Information Leakage in Machine Learning. Neurocomputing, 2023, 534, pp.1-17. ⟨10.1016/j.neucom.2023.02.058⟩. ⟨hal-04349219⟩
36 View
8 Download

Altmetric

Share

Gmail Facebook X LinkedIn More