Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published April 2019 | Submitted + Published + Supplemental Material
Journal Article Open

On Connecting Stochastic Gradient MCMC and Differential Privacy

Abstract

Concerns related to data security and confidentiality have been raised when applying machine learning to real-world applications. Differential privacy provides a principled and rigorous privacy guarantee for machine learning models. While it is common to inject noise to design a model satisfying a required differential-privacy property, it is generally hard to balance the trade-off between privacy and utility. We show that stochastic gradient Markov chain Monte Carlo (SG-MCMC) – a class of scalable Bayesian posterior sampling algorithms – satisfies strong differential privacy, when carefully chosen stepsizes are employed. We develop theory on the performance of the proposed differentially-private SG-MCMC method. We conduct experiments to support our analysis, and show that a standard SG-MCMC sampler with minor modification can reach state-of-the-art performance in terms of both privacy and utility on Bayesian learning.

Additional Information

© 2019 by the author(s). This research was supported in part by DARPA, DOE, NIH, NSF and ONR. We thank Ruiyi Zhang for providing the code base.

Attached Files

Published - li19a.pdf

Submitted - 1712.09097.pdf

Supplemental Material - li19a-supp.pdf

Files

1712.09097.pdf
Files (1.6 MB)
Name Size Download all
md5:6c013b9943da74ad15a23e4625252305
310.2 kB Preview Download
md5:49139820a7b8e9629f4825247addacb9
757.4 kB Preview Download
md5:b0a0fd4c36bc78a5cc561d41026f0cbc
502.9 kB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 19, 2023