Private Posterior Distributions from Variational Approximations

Citation:

Vishesh Karwa, Dan Kifer, and Aleksandra Slavkovic. 2016. “Private Posterior Distributions from Variational Approximations.” NIPS 2015 Workshop on Learning and Privacy with Incomplete Data and Weak Supervision. arXiv Version
PDF436 KB

Abstract:

Privacy preserving mechanisms such as differential privacy inject additional randomness in the form of noise in the data, beyond the sampling mechanism. Ignoring this additional noise can lead to inaccurate and invalid inferences. In this paper, we incorporate the privacy mechanism explicitly into the likelihood function by treating the original data as missing, with an end goal of estimating posterior distributions over model parameters. This leads to a principled way of performing valid statistical inference using private data, however, the corresponding likelihoods are intractable. In this paper, we derive fast and accurate variational approximations to tackle such intractable likelihoods that arise due to privacy. We focus on estimating posterior distributions of parameters of the naive Bayes log-linear model, where the sufficient statistics of this model are shared using a differentially private interface. Using a simulation study, we show that the posterior approximations outperform the naive method of ignoring the noise addition mechanism.