Conference paper
Proceedings of the International Conference on Machine Learning (ICML), 2019
Assistant Professor of Stats & Data Science at UChicago
APA
Click to copy
Schein, A., Wu, Z. S., Schofield, A., Zhou, M., & Wallach, H. (2019). Locally Private Bayesian Inference for Count Models. In Proceedings of the International Conference on Machine Learning (ICML).
Chicago/Turabian
Click to copy
Schein, Aaron, Zhiwei Steven Wu, Alexandra Schofield, Mingyuan Zhou, and Hanna Wallach. “Locally Private Bayesian Inference for Count Models.” In Proceedings of the International Conference on Machine Learning (ICML), 2019.
MLA
Click to copy
Schein, Aaron, et al. “Locally Private Bayesian Inference for Count Models.” Proceedings of the International Conference on Machine Learning (ICML), 2019.
BibTeX Click to copy
@inproceedings{aaron2019a,
title = {Locally Private Bayesian Inference for Count Models},
year = {2019},
author = {Schein, Aaron and Wu, Zhiwei Steven and Schofield, Alexandra and Zhou, Mingyuan and Wallach, Hanna},
booktitle = {Proceedings of the International Conference on Machine Learning (ICML)}
}
Abstract: We present a general method for privacy-preserving Bayesian inference in Poisson factorization, a broad class of models that includes some of the most widely used models in the social sciences. Our method satisfies limited precision local privacy, a generalization of local differential privacy, which we introduce to formulate privacy guarantees appropriate for sparse count data. We develop an MCMC algorithm that approximates the locally private posterior over model parameters given data that has been locally privatized by the geometric mechanism (Ghosh et al., 2012). Our solution is based on two insights: 1) a novel reinterpretation of the geometric mechanism in terms of the Skellam distribution (Skellam, 1946) and 2) a general theorem that relates the Skellam to the Bessel distribution (Yuan & Kalbfleisch, 2000). We demonstrate our method in two case studies on real-world email data in which we show that our method consistently outperforms the commonly-used naive approach, obtaining higher quality topics in text and more accurate link prediction in networks. On some tasks, our privacy-preserving method even outperforms non-private inference which conditions on the true data.