Differentially private Bayesian learning on distributed data

Mikko Heikkilä, Eemil Lagerspetz, Samuel Kaski, Kana Shimizu, Sasu Tarkoma, Antti Honkela

    Research output: Contribution to journalConference article

    4 Citations (Scopus)

    Abstract

    Many applications of machine learning, for example in health care, would benefit from methods that can guarantee privacy of data subjects. Differential privacy (DP) has become established as a standard for protecting learning results. The standard DP algorithms require a single trusted party to have access to the entire data, which is a clear weakness, or add prohibitive amounts of noise. We consider DP Bayesian learning in a distributed setting, where each party only holds a single sample or a few samples of the data. We propose a learning strategy based on a secure multi-party sum function for aggregating summaries from data holders and the Gaussian mechanism for DP. Our method builds on an asymptotically optimal and practically efficient DP Bayesian inference with rapidly diminishing extra cost.

    Original languageEnglish
    Pages (from-to)3227-3236
    Number of pages10
    JournalAdvances in Neural Information Processing Systems
    Volume2017-December
    Publication statusPublished - 2017 Jan 1
    Event31st Annual Conference on Neural Information Processing Systems, NIPS 2017 - Long Beach, United States
    Duration: 2017 Dec 42017 Dec 9

    ASJC Scopus subject areas

    • Computer Networks and Communications
    • Information Systems
    • Signal Processing

    Fingerprint Dive into the research topics of 'Differentially private Bayesian learning on distributed data'. Together they form a unique fingerprint.

  • Cite this

    Heikkilä, M., Lagerspetz, E., Kaski, S., Shimizu, K., Tarkoma, S., & Honkela, A. (2017). Differentially private Bayesian learning on distributed data. Advances in Neural Information Processing Systems, 2017-December, 3227-3236.