Literature DB >> 21892342

Differentially Private Empirical Risk Minimization.

Kamalika Chaudhuri1, Claire Monteleoni, Anand D Sarwate.   

Abstract

Privacy-preserving machine learning algorithms are crucial for the increasingly common setting in which personal data, such as medical or financial records, are analyzed. We provide general techniques to produce privacy-preserving approximations of classifiers learned via (regularized) empirical risk minimization (ERM). These algorithms are private under the ε-differential privacy definition due to Dwork et al. (2006). First we apply the output perturbation ideas of Dwork et al. (2006), to ERM classification. Then we propose a new method, objective perturbation, for privacy-preserving machine learning algorithm design. This method entails perturbing the objective function before optimizing over classifiers. If the loss and regularizer satisfy certain convexity and differentiability criteria, we prove theoretical results showing that our algorithms preserve privacy, and provide generalization bounds for linear and nonlinear kernels. We further present a privacy-preserving technique for tuning the parameters in general machine learning algorithms, thereby providing end-to-end privacy guarantees for the training process. We apply these results to produce privacy-preserving analogues of regularized logistic regression and support vector machines. We obtain encouraging results from evaluating their performance on real demographic and benchmark data sets. Our results show that both theoretically and empirically, objective perturbation is superior to the previous state-of-the-art, output perturbation, in managing the inherent tradeoff between privacy and learning performance.

Entities:  

Year:  2011        PMID: 21892342      PMCID: PMC3164588     

Source DB:  PubMed          Journal:  J Mach Learn Res        ISSN: 1532-4435            Impact factor:   3.654


  3 in total

Review 1.  Weaving technology and policy together to maintain confidentiality.

Authors:  L Sweeney
Journal:  J Law Med Ethics       Date:  1997 Summer-Fall       Impact factor: 1.718

2.  Training a support vector machine in the primal.

Authors:  Olivier Chapelle
Journal:  Neural Comput       Date:  2007-05       Impact factor: 2.026

3.  Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays.

Authors:  Nils Homer; Szabolcs Szelinger; Margot Redman; David Duggan; Waibhav Tembe; Jill Muehling; John V Pearson; Dietrich A Stephan; Stanley F Nelson; David W Craig
Journal:  PLoS Genet       Date:  2008-08-29       Impact factor: 5.917

  3 in total
  31 in total

1.  iDASH: integrating data for analysis, anonymization, and sharing.

Authors:  Lucila Ohno-Machado; Vineet Bafna; Aziz A Boxwala; Brian E Chapman; Wendy W Chapman; Kamalika Chaudhuri; Michele E Day; Claudiu Farcas; Nathaniel D Heintzman; Xiaoqian Jiang; Hyeoneui Kim; Jihoon Kim; Michael E Matheny; Frederic S Resnic; Staal A Vinterbo
Journal:  J Am Med Inform Assoc       Date:  2011-11-10       Impact factor: 4.497

2.  Privacy-Preserving Methods for Vertically Partitioned Incomplete Data.

Authors:  Yi Deng; Xiaoqian Jiang; Qi Long
Journal:  AMIA Annu Symp Proc       Date:  2021-01-25

Review 3.  Deep learning for healthcare: review, opportunities and challenges.

Authors:  Riccardo Miotto; Fei Wang; Shuang Wang; Xiaoqian Jiang; Joel T Dudley
Journal:  Brief Bioinform       Date:  2018-11-27       Impact factor: 11.622

4.  Toward practicing privacy.

Authors:  Cynthia Dwork; Rebecca Pottenger
Journal:  J Am Med Inform Assoc       Date:  2013-01-01       Impact factor: 4.497

5.  Scalable privacy-preserving data sharing methodology for genome-wide association studies.

Authors:  Fei Yu; Stephen E Fienberg; Aleksandra B Slavković; Caroline Uhler
Journal:  J Biomed Inform       Date:  2014-02-06       Impact factor: 6.317

6.  Differential privacy based on importance weighting.

Authors:  Zhanglong Ji; Charles Elkan
Journal:  Mach Learn       Date:  2013-10       Impact factor: 2.940

Review 7.  A Comprehensive Survey on Local Differential Privacy toward Data Statistics and Analysis.

Authors:  Teng Wang; Xuefeng Zhang; Jingyu Feng; Xinyu Yang
Journal:  Sensors (Basel)       Date:  2020-12-08       Impact factor: 3.576

8.  Partitioning-based mechanisms under personalized differential privacy.

Authors:  Haoran Li; Li Xiong; Zhanglong Ji; Xiaoqian Jiang
Journal:  Adv Knowl Discov Data Min (2017)       Date:  2017-04-23

9.  Differentially Private Distributed Online Learning.

Authors:  Chencheng Li; Pan Zhou; Li Xiong; Qian Wang; Ting Wang
Journal:  IEEE Trans Knowl Data Eng       Date:  2018-01-17       Impact factor: 6.977

10.  Federated Tensor Factorization for Computational Phenotyping.

Authors:  Yejin Kim; Jimeng Sun; Hwanjo Yu; Xiaoqian Jiang
Journal:  KDD       Date:  2017-08
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.