An Uplink Communication-Efficient Approach to Featurewise Distributed Sparse Optimization With Differential Privacy

被引:6
作者
Lou, Jian [1 ]
Cheung, Yiu-ming [2 ]
机构
[1] Emory Univ, Dept Comp Sci, Atlanta, GA 30322 USA
[2] Hong Kong Baptist Univ, Dept Comp Sci, Hong Kong, Peoples R China
关键词
Uplink; Convergence; Privacy; Training; Optimization; Servers; Organizations; Differential privacy (DP); distributed optimization; empirical risk minimization (ERM); sparse optimization; RISK;
D O I
10.1109/TNNLS.2020.3020955
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In sparse empirical risk minimization (ERM) models, when sensitive personal data are used, e.g., genetic, healthcare, and financial data, it is crucial to preserve the differential privacy (DP) in training. In many applications, the information (i.e., features) of an individual is held by different organizations, which give rise to the prevalent yet challenging setting of the featurewise distributed multiparty model training. Such a setting is also beneficial to the scalability when the number of features exceeds the computation and storage capacity of a single node. However, existing private sparse optimizations are limited to centralized and samplewise distributed datasets only. In this article, we develop a differentially private algorithm for the sparse ERM model training under the featurewise distributed datasets setting. Our algorithm comes with guaranteed DP, nearly optimal utility, and reduced uplink communication complexity. Accordingly, we present a more generalized convergence analysis for block-coordinate Frank-Wolfe (BCFW) under arbitrary sampling (denoted as BCFW-AS in short), which significantly extends the known convergence results that apply to two specific sampling distributions only. To further reduce the uplink communication cost, we design an active private feature sharing scheme, which is new in both design and analysis of BCFW, to guarantee the convergence of communicating Johnson-Lindenstrauss transformed features. Empirical studies justify the new convergence as well as the nearly optimal utility theoretical results.
引用
收藏
页码:4529 / 4543
页数:15
相关论文
共 50 条
  • [1] Agarwal N, 2018, ADV NEUR IN, V31
  • [2] [Anonymous], 2013, P 30 INT C MACH LEAR
  • [3] Bassily R, 2019, ADV NEUR IN, V32
  • [4] Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds
    Bassily, Raef
    Smith, Adam
    Thakurta, Abhradeep
    [J]. 2014 55TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2014), 2014, : 464 - 473
  • [5] Basu D, 2019, ADV NEUR IN, V32
  • [6] Bellet Aurelien, 2015, SIAM INT C DAT MIN, P478
  • [7] Bernstein J, 2018, PR MACH LEARN RES, V80
  • [8] Chaudhuri K, 2011, J MACH LEARN RES, V12, P1069
  • [9] The Algorithmic Foundations of Differential Privacy
    Dwork, Cynthia
    Roth, Aaron
    [J]. FOUNDATIONS AND TRENDS IN THEORETICAL COMPUTER SCIENCE, 2013, 9 (3-4): : 211 - 406
  • [10] Gandikota V., 2019, ARXIV191107971