A Federated Learning Framework Based on Differentially Private Continuous Data Release

被引:5
作者
Cai, Jianping [1 ]
Liu, Ximeng [1 ]
Ye, Qingqing [2 ]
Liu, Yang [3 ]
Wang, Yuyang [1 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou 350108, Peoples R China
[2] Hong Kong Polytech Univ, Dept Elect & Informat Engn, Hong Kong 999077, Peoples R China
[3] Tsinghua Univ, Inst AI Ind Res, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Privacy; Data models; Artificial intelligence; Differential privacy; Biomedical imaging; Computational modeling; Security; Federated learning; differential privacy; continuous data release; binary indexed tree; matrix mechanism; NOISE;
D O I
10.1109/TDSC.2024.3364060
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) provides a learning framework without participants sharing local raw data, but individual privacy is still at risk of disclosure through attacking the trained models. Due to the strong privacy guarantee, differential privacy (DP) is widely applied to FL to avoid privacy leakage. Traditional private learning adds noise directly to the gradients. The continuous accumulated noise on parameter models severely impairs learning effectiveness. To solve this problem, we introduce the idea of differentially private continuous data release (DPCR) into FL and propose an FL framework based on DPCR (FL-DPCR). Meanwhile, our proposed Equivalent Aggregation Theorem demonstrates that DPCR effectively reduces the overall error added to parameter models and improves FL's accuracy. To improve FL-DPCR's learning effectiveness, we introduce Matrix Mechanism to construct a release strategy and design a binary-indexed-tree (BIT) based DPCR model for Gaussian mechanism (BCRG). By solving a complex nonlinear programming problem with negative exponents, BCRG achieves optimal release accuracy efficiently. Besides, we exploit the residual privacy budget to boost the accuracy further and propose an advanced BCRG version (ABCRG). Our experiments show that, compared to traditional FL with DP, our achievements improve the accuracy with gains ranging from 3.4% on FMNIST to 65.7% on PAMAP2.
引用
收藏
页码:4879 / 4894
页数:16
相关论文
共 47 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
AlShorman O., 2020, Indonesian J. Elect. Eng. Comput. Sci., V20, P422
[3]  
Balle B, 2022, J ROY STAT SOC B, V84, P37, DOI 10.1111/rssb.12455
[4]  
Boyd S., 2004, Convex Optimization
[5]  
[蔡剑平 Cai Jianping], 2016, [计算机科学与探索, Journal of Frontiers of Computer Science & Technology], V10, P481
[6]   Private and Continual Release of Statistics [J].
Chan, T. -H. Hubert ;
Shi, Elaine ;
Song, Dawn .
ACM TRANSACTIONS ON INFORMATION AND SYSTEM SECURITY, 2011, 14 (03)
[7]  
Chaudhuri K, 2011, J MACH LEARN RES, V12, P1069
[8]   Health Insurance Portability and Accountability Act (HIPAA) Compliant Access Control Model for Web Services [J].
Cheng, Vivying S. Y. ;
Hung, Patrick C. K. .
INTERNATIONAL JOURNAL OF HEALTHCARE INFORMATION SYSTEMS AND INFORMATICS, 2006, 1 (01) :22-39
[9]  
Choudhury O, 2020, Arxiv, DOI arXiv:1910.02578
[10]  
Dwork C, 2006, LECT NOTES COMPUT SC, V4004, P486