A privacy-preserving and non-interactive federated learning scheme for regression training with gradient descent

被引:57
作者
Wang, Fengwei [1 ,2 ]
Zhu, Hui [1 ,3 ]
Lu, Rongxing [2 ]
Zheng, Yandong [2 ]
Li, Hui [1 ]
机构
[1] Xidian Univ, Natl Key Lab Integrated Networks Serv, Xian, Peoples R China
[2] Univ New Brunswick, Fac Comp Sci, Fredericton, NB, Canada
[3] Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Regression training; Privacy-preserving; Secure data aggregation; Gradient descent; LINEAR-REGRESSION; LOGISTIC-REGRESSION; RIDGE-REGRESSION; DISTRIBUTED DATA; EFFICIENT;
D O I
10.1016/j.ins.2020.12.007
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, the extensive application of machine learning technologies has been witnessed in various fields. However, in many applications, massive data are distributively stored in multiple data owners. Meanwhile, due to the privacy concerns and communication constraints, it is difficult to bridge the data silos among data owners for training a global machine learning model. In this paper, we propose a privacy-preserving and noninteractive federated learning scheme for regression training with gradient descent, named VANE. With VANE, multiple data owners are able to train a global linear, ridge or logistic regression model with the assistance of cloud, while their private local training data can be well protected. Specifically, we first design a secure data aggregation algorithm, with which local training data from multiple data owners can be aggregated and trained to a global model without disclosing any private information. Meanwhile, benefit from our data pre-processing method, the whole training process is non-interactive, i.e., there is no interaction between data owners and the cloud. Detailed security analysis shows that VANE can well protect the local training data of data owners. The performance evaluation results demonstrate that the training performance of our VANE is around 10(3) times faster than existing schemes. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页码:183 / 200
页数:18
相关论文
共 37 条
[1]   Robust Federated Learning With Noisy Communication [J].
Ang, Fan ;
Chen, Li ;
Zhao, Nan ;
Chen, Yunfei ;
Wang, Weidong ;
Yu, F. Richard .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (06) :3452-3464
[2]  
[Anonymous], TECHNOMETRICS
[3]  
[Anonymous], **DATA OBJECT**
[4]  
[Anonymous], **DATA OBJECT**
[5]   Privacy-Preserving Logistic Regression with Distributed Data Sources via Homomorphic Encryption [J].
Aono, Yoshinori ;
Hayashi, Takuya ;
Phong, Le Trieu ;
Wang, Lihua .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (08) :2079-2089
[6]  
Asuncion A., 2007, UCI Machine Learning Repository
[7]  
Chen XH, 2019, AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, P1880
[8]   Privacy-preserving ridge regression on distributed data [J].
Chen, Yi-Ruei ;
Rezapour, Amir ;
Tzeng, Wen-Guey .
INFORMATION SCIENCES, 2018, 451 :34-49
[9]  
Cock MD, 2015, Proceedings of the 8th ACM Workshop on Arti cial Intelligence and Security, P3
[10]   A survey on application of machine learning for Internet of Things [J].
Cui, Laizhong ;
Yang, Shu ;
Chen, Fei ;
Ming, Zhong ;
Lu, Nan ;
Qin, Jing .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2018, 9 (08) :1399-1417