Privacy-preserving federated learning based on partial low-quality data

被引:4
作者
Wang, Huiyong [1 ,2 ]
Wang, Qi [1 ]
Ding, Yong [2 ,3 ]
Tang, Shijie [4 ]
Wang, Yujue [5 ]
机构
[1] Guilin Univ Elect Technol, Sch Math & Comp Sci, Guilin 541004, Peoples R China
[2] Guilin Univ Elect Technol, Sch Comp Sci & Informat Secur, Guangxi Key Lab Cryptog & Informat Secur, Guilin 541004, Peoples R China
[3] Cyberspace Secur Res Ctr, Peng Cheng Lab, Shenzhen 518055, Peoples R China
[4] Guilin Univ Elect Technol, Sch Elect Engn & Automat, Guilin 541004, Peoples R China
[5] Beihang Univ, Hangzhou Innovat Inst, Hangzhou 310052, Peoples R China
来源
JOURNAL OF CLOUD COMPUTING-ADVANCES SYSTEMS AND APPLICATIONS | 2024年 / 13卷 / 01期
关键词
Federated learning; Privacy protection; Low-quality data; Distributed homomorphic encryption; SECURITY;
D O I
10.1186/s13677-024-00618-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Traditional machine learning requires collecting data from participants for training, which may lead to malicious acquisition of privacy in participants' data. Federated learning provides a method to protect participants' data privacy by transferring the training process from a centralized server to terminal devices. However, the server may still obtain participants' privacy through inference attacks and other methods. In addition, the data provided by participants varies in quality, and the excessive involvement of low-quality data in the training process can render the model unusable, which is an important issue in current mainstream federated learning. To address the aforementioned issues, this paper proposes a Privacy Preserving Federated Learning Scheme with Partial Low-Quality Data (PPFL-LQDP). It can achieve good training results while allowing participants to utilize partial low-quality data, thereby enhancing the privacy and security of the federated learning scheme. Specifically, we use a distributed Paillier cryptographic mechanism to protect the privacy and security of participants' data during the Federated training process. Additionally, we construct composite evaluation values for the data held by participants to reduce the involvement of low-quality data, thereby minimizing the negative impact of such data on the model. Through experiments on the MNIST dataset, we demonstrate that this scheme can complete the model training of federated learning with the participation of partial low-quality data, while effectively protecting the security and privacy of participants' data. Comparisons with related schemes also show that our scheme has good overall performance.
引用
收藏
页数:16
相关论文
共 32 条
  • [1] Security and Privacy-Enhancing Multicloud Architectures
    Bohli, Jens-Matthias
    Gruschka, Nils
    Jensen, Meiko
    Lo Iacono, Luigi
    Marnau, Ninja
    [J]. IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2013, 10 (04) : 212 - 224
  • [2] Practical Secure Aggregation for Privacy-Preserving Machine Learning
    Bonawitz, Keith
    Ivanov, Vladimir
    Kreuter, Ben
    Marcedone, Antonio
    McMahan, H. Brendan
    Patel, Sarvar
    Ramage, Daniel
    Segal, Aaron
    Seth, Karn
    [J]. CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, : 1175 - 1191
  • [3] Federated-Cloud Based Deep Neural Networks with Privacy Preserving Image Filtering Techniques
    Choi, Isabelle
    Song, Qiyang
    Sun, Kun
    [J]. 2019 IEEE CONFERENCE ON DEPENDABLE AND SECURE COMPUTING (DSC), 2019, : 227 - 234
  • [4] Damgård I, 2001, LECT NOTES COMPUT SC, V1992, P119
  • [5] Danner Gabor, 2015, Distributed Applications and Interoperable Systems. 15th IFIP WG 6.1 International Conference, DAIS 2015, held as part of the 10th International Federated Conference on Distributed Computing Techniques, DisCoTec 2015. Proceedings: LNCS 9038, P30, DOI 10.1007/978-3-319-19129-4_3
  • [6] A Critical Overview of Privacy in Machine Learning
    De Cristofaro, Emiliano
    [J]. IEEE SECURITY & PRIVACY, 2021, 19 (04) : 19 - 27
  • [7] Diaz C, 2022, IEEE 4 INT C BIOINSP, P1, DOI [10.1109/BIP56202.2022.10032475, DOI 10.1109/BIP56202.2022.10032475]
  • [8] Privacy-Preserving Deep Learning With Homomorphic Encryption: An Introduction
    Falcetta, Alessandro
    Roveri, Manuel
    [J]. IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2022, 17 (03) : 14 - 25
  • [9] Privacy Preserving Machine Learning with Homomorphic Encryption and Federated Learning
    Fang, Haokun
    Qian, Quan
    [J]. FUTURE INTERNET, 2021, 13 (04):
  • [10] Fleming N, 2018, NATURE, V557, pS55, DOI 10.1038/d41586-018-05267-x