Secure Aggregation is Insecure: Category Inference Attack on Federated Learning

被引:31
作者
Gao, Jiqiang [1 ]
Hou, Boyu [1 ]
Guo, Xiaojie [1 ]
Liu, Zheli [1 ]
Zhang, Ying [1 ]
Chen, Kai [2 ]
Li, Jin [3 ,4 ]
机构
[1] Nankai Univ, Coll Cyber Sci, Coll Comp Sci, Tianjin Key Lab Network & Data Secur Technol, Tianjin 300071, Peoples R China
[2] Chinese Acad Sci, Inst Informat Engn, Beijing 100864, Peoples R China
[3] Guangzhou Univ, Sch Comp Sci, Guangzhou 510006, Peoples R China
[4] Nankai Univ, Coll Cyber Sci, Tianjin 300071, Peoples R China
基金
中国国家自然科学基金;
关键词
Collaborative work; Training; Privacy; Servers; Data models; Data privacy; Protocols; Federated learning; inference attack; secure aggregation; machine learning; PRIVACY;
D O I
10.1109/TDSC.2021.3128679
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning allows a large number of resource-constrained clients to train a globally-shared model together without sharing local data. These clients usually have only a few classes (categories) of data for training, where the data distribution is non-iid (not independent identically distributed). In this article, we put forward the concept of category privacy for the first time to indicate which classes of data a client has, which is an important but ignored privacy goal in the federated learning with non-iid data. Although secure aggregation protocols are designed for federated learning to protect the input privacy of clients, we perform the first systematic study on category inference attack and demonstrate that these protocols cannot fully protect category privacy. We design a differential selection strategy and two de-noising approaches to achieve the attack goal successfully. In our evaluation, we apply the attack to non-iid federated learning settings with various datasets. On MNIST, CIFAR-10, AG_news, and DBPedia dataset, our attack achieves > 90% accuracy measured in F1-score in most cases. We further consider a possible detection method and propose two strategies to make the attack more inconspicuous.
引用
收藏
页码:147 / 160
页数:14
相关论文
共 40 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Bagdasaryan E, 2020, PR MACH LEARN RES, V108, P2938
[3]   Practical Secure Aggregation for Privacy-Preserving Machine Learning [J].
Bonawitz, Keith ;
Ivanov, Vladimir ;
Kreuter, Ben ;
Marcedone, Antonio ;
McMahan, H. Brendan ;
Patel, Sarvar ;
Ramage, Daniel ;
Segal, Aaron ;
Seth, Karn .
CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, :1175-1191
[4]   Collective Data-Sanitization for Preventing Sensitive Information Inference Attacks in Social Networks [J].
Cai, Zhipeng ;
He, Zaobo ;
Guan, Xin ;
Li, Yingshu .
IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2018, 15 (04) :577-590
[5]  
Chen XY, 2017, Arxiv, DOI arXiv:1712.05526
[6]  
Dean Jeffrey., 2012, Advances in Neural Information Processing Systems, P25
[7]   The Algorithmic Foundations of Differential Privacy [J].
Dwork, Cynthia ;
Roth, Aaron .
FOUNDATIONS AND TRENDS IN THEORETICAL COMPUTER SCIENCE, 2013, 9 (3-4) :211-406
[8]  
Fang MH, 2020, PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, P1623
[9]   Property Inference Attacks on Fully Connected Neural Networks using Permutation Invariant Representations [J].
Ganju, Karan ;
Wang, Qi ;
Yang, Wei ;
Gunter, Carl A. ;
Borisov, Nikita .
PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, :619-633
[10]  
Ghamrawi Nadia., 2005, CIKM 05 P 14 ACM INT, P195, DOI [10.1145/1099554.109959, 10.1145/1099554.109959andlt