When AI Meets Information Privacy: The Adversarial Role of AI in Data Sharing Scenario

被引:11
作者
Majeed, Abdul [1 ]
Hwang, Seong Oun [1 ]
机构
[1] Gachon Univ, Dept Comp Engn, Seongnam 13120, South Korea
关键词
AI-powered attacks; artificial intelligence; background knowledge; compromising privacy; data publishing; personal data; privacy; safeguarding privacy; synthetic data; utility; UTILITY;
D O I
10.1109/ACCESS.2023.3297646
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Artificial intelligence (AI) is a transformative technology with a substantial number of practical applications in commercial sectors such as healthcare, finance, aviation, and smart cities. AI also has strong synergy with the information privacy (IP) domain from two distinct aspects: as a protection tool (i.e., safeguarding privacy), and as a threat tool (i.e., compromising privacy). In the former case, AI techniques are amalgamated with the traditional anonymization techniques to improve various key components of the anonymity process, and therefore, privacy is safeguarded effectively. In the latter case, some adversarial knowledge is aggregated with the help of AI techniques and subsequently used to compromise the privacy of individuals. To the best of our knowledge, threats posed by AI-generated knowledge such as synthetic data (SD) to information privacy are often underestimated, and most of the existing anonymization methods do not consider/model this SD-based knowledge that can be available to the adversary, leading to privacy breaches in some cases. In this paper, we highlight the role of AI as a threat tool (i.e., AI used to compromise an individual's privacy), with a special focus on SD that can serve as background knowledge leading to various kinds of privacy breaches. For instance, SD can encompass pertinent information (e.g., total # of attributes in data, distributions of sensitive information, category values of each attribute, minor and major values of some attributes, etc.) about real data that can offer a helpful hint to the adversary regarding the composition of anonymized data, that can subsequently lead to uncovering the identity or private information. We perform reasonable experiments on a real-life benchmark dataset to prove the pitfalls of AI in the data publishing scenario (when a database is either fully or partially released to public domains for conducting analytics).
引用
收藏
页码:76177 / 76195
页数:19
相关论文
共 66 条
[51]  
Soliman A., 2023, SCI REP-UK, V13, P1
[52]   Design of privacy preserving model based on clustering involved anonymization along with feature selection [J].
Srijayanthi, S. ;
Sethukarasi, T. .
COMPUTERS & SECURITY, 2023, 126
[53]   Data Privacy and Trustworthy Machine Learning [J].
Strobel, Martin ;
Shokri, Reza .
IEEE SECURITY & PRIVACY, 2022, 20 (05) :44-49
[54]   K-Anonymity Privacy Protection Algorithm for Multi-Dimensional Data against Skewness and Similarity Attacks [J].
Su, Bing ;
Huang, Jiaxuan ;
Miao, Kelei ;
Wang, Zhangquan ;
Zhang, Xudong ;
Chen, Yourong .
SENSORS, 2023, 23 (03)
[55]   k-anonymity:: A model for protecting privacy [J].
Sweeney, L .
INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2002, 10 (05) :557-570
[56]  
Tang P., 2023, IEEE T KNOWL DATA EN, DOI [10.1109/TKDE.2023.324166, DOI 10.1109/TKDE.2023.3241661]
[57]   Data Anonymization: An Experimental Evaluation Using Open-Source Tools [J].
Tomas, Joana ;
Rasteiro, Deolinda ;
Bernardino, Jorge .
FUTURE INTERNET, 2022, 14 (06)
[58]  
Wang Yinqiu, 2023, Algorithms and Architectures for Parallel Processing: 22nd International Conference, ICA3PP 2022, Proceedings. Lecture Notes in Computer Science (13777), P21, DOI 10.1007/978-3-031-22677-9_2
[59]  
Xu L, 2019, ADV NEUR IN, V32
[60]   Attack Risk Analysis in Data Anonymization in Internet of Things [J].
Yang, Tianli ;
Cang, Li Shan ;
Iqbal, Muddesar ;
Almakhles, Dhafer .
IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (04) :4986-4993