GDPR and Large Language Models: Technical and Legal Obstacles

被引:2
作者
Feretzakis, Georgios [1 ]
Vagena, Evangelia [2 ]
Kalodanis, Konstantinos [3 ]
Peristera, Paraskevi [4 ]
Kalles, Dimitris [1 ]
Anastasiou, Athanasios [5 ]
机构
[1] Hellen Open Univ, Sch Sci & Technol, Patras 26335, Greece
[2] Athens Univ Econ & Business, Athens 10434, Greece
[3] Harokopio Univ Athens, Dept Informat & Telemat, Kallithea 17676, Greece
[4] Stockholm Univ, Dept Psychol, Div Psychobiol & Epidemiol, S-10691 Stockholm, Sweden
[5] Natl Tech Univ Athens, Biomed Engn Lab, Athens 15780, Greece
关键词
GDPR; artificial intelligence; large language models; AI Act; LLM; LLMs; data privacy; AI; Legal Obstacles;
D O I
10.3390/fi17040151
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large Language Models (LLMs) have revolutionized natural language processing but present significant technical and legal challenges when confronted with the General Data Protection Regulation (GDPR). This paper examines the complexities involved in reconciling the design and operation of LLMs with GDPR requirements. In particular, we analyze how key GDPR provisions-including the Right to Erasure, Right of Access, Right to Rectification, and restrictions on Automated Decision-Making-are challenged by the opaque and distributed nature of LLMs. We discuss issues such as the transformation of personal data into non-interpretable model parameters, difficulties in ensuring transparency and accountability, and the risks of bias and data over-collection. Moreover, the paper explores potential technical solutions such as machine unlearning, explainable AI (XAI), differential privacy, and federated learning, alongside strategies for embedding privacy-by-design principles and automated compliance tools into LLM development. The analysis is further enriched by considering the implications of emerging regulations like the EU's Artificial Intelligence Act. In addition, we propose a four-layer governance framework that addresses data governance, technical privacy enhancements, continuous compliance monitoring, and explainability and oversight, thereby offering a practical roadmap for GDPR alignment in LLM systems. Through this comprehensive examination, we aim to bridge the gap between the technical capabilities of LLMs and the stringent data protection standards mandated by GDPR, ultimately contributing to more responsible and ethical AI practices.
引用
收藏
页数:26
相关论文
共 90 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
2023, Arxiv, DOI arXiv:2303.08774
[3]  
Adamopoulou E., 2020, AIAI 2020, V584, P373, DOI [DOI 10.1007/978-3-030-49186-4_31, 10.1007/978-3-030-49186-431, 10.1007978-3-030-49186-431]
[4]  
Albrecht J. P., 2016, European Data Protection Law Review, V2, P287, DOI [DOI 10.21552/EDPL/2016/3/4, 10.21552/EDPL/2016/3/4]
[5]  
Anil Rohan, 2022, FINDINGS ASS COMPUTA, P6481, DOI DOI 10.18653/V1/2022.FINDINGS-EMNLP
[6]  
[Anonymous], 2016, Regulation (EU) 2016/679 of the European Parliament and of the Council
[7]  
[Anonymous], 2016, Regulation (EU) 2016/679 of the European Parliament and of the Council
[8]  
[Anonymous], 2016, Regulation (EU) 2016/679 of the European Parliament and of the Council
[9]  
[Anonymous], 2016, Regulation (EU) 2016/679 of the European Parliament and of the Council
[10]  
[Anonymous], 2016, Regulation (EU) 2016/679 of the European Parliament and of the Council