Privacy-preserving non-negative matrix factorization for decentralized-data using correlated noise

被引:0
作者
Imtiaz, Hafiz [1 ]
Karmakar, Tusher [1 ]
Mohanta, Protoye Kumar [1 ]
机构
[1] Bangladesh Univ Engn & Technol BUET, Elect & Elect Engn, Dhaka 1205, Bangladesh
关键词
Non-negative matrix factorization (NMF); Differential privacy; Decentralized data; Correlation assisted private estimation (CAPE); Renyi differential privacy; DISTRIBUTED CONVEX-OPTIMIZATION; ALGORITHMS; MECHANISM;
D O I
10.1007/s11760-025-03887-1
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Several matrix factorization algorithms are employed in machine learning applications. Among these, Non-negative Matrix Factorization (NMF) gained attention due to the ability to extract meaningful features from inherently non-negative data, such as documents, images or videos. However, such data are often privacy-sensitive, which necessitates formal privacy guarantees of the machine learning model training algorithm. Additionally, modern data are typically stored in different nodes or clients, rather than a centralized server. Conventional decentralized privacy-preserving schemes suffer from too much noise and consequently, much lower utility compared to their centralized counterparts. This motivates us to develop an efficient privacy-preserving NMF algorithm that can operate on decentralized data, and can closely approximate the performance of centralized/non-privacy-preserving approach, while offering strict privacy guarantees. We design our method and demonstrate our results in such a way that the clients/data holders have the control to select the degree of privacy guarantee based on the required utility. We show the effectiveness of our proposed algorithm on six real datasets. Our experimental results show that our proposed method easily outperforms conventional privacy-preserving scheme, while achieving close approximation of the non-privacy-preserving approach under some parameter choices.
引用
收藏
页数:16
相关论文
共 61 条
  • [1] Deep Learning with Differential Privacy
    Abadi, Martin
    Chu, Andy
    Goodfellow, Ian
    McMahan, H. Brendan
    Mironov, Ilya
    Talwar, Kunal
    Zhang, Li
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 308 - 318
  • [2] Andrew G., 2021, Adv Neural Inf Process Syst, V34, P17455
  • [3] [Anonymous], 1998, Online learning in neural networks
  • [4] Apple D, 2017, Apple Machine Learning Journal, V1
  • [5] Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds
    Bassily, Raef
    Smith, Adam
    Thakurta, Abhradeep
    [J]. 2014 55TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2014), 2014, : 464 - 473
  • [6] Practical Secure Aggregation for Privacy-Preserving Machine Learning
    Bonawitz, Keith
    Ivanov, Vladimir
    Kreuter, Ben
    Marcedone, Antonio
    McMahan, H. Brendan
    Patel, Sarvar
    Ramage, Daniel
    Segal, Aaron
    Seth, Karn
    [J]. CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, : 1175 - 1191
  • [7] SVD based initialization: A head start for nonnegative matrix factorization
    Boutsidis, C.
    Gallopoulos, E.
    [J]. PATTERN RECOGNITION, 2008, 41 (04) : 1350 - 1362
  • [8] Distributed optimization and statistical learning via the alternating direction method of multipliers
    Boyd S.
    Parikh N.
    Chu E.
    Peleato B.
    Eckstein J.
    [J]. Foundations and Trends in Machine Learning, 2010, 3 (01): : 1 - 122
  • [9] Bureau U.C., 2018, Census Blogs
  • [10] Chaudhuri K, 2013, J MACH LEARN RES, V14, P2905