Matryoshka: Exploiting the Over-Parametrization of Deep Learning Models for Covert Data Transmission

被引:0
|
作者
Pan, Xudong [1 ]
Zhang, Mi [1 ]
Yan, Yifan [1 ]
Zhang, Shengyao [1 ]
Yang, Min [1 ,2 ,3 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200437, Peoples R China
[2] Minist Educ, Fac Shanghai Inst Intelligent Elect & Syst, Shanghai 200437, Peoples R China
[3] Minist Educ, Engn Res Ctr Cyber Secur Auditing & Monitoring, Shanghai 200437, Peoples R China
基金
中国国家自然科学基金;
关键词
Data models; Training; Predictive models; Training data; Computational modeling; Task analysis; Data privacy; Training data privacy; deep learning privacy; steganography; covert transmission; AI security;
D O I
10.1109/TPAMI.2024.3434417
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
High-quality private machine learning (ML) data stored in local data centers becomes a key competitive factor for AI corporations. In this paper, we present a novel insider attack called Matryoshka to reveal the possibility of breaking the privacy of ML data even with no exposed interface. Our attack employs a scheduled-to-publish DNN model as a carrier model for covert transmission of secret models which memorize the information of private ML data that otherwise has no interface to the outsider. At the core of our attack, we present a novel parameter sharing approach which exploits the learning capacity of the carrier model for information hiding. Our approach simultaneously achieves: (i) High Capacity - With almost no utility loss of the carrier model, Matryoshka can transmit over 10,000 real-world data samples within a carrier model which has $220\times$220x less parameters than the total size of the stolen data, and simultaneously transmit multiple heterogeneous datasets or models within a single carrier model under a trivial distortion rate, neither of which can be done with existing steganography techniques; (ii) Decoding Efficiency - once downloading the published carrier model, an outside colluder can exclusively decode the hidden models from the carrier model with only several integer secrets and the knowledge of the hidden model architecture; (iii) Effectiveness - Moreover, almost all the recovered models either have similar performance as if it is trained independently on the private data, or can be further used to extract memorized raw training data with low error; (iv) Robustness - Information redundancy is naturally implemented to achieve resilience against common post-processing techniques on the carrier before its publishing; (v) Covertness - A model inspector with different levels of prior knowledge could hardly differentiate a carrier model from a normal model.
引用
收藏
页码:663 / 678
页数:16
相关论文
共 20 条
  • [1] Leveraging Haptic Feedback to Improve Data Quality and Quantity for Deep Imitation Learning Models
    Cuan, Catie
    Okamura, Allison
    Khansari, Mohi
    IEEE TRANSACTIONS ON HAPTICS, 2024, 17 (04) : 984 - 991
  • [2] A Survey on Mathematical, Machine Learning and Deep Learning Models for COVID-19 Transmission and Diagnosis
    John, Christopher Clement
    Ponnusamy, VijayaKumar
    Krishnan Chandrasekaran, Sriharipriya
    Nandakumar, R.
    IEEE REVIEWS IN BIOMEDICAL ENGINEERING, 2022, 15 : 325 - 340
  • [3] Progress Estimation for End-to-End Training of Deep Learning Models With Online Data Preprocessing
    Dong, Qifei
    Luo, Gang
    IEEE ACCESS, 2024, 12 : 18658 - 18684
  • [4] The Cost of Training Machine Learning Models Over Distributed Data Sources
    Guerra, Elia
    Wilhelmi, Francesc
    Miozzo, Marco
    Dini, Paolo
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2023, 4 : 1111 - 1126
  • [5] DeepTrigger: A Watermarking Scheme of Deep Learning Models Based on Chaotic Automatic Data Annotation
    Zhang, Ying-Qian
    Jia, Yi-Ran
    Wang, Xingyuan
    Niu, Qiong
    Chen, Nian-Dong
    IEEE ACCESS, 2020, 8 : 213296 - 213305
  • [6] CT Lung Nodule Segmentation: A Comparative Study of Data Preprocessing and Deep Learning Models
    Chen, Weihao
    Wang, Yu
    Tian, Dingcheng
    Yao, Yudong
    IEEE ACCESS, 2023, 11 : 34925 - 34931
  • [7] Attention-Based Multimodal Deep Learning on Vision-Language Data: Models, Datasets, Tasks, Evaluation Metrics and Applications
    Bose, Priyankar
    Rana, Pratip
    Ghosh, Preetam
    IEEE ACCESS, 2023, 11 : 80624 - 80646
  • [8] Adversarially-Regularized Mixed Effects Deep Learning (ARMED) Models Improve Interpretability, Performance, and Generalization on Clustered (non-iid) Data
    Nguyen, Kevin P.
    Treacher, Alex H.
    Montillo, Albert A.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (07) : 8081 - 8093
  • [9] Increasing the Robustness of Deep Learning Models for Object Segmentation: A Framework for Blending Automatically Annotated Real and Synthetic Data
    Karoly, Artur Istvan
    Tirczka, Sebestyen
    Gao, Huijun
    Rudas, Imre J.
    Galambos, Peter
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (01) : 25 - 38
  • [10] A Comparative Multivariate Analysis of VAR and Deep Learning-Based Models for Forecasting Volatile Time Series Data
    Gopali, Saroj
    Siami-Namini, Sima
    Abri, Faranak
    Namin, Akbar Siami
    IEEE ACCESS, 2024, 12 : 155423 - 155436