Semi-Fragile Neural Network Watermarking Based on Adversarial Examples

被引:2
作者
Yuan, Zihan [1 ]
Zhang, Xinpeng [1 ]
Wang, Zichi [1 ]
Yin, Zhaoxia [2 ]
机构
[1] Shanghai Univ, Sch Commun & Informat Engn, Shanghai 200444, Peoples R China
[2] East China Normal Univ, Sch Commun & Elect Engn, Shanghai 200444, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年 / 8卷 / 04期
基金
中国国家自然科学基金;
关键词
Semi-fragile watermarking; neural network; black-box; malicious tampering; privacy and security; MODEL;
D O I
10.1109/TETCI.2024.3372373
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) may be subject to various modifications during transmission and use. Regular processing operations do not affect the functionality of a model, while malicious tampering will cause serious damage. Therefore, it is crucial to determine the availability of a DNN model. To address this issue, we propose a semi-fragile black-box watermarking method that can distinguish between accidental modification and malicious tampering of DNNs, focusing on the privacy and security of neural network models. Specifically, for a given model, a strategy is designed to generate semi-fragile and sensitive samples using adversarial example techniques without decreasing the model accuracy. The model outputs for these samples are extremely sensitive to malicious tampering and robust to accidental modification. According to these properties, accidental modification and malicious tampering can be distinguished to assess the availability of a watermarked model. Extensive experiments demonstrate that the proposed method can detect malicious model tampering with high accuracy up to 100% while tolerating accidental modifications such as fine-tuning, pruning, and quantitation with the accuracy exceed 75%. Moreover, our semi-fragile neural network watermarking approach can be easily extended to various DNNs.
引用
收藏
页码:2775 / 2790
页数:16
相关论文
共 36 条
[1]   DeepiSign: Invisible Fragile Watermark to Protect the Integrity and Authenticity of CNN [J].
Abuadbba, Alsharif ;
Kim, Hyoungshick ;
Nepal, Surya .
36TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2021, 2021, :952-959
[2]   NeuNAC: A novel fragile watermarking algorithm for integrity protection of neural networks [J].
Botta, Marco ;
Cavagnino, Davide ;
Esposito, Roberto .
INFORMATION SCIENCES, 2021, 576 :228-241
[3]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[4]  
Chen B., 2019, IN PRESS, P707, DOI [10.48550/arXiv.1904.00344, DOI 10.48550/ARXIV.1904.00344]
[5]  
Chen Chien-Hua, ARXIV
[6]  
Cox M., 2007, DIGITAL WATER MARKIN
[7]   A Generalized Deep Neural Network Approach for Digital Watermarking Analysis [J].
Ding, Weiping ;
Ming, Yurui ;
Cao, Zehong ;
Lin, Chin-Teng .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (03) :613-627
[8]   Efficient Privacy Preserving Edge Intelligent Computing Framework for Image Classification in IoT [J].
Fagbohungbe, Omobayode ;
Reza, Sheikh Rufsan ;
Dong, Xishuang ;
Qian, Lijun .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (04) :941-956
[9]   A watermarking scheme based on rotating vector for image content authentication [J].
Fu, Jianjing ;
Mao, Jiafa ;
Xue, Dawen ;
Chen, Deren .
SOFT COMPUTING, 2020, 24 (08) :5755-5772
[10]   Understanding action recognition in still images [J].
Girish, Deeptha ;
Singh, Vineeta ;
Ralescu, Anca .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :1523-1529