DeepiSign: Invisible Fragile Watermark to Protect the Integrity and Authenticity of CNN

被引:9
作者
Abuadbba, Alsharif [1 ,2 ]
Kim, Hyoungshick [1 ,3 ]
Nepal, Surya [2 ]
机构
[1] CSIRO, Data61, Canberra, ACT, Australia
[2] Cyber Secur CRC, Joondalup, Australia
[3] Sungkyunkwan Univ, Seoul, South Korea
来源
36TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2021 | 2021年
基金
新加坡国家研究基金会;
关键词
Watermarking; CNN; integrity; authenticity;
D O I
10.1145/3412841.3441970
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Networks (CNNs) deployed in real-life applications such as autonomous vehicles have shown to be vulnerable to manipulation attacks, such as poisoning attacks and fine-tuning. Hence, it is essential to ensure the integrity and authenticity of CNNs because compromised models can produce incorrect outputs and behave maliciously. In this paper, we propose a self-contained tamper-proofing method, called DeepiSign, to ensure the integrity and authenticity of CNN models against such manipulation attacks. DeepiSign applies the idea of fragile invisible watermarking to securely embed a secret and its hash value into a CNN model. To verify the integrity and authenticity of the model, we retrieve the secret from the model, compute the hash value of the secret, and compare it with the embedded hash value. To minimize the effects of the embedded secret on the CNN model, we use a wavelet-based technique to transform weights into the frequency domain and embed the secret into less significant coefficients. Our theoretical analysis shows that DeepiSign can hide up to 1KB secret in each layer with minimal loss of the model's accuracy. To evaluate the security and performance of DeepiSign, we performed experiments on four pre-trained models (ResNet18, VGG16, AlexNet, and MobileNet) using three datasets (MNIST, CIFAR-10, and Imagenet) against three types of manipulation attacks (targeted input poisoning, output poisoning, and fine-tuning). The results demonstrate that DeepiSign is verifiable without degrading the classification accuracy, and robust against representative CNN manipulation attacks.
引用
收藏
页码:952 / 959
页数:8
相关论文
共 26 条
[1]  
Adi Y, 2018, PROCEEDINGS OF THE 27TH USENIX SECURITY SYMPOSIUM, P1615
[2]   Digital image steganography: Survey and analysis of current methods [J].
Cheddad, Abbas ;
Condell, Joan ;
Curran, Kevin ;
Mc Kevitt, Paul .
SIGNAL PROCESSING, 2010, 90 (03) :727-752
[3]   DeepDriving: Learning Affordance for Direct Perception in Autonomous Driving [J].
Chen, Chenyi ;
Seff, Ari ;
Kornhauser, Alain ;
Xiao, Jianxiong .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2722-2730
[4]  
Chen XY, 2017, Arxiv, DOI arXiv:1712.05526
[5]  
Rouhani BD, 2018, Arxiv, DOI arXiv:1804.00750
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]  
Desai Madhavi B., 2014, International Journal of Computer Science and Information Technologies, V5, P4752
[8]  
Gu TY, 2019, Arxiv, DOI arXiv:1708.06733
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
He Z., 2018, arXiv