Unambiguous and High-Fidelity Backdoor Watermarking for Deep Neural Networks

被引:17
作者
Hua, Guang [1 ,2 ]
Teoh, Andrew Beng Jin [3 ]
Xiang, Yong [4 ]
Jiang, Hao [5 ]
机构
[1] Wuhan Univ, Sch Elect Informat, Wuhan 430072, Peoples R China
[2] ASTAR, Inst Infocomm Res I2R, Singapore 138632, Singapore
[3] Yonsei Univ, Coll Engn, Sch Elect & Elect Engn, Seoul 120749, South Korea
[4] Deakin Univ, Sch Informat Technol, Burwood, Vic 3125, Australia
[5] Wuhan Univ, Sch Elect Informat, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Backdoor watermarking; black-box ambiguity attack; high-fidelity deep neural network (DNN) watermarking; neural network watermarking; unambiguous verification;
D O I
10.1109/TNNLS.2023.3250210
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The unprecedented success of deep learning could not be achieved without the synergy of big data, computing power, and human knowledge, among which none is free. This calls for the copyright protection of deep neural networks (DNNs), which has been tackled via DNN watermarking. Due to the special structure of DNNs, backdoor watermarks have been one of the popular solutions. In this article, we first present a big picture of DNN watermarking scenarios with rigorous definitions unifying the black-and white-box concepts across watermark embedding, attack, and verification phases. Then, from the perspective of data diversity, especially adversarial and open set examples overlooked in the existing works, we rigorously reveal the vulnerability of backdoor watermarks against black-box ambiguity attacks. To solve this problem, we propose an unambiguous backdoor watermarking scheme via the design of deterministically dependent trigger samples and labels, showing that the cost of ambiguity attacks will increase from the existing linear complexity to exponential complexity. Furthermore, noting that the existing definition of backdoor fidelity is solely concerned with classification accuracy, we propose to more rigorously evaluate fidelity via examining training data feature distributions and decision boundaries before and after backdoor embedding. Incorporating the proposed prototype guided regularizer (PGR) and fine-tune all layers (FTAL) strategy, we show that backdoor fidelity can be substantially improved. Experimental results using two versions of the basic ResNet18, advanced wide residual network (WRN28_10) and EfficientNet-B0, on MNIST, CIFAR-10, CIFAR-100, and FOOD-101 classification tasks, respectively, illustrate the advantages of the proposed method.
引用
收藏
页码:11204 / 11217
页数:14
相关论文
共 48 条
[1]  
Adi Y, 2018, PROCEEDINGS OF THE 27TH USENIX SECURITY SYMPOSIUM, P1615
[2]   A New Measure of Watermarking Security: The Effective Key Length [J].
Bas, Patrick ;
Furon, Teddy .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2013, 8 (08) :1306-1317
[3]  
Bassett R, 2020, ARXIV
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]  
Cecil R. R., 2019, Pharmaceutical Supply ChainsMedicines Shortages, P183
[6]   SpecMark: A Spectral Watermarking Framework for IP Protection of Speech Recognition Systems [J].
Chen, Huili ;
Darvish, Bita ;
Koushanfar, Farinaz .
INTERSPEECH 2020, 2020, :2312-2316
[7]   DeepMarks: A Secure Fingerprinting Framework for Digital Rights Management of Deep Learning Models [J].
Chen, Huili ;
Rouhani, Bita Darvish ;
Fu, Cheng ;
Zhao, Jishen ;
Koushanfar, Farinaz .
ICMR'19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2019, :105-113
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Duda R.O., 2001, PATTERN CLASSIFICATI
[10]  
Fan LX, 2019, ADV NEUR IN, V32