Multi-bit, Black-box Watermarking of Deep Neural Networks in Embedded Applications

被引:2
作者
Leroux, Sam [1 ]
Vanassche, Stijn [1 ]
Simoens, Pieter [1 ]
机构
[1] Ghent Univ Imec, Dept Informat Technol, IDLab, Ghent, Belgium
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW | 2024年
关键词
D O I
10.1109/CVPRW63382.2024.00217
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The effort required to collect data and train a large neural network requires a significant investment from organizations. Therefore, trained neural networks are often seen as valuable intellectual property that needs to be protected. At the same time, we are increasingly seeing applications where a model is deployed on an edge device. This has several benefits, including improved privacy and reduced latency but it also opens up the possibility for third parties to extract the trained model from the device and to use it for their own purposes. Watermarking techniques aim to safeguard neural networks from unauthorized usage. These methods alter the model's behavior for specific trigger inputs, enabling the owner to recognize stolen instances. However, existing watermarking algorithms are not suited for distributed edge AI scenarios as they only create a single watermarked model instance. We introduce a novel multi-bit watermarking approach capable of efficiently generating a large number of model instances. Each of these instances maintains functional equivalence but exhibits unique behaviors when prompted with a predefined trigger input. This allows the owner to trace the source of a model leak to a potentially malicious user. We experimentally validate our approach on the MNIST, CIFAR-10, and ImageNet datasets, evaluating model performance and resilience against watermark removal attacks.
引用
收藏
页码:2121 / 2130
页数:10
相关论文
共 46 条
[1]  
Adi Y, 2018, PROCEEDINGS OF THE 27TH USENIX SECURITY SYMPOSIUM, P1615
[2]   Neural network laundering: Removing black-box backdoor watermarks from deep neural networks [J].
Aiken, William ;
Kim, Hyoungshick ;
Woo, Simon ;
Ryoo, Jungwoo .
COMPUTERS & SECURITY, 2021, 106
[3]  
[Anonymous], 2018, RES ATTACKS INTRUSIO, DOI DOI 10.5220/0006525502730275
[4]  
[Anonymous], 2022 IEEE S SEC PRIV
[5]   A Systematic Review on Model Watermarking for Neural Networks [J].
Boenisch, Franziska .
FRONTIERS IN BIG DATA, 2021, 4
[6]  
Cao Xiaoyu, 2019, ARXIV
[7]   Hardware-Assisted Intellectual Property Protection of Deep Learning Models [J].
Chakraborty, Abhishek ;
Mondal, Ankit ;
Srivastava, Ankur .
PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
[8]  
Chen, 2023, ARXIV
[9]  
Chen H, 2019, ARXIV
[10]   Perceptual Hashing of Deep Convolutional Neural Networks for Model Copy Detection [J].
Chen, Haozhe ;
Zhou, Hang ;
Zhang, Jie ;
Chen, Dongdong ;
Zhang, Weiming ;
Chen, Kejiang ;
Hua, Gang ;
Yu, Nenghai .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (03)