Sugar-coated poison defense on deepfake face-swapping attacks

被引:0
作者
Guo, Cheng-Yao [1 ]
Yu, Fang [1 ]
机构
[1] Natl Chengchi Univ, Taipei, Taiwan
来源
PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATION OF SOFTWARE TEST, AST 2024 | 2024年
关键词
Poison defense; Deepfake; Face-swapping;
D O I
10.1145/3644032.3644459
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The deployment of deepfake face-swapping technology has matured, becoming widespread on the Internet. The misuse of this technology raises significant concerns for application security and privacy. To counter deepfake threats, we propose a sugar-coated poison defense targeting the latent vectors of generative models. This strategy aims to impact visual effects without substantially increasing reconstruction loss. We establish metrics for visual effects and reconstruction loss to assess perturbation effects on latent vectors, emphasizing those with the most significant impact on visual effects while minimizing reconstruction loss. Our approach begins by utilizing a facial feature extraction model to convert faces into latent representations. We then introduce two latent selection methods: 1) shap-based latent selection using a linear regression model for approximation, and 2) grid search latent selection employing heuristics of adversarial attacks. These methods pinpoint vectors that, when perturbed, can increase face landmark distances while maintaining low mean square errors, commonly used as the optimization metric in deepfake reconstruction models. We apply inconsistent perturbations to selected latent vectors in video frames, acting as sugar-coated poison for deepfake face-swapping applications. Preliminary results demonstrate that these perturbations can be applied to individual videos, resulting in low reconstruction loss. Importantly, they induce measurable consistency reduction in deepfake videos, making them more discernible and accessible to identify.
引用
收藏
页码:78 / 87
页数:10
相关论文
共 40 条
  • [1] Agarwal S., 2019, P IEEE C COMP VIS PA, V1, P38
  • [2] TAFIM: Targeted Adversarial Attacks Against Facial Image Manipulations
    Aneja, Shivangi
    Markhasin, Lev
    Niessner, Matthias
    [J]. COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 58 - 75
  • [3] [Anonymous], 2022, Dlib library
  • [4] [Anonymous], 2022, Faceswap
  • [5] Deep Generative Modelling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models
    Bond-Taylor, Sam
    Leach, Adam
    Long, Yang
    Willcocks, Chris G.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (11) : 7327 - 7347
  • [6] Buitinck L, 2013, API DESIGN MACHINE L
  • [7] Choi Y, 2020, PROC CVPR IEEE, P8185, DOI 10.1109/CVPR42600.2020.00821
  • [8] StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation
    Choi, Yunjey
    Choi, Minje
    Kim, Munyoung
    Ha, Jung-Woo
    Kim, Sunghun
    Choo, Jaegul
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8789 - 8797
  • [9] Ciftci Umur Aybars, 2020, IEEE Trans Pattern Anal Mach Intell, VPP, DOI 10.1109/TPAMI.2020.3009287
  • [10] UNIVERSAL ADVERSARIAL ROBUSTNESS OF TEXTURE AND SHAPE-BIASED MODELS
    Co, Kenneth T.
    Munoz-Gonzalez, Luis
    Kanthan, Leslie
    Glocker, Ben
    Lupu, Emil C.
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 799 - 803