BadCM: Invisible Backdoor Attack Against Cross-Modal Learning

被引:5
作者
Zhang, Zheng [1 ,2 ]
Yuan, Xu [1 ]
Zhu, Lei [3 ]
Song, Jingkuan [4 ]
Nie, Liqiang [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Tongji Univ, Sch Elect & Informat Engn, Shanghai 200092, Peoples R China
[4] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
关键词
Visualization; Training; Flowering plants; Perturbation methods; Dogs; Generators; Computational modeling; Backdoor attacks; cross-modal learning; dataset security; imperceptibility;
D O I
10.1109/TIP.2024.3378918
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite remarkable successes in unimodal learning tasks, backdoor attacks against cross-modal learning are still underexplored due to the limited generalization and inferior stealthiness when involving multiple modalities. Notably, since works in this area mainly inherit ideas from unimodal visual attacks, they struggle with dealing with diverse cross-modal attack circumstances and manipulating imperceptible trigger samples, which hinders their practicability in real-world applications. In this paper, we introduce a novel bilateral backdoor to fill in the missing pieces of the puzzle in the cross-modal backdoor and propose a generalized invisible backdoor framework against cross-modal learning (BadCM). Specifically, a cross-modal mining scheme is developed to capture the modality-invariant components as target poisoning areas, where well-designed trigger patterns injected into these regions can be efficiently recognized by the victim models. This strategy is adapted to different image-text cross-modal models, making our framework available to various attack scenarios. Furthermore, for generating poisoned samples of high stealthiness, we conceive modality-specific generators for visual and linguistic modalities that facilitate hiding explicit trigger patterns in modality-invariant regions. To the best of our knowledge, BadCM is the first invisible backdoor method deliberately designed for diverse cross-modal attacks within one unified framework. Comprehensive experimental evaluations on two typical applications, i.e., cross-modal retrieval and VQA, demonstrate the effectiveness and generalization of our method under multiple kinds of attack scenarios. Moreover, we show that BadCM can robustly evade existing backdoor defenses. Our code is available at https://github.com/xandery-geek/BadCM.
引用
收藏
页码:2558 / 2571
页数:14
相关论文
共 65 条
  • [1] Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering
    Anderson, Peter
    He, Xiaodong
    Buehler, Chris
    Teney, Damien
    Johnson, Mark
    Gould, Stephen
    Zhang, Lei
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6077 - 6086
  • [2] VQA: Visual Question Answering
    Antol, Stanislaw
    Agrawal, Aishwarya
    Lu, Jiasen
    Mitchell, Margaret
    Batra, Dhruv
    Zitnick, C. Lawrence
    Parikh, Devi
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2425 - 2433
  • [3] Adversarial Attack Against Deep Saliency Models Powered by Non-Redundant Priors
    Che, Zhaohui
    Borji, Ali
    Zhai, Guangtao
    Ling, Suiyi
    Li, Jing
    Tian, Yuan
    Guo, Guodong
    Le Callet, Patrick
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1973 - 1988
  • [4] Chen Wuyang, 2022, Advances in Neural Information Processing Systems
  • [5] Kallima: A Clean-Label Framework for Textual Backdoor Attacks
    Chen, Xiaoyi
    Dong, Yinpeng
    Sun, Zeyu
    Zhai, Shengfang
    Shen, Qingni
    Wu, Zhonghai
    [J]. COMPUTER SECURITY - ESORICS 2022, PT I, 2022, 13554 : 447 - 466
  • [6] BadNL: Backdoor Attacks against NLP Models with Semantic-preserving Improvements
    Chen, Xiaoyi
    Salem, Ahmed
    Chen, Dingfan
    Backes, Michael
    Ma, Shiqing
    Shen, Qingni
    Wu, Zhonghai
    Zhang, Yang
    [J]. 37TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2021, 2021, : 554 - 569
  • [7] Chen XY, 2017, Arxiv, DOI arXiv:1712.05526
  • [8] A Backdoor Attack Against LSTM-Based Text Classification Systems
    Dai, Jiazhu
    Chen, Chuanshuai
    Li, Yufeng
    [J]. IEEE ACCESS, 2019, 7 : 138872 - 138878
  • [9] Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
  • [10] Februus: Input Purification Defense Against Trojan Attacks on Deep Neural Network Systems
    Doan, Bao Gia
    Abbasnejad, Ehsan
    Ranasinghe, Damith C.
    [J]. 36TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSAC 2020), 2020, : 897 - 912