StyleAdv: Meta Style Adversarial Training for Cross-Domain Few-Shot Learning

被引:42
作者
Fu, Yuqian [1 ]
Xie, Yu [2 ]
Fu, Yanwei [3 ]
Jiang, Yu-Gang [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Intelligent Informat Proc, Shanghai, Peoples R China
[2] Purple Mt Labs, Nanjing, Peoples R China
[3] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
基金
国家重点研发计划;
关键词
D O I
10.1109/CVPR52729.2023.02354
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-Domain Few-Shot Learning (CD-FSL) is a recently emerging task that tackles few-shot learning across different domains. It aims at transferring prior knowledge learned on the source dataset to novel target datasets. The CD-FSL task is especially challenged by the huge domain gap between different datasets. Critically, such a domain gap actually comes from the changes of visual styles, and wave-SAN [10] empirically shows that spanning the style distribution of the source data helps alleviate this issue. However, wave-SAN simply swaps styles of two images. Such a vanilla operation makes the generated styles "real" and "easy", which still fall into the original set of the source styles. Thus, inspired by vanilla adversarial learning, a novel model-agnostic meta Style Adversarial training (StyleAdv) method together with a novel style adversarial attack method is proposed for CD-FSL. Particularly, our style attack method synthesizes both "virtual" and "hard" adversarial styles for model training. This is achieved by perturbing the original style with the signed style gradients. By continually attacking styles and forcing the model to recognize these challenging adversarial styles, our model is gradually robust to the visual styles, thus boosting the generalization ability for novel target datasets. Besides the typical CNN-based backbone, we also employ our StyleAdv method on large-scale pre-trained vision transformer. Extensive experiments conducted on eight various target datasets show the effectiveness of our method. Whether built upon ResNet or ViT, we achieve the new state of the art for CD-FSL. Code is available at https://github.com/lovelyqian/StyleAdv-CDFSL.
引用
收藏
页码:24575 / 24584
页数:10
相关论文
共 67 条
[1]  
[Anonymous], 2021, ICML
[2]  
Cai John, 2021, CVPR
[3]   Emerging Properties in Self-Supervised Vision Transformers [J].
Caron, Mathilde ;
Touvron, Hugo ;
Misra, Ishan ;
Jegou, Herve ;
Mairal, Julien ;
Bojanowski, Piotr ;
Joulin, Armand .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9630-9640
[4]  
Chen Chen, 2022, MAXSTYLE ADVERSARIAL, P3
[5]  
Codella N., 2019, arXiv
[6]   New Analytical Solution to Predict the Vertical Impedance of a Large-Diameter Pipe Pile in Soil Considering Wave Propagation in Visco-Elastic Continuum [J].
Cui, Chunyi ;
Liang, Zhimeng ;
Xu, Chengshun ;
Xin, Yu ;
Wang, Benlong ;
Meng, Kun .
JOURNAL OF EARTHQUAKE AND TSUNAMI, 2022, 16 (02)
[7]  
Dosovitskiy A., 2020, ICLR 2021
[8]   Adversarial Camouflage: Hiding Physical-World Attacks with Natural Styles [J].
Duan, Ranjie ;
Ma, Xingjun ;
Wang, Yisen ;
Bailey, James ;
Qin, A. K. ;
Yang, Yun .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :997-1005
[9]  
Fu Y., 2021, ACM Multimedia
[10]  
Fu Yuqian, 2022, ACM MULTIMEDIA