Counterfactual Generation Framework for Few-Shot Learning

被引:22
作者
Dang, Zhuohang [1 ]
Luo, Minnan [1 ]
Jia, Chengyou [1 ]
Yan, Caixia [1 ]
Chang, Xiaojun [2 ]
Zheng, Qinghua [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Comp Sci & Technol, Xian 710049, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ultimo, NSW 2007, Australia
基金
中国国家自然科学基金;
关键词
Few-shot learning; counterfactual inference; data augmentation; prototype learning; IMAGE CLASSIFICATION; ALIGNMENT; NETWORK;
D O I
10.1109/TCSVT.2023.3241651
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Few-shot learning (FSL) that aims to recognize novel classes with few labeled samples is troubled by its data scarcity. Though recent works tackle FSL with data augmentation-based methods, these models fail to maintain the discrimination and diversity of the generated samples due to the distribution shift and intra-class bias caused by the data scarcity, therefore greatly undermining the performance. To this end, we use causal mechanisms, which are constant among independent variables across data distribution, to alleviate such effects. In this sense, we decompose the image information into two independent components: sample-specific and class-agnostic information, and further propose a novel Counterfactual Generation Framework (CGF) to learn the underlying causal mechanisms to synthesize faithful samples for FSL. Specifically, based on the counterfactual inference, we design a class-agnostic feature extractor to capture the sample-specific information, together with a counterfactual generation network to simulate the data generation process from a causal perspective. Moreover, to leverage the power of CGF in counterfactual inference, we further develop a novel classifier that classifies samples based on their distributions of counterfactual generations. Extensive experiments demonstrate the effectiveness of CGF on four FSL benchmarks, e.g., 80.12/86.13% accuracy on 5-way 1/5-shot miniImageNet FSL tasks, significantly improving the performance. Our codes and models are available at https://github.com/eric-hang/CGF.
引用
收藏
页码:3747 / 3758
页数:12
相关论文
共 73 条
[1]  
Asgari S., 2022, ADV NEURAL INF PROCE, P1
[2]   Recognition in Terra Incognita [J].
Beery, Sara ;
Van Horn, Grant ;
Perona, Pietro .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :472-489
[3]  
Bertinetto L., 2019, INT C LEARNING REPRE
[4]   Hierarchical Graph Neural Networks for Few-Shot Learning [J].
Chen, Cen ;
Li, Kenli ;
Wei, Wei ;
Zhou, Joey Tianyi ;
Zeng, Zeng .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) :240-252
[5]   Image Deformation Meta-Networks for One-Shot Learning [J].
Chen, Zitian ;
Fu, Yanwei ;
Wang, Yu-Xiong ;
Ma, Lin ;
Liu, Wei ;
Hebert, Martial .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :8672-8681
[6]   Multi-Level Semantic Feature Augmentation for One-Shot Learning [J].
Chen, Zitian ;
Fu, Yanwei ;
Zhang, Yinda ;
Jiang, Yu-Gang ;
Xue, Xiangyang ;
Sigal, Leonid .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (09) :4594-4605
[7]   Meta-Learning-Based Incremental Few-Shot Object Detection [J].
Cheng, Meng ;
Wang, Hanli ;
Long, Yu .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (04) :2158-2169
[8]  
Cheng PY, 2020, PR MACH LEARN RES, V119
[9]   Learning to Capture the Query Distribution for Few-Shot Learning [J].
Chi, Ziqiu ;
Wang, Zhe ;
Yang, Mengping ;
Li, Dongdong ;
Du, Wenli .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) :4163-4173
[10]  
Chickering D. M., 2003, Journal of Machine Learning Research, V3, P507, DOI 10.1162/153244303321897717