Few-shot Learning for Multi-Modality Tasks

被引:1
作者
Chen, Jie [1 ,2 ]
Ye, Qixiang [2 ,3 ]
Yang, Xiaoshan [2 ,4 ]
Zhou, S. Kevin [2 ,5 ]
Hong, Xiaopeng [2 ,6 ]
Zhang, Li [7 ]
机构
[1] Peking Univ, Sch Elect & Comp Engn, Beijing, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
[4] Univ Chinese Acad Sci, NLPR, CASIA, Beijing, Peoples R China
[5] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
[6] Xi An Jiao Tong Univ, Xian, Peoples R China
[7] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021 | 2021年
关键词
Few-shot learning; multi-modal learning;
D O I
10.1145/3474085.3478873
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent deep learning methods rely on a large amount of labeled data to achieve high performance. These methods may be impractical in some scenarios, where manual data annotation is costly or the samples of certain categories are scarce (e.g., tumor lesions, endangered animals and rare individual activities). When only limited annotated samples are available, these methods usually suffer from the overfitting problem severely, which degrades the performance significantly. In contrast, humans can recognize the objects in the images rapidly and correctly with their prior knowledge after exposed to only a few annotated samples. To simulate the learning schema of humans and relieve the reliance on the large-scale annotation benchmarks, researchers start shifting towards the few-shot learning problem: they try to learn a model to correctly recognize novel categories with only a few annotated samples.
引用
收藏
页码:5673 / 5674
页数:2
相关论文
empty
未找到相关数据