RACP: A network with attention corrected prototype for few-shot speaker recognition using indefinite distance metric

被引:9
作者
Wang, Xingmei [1 ]
Meng, Jiaxiang [1 ]
Wen, Bin [1 ]
Xue, Fuzhao [2 ]
机构
[1] Harbin Engn Univ, Coll Comp Sci & Technol, Harbin 150001, Peoples R China
[2] Natl Univ Singapore, Sch Comp, Singapore, Singapore
关键词
Few-shot learning; Speaker recognition; Prototypical networks; Indefinite distance metric; Attention;
D O I
10.1016/j.neucom.2021.11.092
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot speaker recognition task is to identify speakers from limited support samples. We argue that query samples and support samples are both informative for classification. To help Prototypical Networks capture information from query samples, this paper proposes the relation-based indefinite distance metric attentive correction prototype network (RACP). Since the mean prototype deviates from the ideal prototype, we calculate attention scores for each query sample to customize the attention prototype. Then, to compensate for the missed query samples information, the prototype is further refined by correction data that is constructed by combining query samples with the global class attention score. Later, the indefinite distance metric of Relation Networks is introduced on Prototypical Networks, and the relation scores between the sample prototypes and the query samples are calculated for final prediction. Compare with existing methods, RACP can consider both query samples and support samples instead of ignoring the query ones. We compare RACP with strong baselines (e.g. GMM-SVM, MAML, Prototypical Networks, Res32, and VGG11). Ablation study and generalizability study of different scenarios are also conducted on different datasets. Results show that RACP achieves better performance and generalization ability. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:283 / 294
页数:12
相关论文
共 50 条
[41]   CLIP-Driven Prototype Network for Few-Shot Semantic Segmentation [J].
Guo, Shi-Cheng ;
Liu, Shang-Kun ;
Wang, Jing-Yu ;
Zheng, Wei-Min ;
Jiang, Cheng-Yu .
ENTROPY, 2023, 25 (09)
[42]   Few-Shot Metric Transfer Learning Network for Surface Defect Detection [J].
Huang J. ;
Zheng C. ;
Zhang J. ;
Wang B. ;
Chen P. .
Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2021, 34 (05) :407-414
[43]   Siamese-Hashing Network for Few-Shot Palmprin Recognition [J].
Liu, Chengcheng ;
Shao, Huikai ;
Thong, Dexing ;
Du, Jun .
2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, :3251-3258
[44]   Embedding Adaptation Network with Transformer for Few-Shot Action Recognition [J].
Jin, Rongrong ;
Wang, Xiao ;
Wang, Guangge ;
Lu, Yang ;
Hu, Hai-Miao ;
Wang, Hanzi .
ASIAN CONFERENCE ON MACHINE LEARNING, VOL 189, 2022, 189
[45]   Few-shot defect segmentation based on cross-modal attention aggregation and adaptive prototype generation network [J].
Liu, Shi-Tong ;
Zhang, Yun-Zhou ;
Shan, De-Xing ;
Jin, Yang ;
Ning, Jian .
Kongzhi yu Juece/Control and Decision, 2024, 39 (11) :3655-3663
[46]   Hybrid attentive prototypical network for few-shot action recognition [J].
Ruan, Zanxi ;
Wei, Yingmei ;
Guo, Yanming ;
Xie, Yuxiang .
COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (06) :8249-8272
[47]   Hierarchical Motion Excitation Network for Few-Shot Video Recognition [J].
Wang, Bing ;
Wang, Xiaohua ;
Ren, Shiwei ;
Wang, Weijiang ;
Shi, Yueting .
ELECTRONICS, 2023, 12 (05)
[48]   Dual-Metric Neural Network With Attention Guidance for Surface Defect Few-Shot Detection in Smart Manufacturing [J].
Gao, Pengjie ;
Wang, Junliang ;
Xia, Min ;
Qin, Zijin ;
Zhang, Jie .
JOURNAL OF MANUFACTURING SCIENCE AND ENGINEERING-TRANSACTIONS OF THE ASME, 2023, 145 (12)
[49]   Multi-level Attention Feature Network for Few-shot Learning [J].
Wang R. ;
Han M. ;
Yang J. ;
Xue L. ;
Hu M. .
Yang, Juan (yangjuan@hfut.edu.cn), 1600, Science Press (42) :772-778
[50]   Bidirectional Patch-Aware Attention Network for Few-Shot Learning [J].
Mao, Yu ;
Lin, Shaojie ;
Lin, Zilong ;
Lin, Yaojin .
IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025,