Transferring Annotator- and Instance-Dependent Transition Matrix for Learning From Crowds

被引:0
|
作者
Li, Shikun [1 ,2 ]
Xia, Xiaobo [3 ]
Deng, Jiankang [4 ]
Ge, Shiming [1 ,2 ]
Liu, Tongliang [3 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing 100095, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing 100049, Peoples R China
[3] Univ Sydney, Fac Engn, Sydney AI Ctr, Sch Comp Sci, Darlington, NSW 2008, Australia
[4] Imperial Coll London, Dept Comp, London SW7 2BX, England
基金
澳大利亚研究理事会;
关键词
Noise; Annotations; Noise measurement; Knowledge transfer; Data models; Sparse matrices; Estimation; Learning from crowds; label-noise learning; noise transition matrix; knowledge transfer;
D O I
10.1109/TPAMI.2024.3388209
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from crowds describes that the annotations of training data are obtained with crowd-sourcing services. Multiple annotators each complete their own small part of the annotations, where labeling mistakes that depend on annotators occur frequently. Modeling the label-noise generation process by the noise transition matrix is a powerful tool to tackle the label noise. In real-world crowd-sourcing scenarios, noise transition matrices are both annotator- and instance-dependent. However, due to the high complexity of annotator- and instance-dependent transition matrices (AIDTM), annotation sparsity, which means each annotator only labels a tiny part of instances, makes modeling AIDTM very challenging. Without prior knowledge, existing works simplify the problem by assuming the transition matrix is instance-independent or using simple parametric ways, which lose modeling generality. Motivated by this, we target a more realistic problem, estimating general AIDTM in practice. Without losing modeling generality, we parameterize AIDTM with deep neural networks. To alleviate the modeling challenge, we suppose every annotator shares its noise pattern with similar annotators, and estimate AIDTM via knowledge transfer. We hence first model the mixture of noise patterns by all annotators, and then transfer this modeling to individual annotators. Furthermore, considering that the transfer from the mixture of noise patterns to individuals may cause two annotators with highly different noise generations to perturb each other, we employ the knowledge transfer between identified neighboring annotators to calibrate the modeling. Theoretical analyses are derived to demonstrate that both the knowledge transfer from global to individuals and the knowledge transfer between neighboring individuals can effectively help mitigate the challenge of modeling general AIDTM. Experiments confirm the superiority of the proposed approach on synthetic and real-world crowd-sourcing data.
引用
收藏
页码:7377 / 7391
页数:15
相关论文
共 8 条
  • [1] A Time-Consistency Curriculum for Learning From Instance-Dependent Noisy Labels
    Wu, Songhua
    Zhou, Tianyi
    Du, Yuxuan
    Yu, Jun
    Han, Bo
    Liu, Tongliang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (07) : 4830 - 4842
  • [2] Instance-Dependent Inaccurate Label Distribution Learning
    Kou, Zhiqiang
    Wang, Jing
    Jia, Yuheng
    Liu, Biao
    Geng, Xin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 1425 - 1437
  • [3] Cognition-Driven Structural Prior for Instance-Dependent Label Transition Matrix Estimation
    Zhang, Ruiheng
    Cao, Zhe
    Yang, Shuo
    Si, Lingyu
    Sun, Haoyang
    Xu, Lixin
    Sun, Fuchun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (02) : 3730 - 3743
  • [4] Instance-dependent Label Distribution Estimation for Learning with Label Noise
    Liao, Zehui
    Hu, Shishuai
    Xie, Yutong
    Xia, Yong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, : 2568 - 2580
  • [5] Instance-Dependent Noisy-Label Learning with Graphical Model Based Noise-Rate Estimation
    Garg, Arpit
    Cuong Nguyen
    Felix, Rafael
    Thanh-Toan Do
    Carneiro, Gustavo
    COMPUTER VISION-ECCV 2024, PT IV, 2025, 15062 : 372 - 389
  • [6] Beyond confusion matrix: learning from multiple annotators with awareness of instance features
    Li, Jingzheng
    Sun, Hailong
    Li, Jiyi
    MACHINE LEARNING, 2023, 112 (03) : 1053 - 1075
  • [7] Beyond confusion matrix: learning from multiple annotators with awareness of instance features
    Jingzheng Li
    Hailong Sun
    Jiyi Li
    Machine Learning, 2023, 112 : 1053 - 1075
  • [8] Learning Sparse High-Dimensional Matrix-Valued Graphical Models From Dependent Data
    Tugnait, Jitendra K.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 3363 - 3379