Learning from crowds with decision trees

被引:14
|
作者
Yang, Wenjun [1 ]
Li, Chaoqun [1 ]
Jiang, Liangxiao [2 ]
机构
[1] China Univ Geosci, Sch Math & Phys, Wuhan 430074, Peoples R China
[2] China Univ Geosci, Sch Comp Sci, Wuhan 430074, Peoples R China
关键词
Crowdsourcing learning; Weighted majority voting; Decision trees; MODEL QUALITY; STATISTICAL COMPARISONS; WEIGHTING FILTER; IMPROVING DATA; CLASSIFIERS; TOOL;
D O I
10.1007/s10115-022-01701-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Crowdsourcing systems provide an efficient way to collect labeled data by employing non-expert crowd workers. In practice, each instance obtains a multiple noisy label set from different workers. Ground truth inference algorithms are designed to infer the unknown true labels of data from multiple noisy label sets. Since there is substantial variation among different workers, evaluating the qualities of workers is crucial for ground truth inference. This paper proposes a novel algorithm called decision tree-based weighted majority voting (DTWMV). DTWMV directly takes the multiple noisy label set of each instance as its feature vector; that is, each worker is a feature of instances. Then sequential decision trees are built to calculate the weight of each feature (worker). Finally weighted majority voting is used to infer the integrated labels of instances. In DTWMV, evaluating the qualities of workers is converted to calculating the weights of features, which provides a new perspective for solving the ground truth inference problem. Then, a novel feature weight measurement based on decision trees is proposed. Our experimental results show that DTWMV can effectively evaluate the qualities of workers and improve the label quality of data.
引用
收藏
页码:2123 / 2140
页数:18
相关论文
共 50 条
  • [31] Versatile Decision Trees for Learning Over Multiple Contexts
    Al-Otaibi, Reem
    Prudencio, Ricardo B. C.
    Kull, Meelis
    Flach, Peter
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2015, PT I, 2015, 9284 : 184 - 199
  • [32] Learning compact Markov logic networks with decision trees
    Khosravi, Hassan
    Schulte, Oliver
    Hu, Jianfeng
    Gao, Tianxiang
    MACHINE LEARNING, 2012, 89 (03) : 257 - 277
  • [33] Properly Learning Decision Trees in almost Polynomial Time
    Blanc, Guy
    Lange, Jane
    Qiao, Mingda
    Tan, Li-Yang
    JOURNAL OF THE ACM, 2022, 69 (06)
  • [34] Learning Optimal Decision Trees Under Memory Constraints
    Aglin, Gael
    Nijssen, Siegfried
    Schaus, Pierre
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT V, 2023, 13717 : 393 - 409
  • [35] Using POMDPs for learning cost sensitive decision trees
    Maliah, Shlomi
    Shani, Guy
    ARTIFICIAL INTELLIGENCE, 2021, 292
  • [36] Learning Behavior Trees From Planning Experts Using Decision Tree and Logic Factorization
    Gugliermo, Simona
    Schaffernicht, Erik
    Koniaris, Christos
    Pecora, Federico
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (06) : 3534 - 3541
  • [37] Learning compact Markov logic networks with decision trees
    Hassan Khosravi
    Oliver Schulte
    Jianfeng Hu
    Tianxiang Gao
    Machine Learning, 2012, 89 : 257 - 277
  • [38] Learning Invariants using Decision Trees and Implication Counterexamples
    Garg, Pranav
    Neider, Daniel
    Madhusudan, P.
    Roth, Dan
    ACM SIGPLAN NOTICES, 2016, 51 (01) : 499 - 512
  • [39] The limitations of decision trees and automatic learning in real world medical decision making
    Kokol, P
    Zorman, M
    Stiglic, MM
    Malèiæ, I
    MEDINFO '98 - 9TH WORLD CONGRESS ON MEDICAL INFORMATICS, PTS 1 AND 2, 1998, 52 : 529 - 533
  • [40] Multidimensional Learning from Crowds: Usefulness and Application of Expertise Detection
    Hernandez-Gonzalez, Jeronimo
    Inza, Inaki
    Lozano, Jose A.
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2015, 30 (03) : 326 - 354