Learning from crowds with decision trees

被引:14
|
作者
Yang, Wenjun [1 ]
Li, Chaoqun [1 ]
Jiang, Liangxiao [2 ]
机构
[1] China Univ Geosci, Sch Math & Phys, Wuhan 430074, Peoples R China
[2] China Univ Geosci, Sch Comp Sci, Wuhan 430074, Peoples R China
关键词
Crowdsourcing learning; Weighted majority voting; Decision trees; MODEL QUALITY; STATISTICAL COMPARISONS; WEIGHTING FILTER; IMPROVING DATA; CLASSIFIERS; TOOL;
D O I
10.1007/s10115-022-01701-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Crowdsourcing systems provide an efficient way to collect labeled data by employing non-expert crowd workers. In practice, each instance obtains a multiple noisy label set from different workers. Ground truth inference algorithms are designed to infer the unknown true labels of data from multiple noisy label sets. Since there is substantial variation among different workers, evaluating the qualities of workers is crucial for ground truth inference. This paper proposes a novel algorithm called decision tree-based weighted majority voting (DTWMV). DTWMV directly takes the multiple noisy label set of each instance as its feature vector; that is, each worker is a feature of instances. Then sequential decision trees are built to calculate the weight of each feature (worker). Finally weighted majority voting is used to infer the integrated labels of instances. In DTWMV, evaluating the qualities of workers is converted to calculating the weights of features, which provides a new perspective for solving the ground truth inference problem. Then, a novel feature weight measurement based on decision trees is proposed. Our experimental results show that DTWMV can effectively evaluate the qualities of workers and improve the label quality of data.
引用
收藏
页码:2123 / 2140
页数:18
相关论文
共 50 条
  • [41] Optimal Decision Trees Generation from OR-Decision Tables
    Grana, Costantino
    Montangero, Manuela
    Borghesani, Daniele
    Cucchiara, Rita
    IMAGE ANALYSIS AND PROCESSING - ICIAP 2011, PT I, 2011, 6978 : 443 - 452
  • [42] Learning From Crowds With Multiple Noisy Label Distribution Propagation
    Jiang, Liangxiao
    Zhang, Hao
    Tao, Fangna
    Li, Chaoqun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) : 6558 - 6568
  • [44] Automated decision-making with DMN: from decision trees to decision tables
    Etinger, D.
    Simic, S. D.
    Buljubasic, L.
    2019 42ND INTERNATIONAL CONVENTION ON INFORMATION AND COMMUNICATION TECHNOLOGY, ELECTRONICS AND MICROELECTRONICS (MIPRO), 2019, : 1309 - 1313
  • [45] A Monte Carlo tree search approach to learning decision trees
    Nunes, Cecilia
    De Craene, Mathieu
    Langet, Helene
    Camara, Oscar
    Jonsson, Anders
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 429 - 435
  • [46] Learning Customised Decision Trees for Domain-knowledge Constraints
    Nanfack, Geraldin
    Temple, Paul
    Frenay, Benoit
    PATTERN RECOGNITION, 2023, 142
  • [47] Properly learning decision trees with queries is NP-hard
    Koch, Caleb
    Strassle, Carmen
    Tan, Li-Yang
    2023 IEEE 64TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, FOCS, 2023, : 2383 - 2407
  • [48] Learning Optimal and Fair Decision Trees for Non-Discriminative Decision-Making
    Aghaei, Sina
    Azizi, Mohammad Javad
    Vayanos, Phebe
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 1418 - 1426
  • [49] A heuristic for learning decision trees and pruning them into classification rules
    Ranilla, J
    Luaces, O
    Bahamonde, A
    AI COMMUNICATIONS, 2003, 16 (02) : 71 - 87
  • [50] Decision Trees in Federated Learning: Current State and Future Opportunities
    Heiyanthuduwage, Sudath R.
    Altas, Irfan
    Bewong, Michael
    Islam, Md Zahidul
    Deho, Oscar B.
    IEEE ACCESS, 2024, 12 : 127943 - 127965