Statistical quality estimation for partially subjective classification tasks through crowdsourcing

被引:0
作者
Yoshinao Sato
Kouki Miyazawa
机构
[1] Fairy Devices Inc.,
来源
Language Resources and Evaluation | 2023年 / 57卷
关键词
Crowdsourcing; Quality estimation; Latent variable model; Partially subjective task;
D O I
暂无
中图分类号
学科分类号
摘要
When constructing a large-scale data resource, the quality of artifacts has great significance, especially when they are generated by creators through crowdsourcing. A widely used approach is to estimate the quality of each artifact based on evaluations by reviewers. However, the commonly used vote-counting method to aggregate reviewers’ evaluations does not work effectively for partially subjective tasks. In such a task, a single correct answer cannot necessarily be defined. We propose a statistical quality estimation method for partially subjective classification tasks to infer the quality of artifacts considering the abilities and biases of creators and reviewers as latent variables. In our experiments, we use the partially subjective task of classifying speech into one of the following four attitudes: agreement, disagreement, stalling, and question. We collect a speech corpus through crowdsourcing and apply the proposed method to it. The results show that the proposed method estimates the quality of speech more effectively than vote aggregation, as measured by correlation with a fine-grained classification performed by experts. Furthermore, we compare the speech attitude classification performance of a neural network model on two subsets of our corpus extracted using the voting and proposed methods. The results indicate that we can effectively extract a consistent and high-quality subset of a corpus using the proposed method. This method facilitates the efficient collection of large-scale data resources for mutually exclusive classification, even if the task is partially subjective.
引用
收藏
页码:31 / 56
页数:25
相关论文
共 45 条
  • [21] Towards an Impact-Driven Quality Control Model for Imbalanced Crowdsourcing Tasks
    El Maarry, Kinda
    Balke, Wolf-Tilo
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2016, PT I, 2016, 10041 : 124 - 139
  • [22] A Monte-Carlo Approach to the Value of Information in Crowdsourcing Quality Control Tasks
    Xiang Jianwei
    Liu Shuang
    Xu Han
    ICMLC 2019: 2019 11TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, 2019, : 119 - 123
  • [23] COMPARISON OF CROWDSOURCING AND LABORATORY SETTINGS FOR SUBJECTIVE ASSESSMENT OF VIDEO QUALITY AND ACCEPTABILITY & ANNOYANCE
    Ak, Ali
    Gera, Abhishek
    Noyes, Denise
    Tmar, Hassene
    Katsavounidis, Ioannis
    Le Callet, Patrick
    2024 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2024, : 1159 - 1164
  • [24] Cost Adjustment for Software Crowdsourcing Tasks Using Ensemble Effort Estimation and Topic Modeling
    Yasmin, Anum
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2024, 49 (09) : 12693 - 12728
  • [25] A technical survey on statistical modelling and design methods for crowdsourcing quality control
    Jin, Yuan
    Carman, Mark
    Zhu, Ye
    Xiang, Yong
    ARTIFICIAL INTELLIGENCE, 2020, 287
  • [26] An Impact-driven Model for Quality Control in Skewed-domain Crowdsourcing Tasks
    El Maarry, Kinda
    Balke, Wolf-Tilo
    PROCEEDINGS OF THE 2016 ACM WEB SCIENCE CONFERENCE (WEBSCI'16), 2016, : 320 - 322
  • [27] A Decision Tree Based Quality Control Framework for Multi-phase Tasks in Crowdsourcing
    Fang, Yili
    Chen, Pengpeng
    Sun, Kai
    Sun, Hailong
    12TH CHINESE CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING (CHINESECSCW 2017), 2017, : 10 - 17
  • [28] On subjective quality assessment of adaptive video streaming via crowdsourcing and laboratory based experiments
    Jacob Søgaard
    Muhammad Shahid
    Jeevan Pokhrel
    Kjell Brunnström
    Multimedia Tools and Applications, 2017, 76 : 16727 - 16748
  • [29] Impact of the Number of Votes on the Reliability and Validity of Subjective Speech Quality Assessment in the Crowdsourcing Approach
    Naderi, Babak
    Hossfeld, Tobias
    Hirth, Matthias
    Metzger, Florian
    Moeller, Sebastian
    Jimenez, Rafael Zequeira
    2020 TWELFTH INTERNATIONAL CONFERENCE ON QUALITY OF MULTIMEDIA EXPERIENCE (QOMEX), 2020,
  • [30] A Probabilistic Graphical Model for Analyzing the Subjective Visual Quality Assessment Data from Crowdsourcing
    Li, Jing
    Ling, Suiyi
    Wang, Junle
    Le Callet, Patrick
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3339 - 3347