Statistical quality estimation for partially subjective classification tasks through crowdsourcing

被引:0
作者
Yoshinao Sato
Kouki Miyazawa
机构
[1] Fairy Devices Inc.,
来源
Language Resources and Evaluation | 2023年 / 57卷
关键词
Crowdsourcing; Quality estimation; Latent variable model; Partially subjective task;
D O I
暂无
中图分类号
学科分类号
摘要
When constructing a large-scale data resource, the quality of artifacts has great significance, especially when they are generated by creators through crowdsourcing. A widely used approach is to estimate the quality of each artifact based on evaluations by reviewers. However, the commonly used vote-counting method to aggregate reviewers’ evaluations does not work effectively for partially subjective tasks. In such a task, a single correct answer cannot necessarily be defined. We propose a statistical quality estimation method for partially subjective classification tasks to infer the quality of artifacts considering the abilities and biases of creators and reviewers as latent variables. In our experiments, we use the partially subjective task of classifying speech into one of the following four attitudes: agreement, disagreement, stalling, and question. We collect a speech corpus through crowdsourcing and apply the proposed method to it. The results show that the proposed method estimates the quality of speech more effectively than vote aggregation, as measured by correlation with a fine-grained classification performed by experts. Furthermore, we compare the speech attitude classification performance of a neural network model on two subsets of our corpus extracted using the voting and proposed methods. The results indicate that we can effectively extract a consistent and high-quality subset of a corpus using the proposed method. This method facilitates the efficient collection of large-scale data resources for mutually exclusive classification, even if the task is partially subjective.
引用
收藏
页码:31 / 56
页数:25
相关论文
共 46 条
[21]   Towards an Impact-Driven Quality Control Model for Imbalanced Crowdsourcing Tasks [J].
El Maarry, Kinda ;
Balke, Wolf-Tilo .
WEB INFORMATION SYSTEMS ENGINEERING - WISE 2016, PT I, 2016, 10041 :124-139
[22]   A Monte-Carlo Approach to the Value of Information in Crowdsourcing Quality Control Tasks [J].
Xiang Jianwei ;
Liu Shuang ;
Xu Han .
ICMLC 2019: 2019 11TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, 2019, :119-123
[23]   COMPARISON OF CROWDSOURCING AND LABORATORY SETTINGS FOR SUBJECTIVE ASSESSMENT OF VIDEO QUALITY AND ACCEPTABILITY & ANNOYANCE [J].
Ak, Ali ;
Gera, Abhishek ;
Noyes, Denise ;
Tmar, Hassene ;
Katsavounidis, Ioannis ;
Le Callet, Patrick .
2024 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2024, :1159-1164
[24]   Cost Adjustment for Software Crowdsourcing Tasks Using Ensemble Effort Estimation and Topic Modeling [J].
Yasmin, Anum .
ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2024, 49 (09) :12693-12728
[25]   A technical survey on statistical modelling and design methods for crowdsourcing quality control [J].
Jin, Yuan ;
Carman, Mark ;
Zhu, Ye ;
Xiang, Yong .
ARTIFICIAL INTELLIGENCE, 2020, 287
[26]   On subjective quality assessment of adaptive video streaming via crowdsourcing and laboratory based experiments [J].
Jacob Søgaard ;
Muhammad Shahid ;
Jeevan Pokhrel ;
Kjell Brunnström .
Multimedia Tools and Applications, 2017, 76 :16727-16748
[27]   An Impact-driven Model for Quality Control in Skewed-domain Crowdsourcing Tasks [J].
El Maarry, Kinda ;
Balke, Wolf-Tilo .
PROCEEDINGS OF THE 2016 ACM WEB SCIENCE CONFERENCE (WEBSCI'16), 2016, :320-322
[28]   A Decision Tree Based Quality Control Framework for Multi-phase Tasks in Crowdsourcing [J].
Fang, Yili ;
Chen, Pengpeng ;
Sun, Kai ;
Sun, Hailong .
12TH CHINESE CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING (CHINESECSCW 2017), 2017, :10-17
[29]   Impact of the Number of Votes on the Reliability and Validity of Subjective Speech Quality Assessment in the Crowdsourcing Approach [J].
Naderi, Babak ;
Hossfeld, Tobias ;
Hirth, Matthias ;
Metzger, Florian ;
Moeller, Sebastian ;
Jimenez, Rafael Zequeira .
2020 TWELFTH INTERNATIONAL CONFERENCE ON QUALITY OF MULTIMEDIA EXPERIENCE (QOMEX), 2020,
[30]   A Probabilistic Graphical Model for Analyzing the Subjective Visual Quality Assessment Data from Crowdsourcing [J].
Li, Jing ;
Ling, Suiyi ;
Wang, Junle ;
Le Callet, Patrick .
MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, :3339-3347