Statistical quality estimation for partially subjective classification tasks through crowdsourcing

被引:0
|
作者
Yoshinao Sato
Kouki Miyazawa
机构
[1] Fairy Devices Inc.,
来源
Language Resources and Evaluation | 2023年 / 57卷
关键词
Crowdsourcing; Quality estimation; Latent variable model; Partially subjective task;
D O I
暂无
中图分类号
学科分类号
摘要
When constructing a large-scale data resource, the quality of artifacts has great significance, especially when they are generated by creators through crowdsourcing. A widely used approach is to estimate the quality of each artifact based on evaluations by reviewers. However, the commonly used vote-counting method to aggregate reviewers’ evaluations does not work effectively for partially subjective tasks. In such a task, a single correct answer cannot necessarily be defined. We propose a statistical quality estimation method for partially subjective classification tasks to infer the quality of artifacts considering the abilities and biases of creators and reviewers as latent variables. In our experiments, we use the partially subjective task of classifying speech into one of the following four attitudes: agreement, disagreement, stalling, and question. We collect a speech corpus through crowdsourcing and apply the proposed method to it. The results show that the proposed method estimates the quality of speech more effectively than vote aggregation, as measured by correlation with a fine-grained classification performed by experts. Furthermore, we compare the speech attitude classification performance of a neural network model on two subsets of our corpus extracted using the voting and proposed methods. The results indicate that we can effectively extract a consistent and high-quality subset of a corpus using the proposed method. This method facilitates the efficient collection of large-scale data resources for mutually exclusive classification, even if the task is partially subjective.
引用
收藏
页码:31 / 56
页数:25
相关论文
共 45 条
  • [1] Statistical quality estimation for partially subjective classification tasks through crowdsourcing
    Sato, Yoshinao
    Miyazawa, Kouki
    LANGUAGE RESOURCES AND EVALUATION, 2023, 57 (01) : 31 - 56
  • [2] Quality Estimation for Partially Subjective Classification Tasks via Crowdsourcing
    Sato, Yoshinao
    Miyazawa, Kouki
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 229 - 235
  • [3] Statistical Quality Estimation for General Crowdsourcing Tasks
    Baba, Yukino
    Kashima, Hisashi
    19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), 2013, : 554 - 562
  • [4] Debiased Label Aggregation for Subjective Crowdsourcing Tasks
    Wallace, Shaun
    Cai, Tianyuan
    Le, Brendan
    Leiva, Luis A.
    EXTENDED ABSTRACTS OF THE 2022 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2022, 2022,
  • [5] Towards a Classification Model for Tasks in Crowdsourcing
    Alabduljabbar, Reham
    Al-Dossari, Hmood
    PROCEEDINGS OF THE SECOND INTERNATIONAL CONFERENCE ON INTERNET OF THINGS, DATA AND CLOUD COMPUTING (ICC 2017), 2017,
  • [6] Knowledge Enhanced Quality Estimation for Crowdsourcing
    Wang, Shaofei
    Dang, Depeng
    Guo, Zixian
    Chen, Chuangxia
    Yu, Wenhui
    IEEE ACCESS, 2019, 7 : 106693 - 106703
  • [7] CROWDSOURCING SUBJECTIVE IMAGE QUALITY EVALUATION
    Ribeiro, Flavio
    Florencio, Dinei
    Nascimento, Vtor
    2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2011,
  • [8] Subjective Quality Evaluations Using Crowdsourcing
    Salas, Oscar Figuerola
    Adzic, Velibor
    Kalva, Hari
    2013 PICTURE CODING SYMPOSIUM (PCS), 2013, : 418 - 421
  • [9] Tasks-based classification of crowdsourcing initiatives
    Estelles-Arolas, Enrique
    Gonzalez-Ladron-De-Guevara, Fernando
    PROFESIONAL DE LA INFORMACION, 2012, 21 (03): : 283 - 291
  • [10] Measuring the Quality of Annotations for a Subjective Crowdsourcing Task
    Justo, Raquel
    Ines Torres, M.
    Alcaide, Jose M.
    PATTERN RECOGNITION AND IMAGE ANALYSIS (IBPRIA 2017), 2017, 10255 : 58 - 68