Redefining crowdsourced test report prioritization: An innovative approach with large language model

被引:0
|
作者
Ling, Yuchen [1 ]
Yu, Shengcheng [1 ]
Fang, Chunrong [1 ]
Pan, Guobin [2 ]
Wang, Jun [2 ]
Liu, Jia [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] China Mobile Suzhou Software Technol Co Ltd, Suzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Crowdsourced testing; Mobile app testing; Test report prioritization; Large language model; DUPLICATE;
D O I
10.1016/j.infsof.2024.107629
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Context: Crowdsourced testing has gained popularity in software testing, especially for mobile app testing, due to its ability to bring diversity and tackle fragmentation issues. However, the openness of crowdsourced testing presents challenges, particularly in the manual review of numerous test reports, which is time-consuming and labor-intensive. Objective: The primary goal of this research is to improve the efficiency of review processes in crowdsourced testing. Traditional approaches to test report prioritization lack a deep understanding of semantic information in textual descriptions of these reports. This paper introduces LLMPrior, a novel approach for prioritizing crowdsourced test reports using large language models (LLMs). Method: LLMPrior leverages LLMs for the analysis and clustering of crowdsourced test reports based on the types of bugs revealed in their textual descriptions. This involves using prompt engineering techniques to enhance the performance of LLMs. Following the clustering, a recurrent selection algorithm is applied to prioritize the reports. Results: Empirical experiments are conducted to evaluate the effectiveness of LLMPrior. The findings indicate that LLMPrior not only surpasses current state-of-the-art approaches in terms of performance but also proves to be more feasible, efficient, and reliable. This success is attributed to the use of prompt engineering techniques and the cluster-based prioritization strategy. Conclusion: LLMPrior represents a significant advancement in crowdsourced test report prioritization. By effectively utilizing large language models and a cluster-based strategy, it addresses the challenges in traditional prioritization approaches, offering a more efficient and reliable solution for app developers dealing with crowdsourced test reports.
引用
收藏
页数:13
相关论文
共 39 条
  • [31] Large language model-based code generation for the control of construction assembly robots: A hierarchical generation approach
    Luo, Hanbin
    Wu, Jianxin
    Liu, Jiajing
    Antwi-Afari, Maxwell Fordjour
    DEVELOPMENTS IN THE BUILT ENVIRONMENT, 2024, 19
  • [32] Graph Neural Network vs. Large Language Model: A Comparative Analysis for Bug Report Priority and Severity Prediction
    Acharya, Jagrit
    Ginde, Gouri
    PROCEEDINGS OF THE 20TH INTERNATIONAL CONFERENCE ON PREDICTIVE MODELS AND DATA ANALYTICS IN SOFTWARE ENGINEERING, PROMISE 2024, 2024, : 2 - 11
  • [33] Generating Plausible and Context-Appropriate Comments on Social Media Posts: A Large Language Model-Based Approach
    Ha, Taehyun
    IEEE ACCESS, 2024, 12 : 161545 - 161556
  • [34] GPT meets PubMed: a novel approach to literature review using a large language model to crowdsource migraine medication reviews
    Mackenzie, Elyse
    Cheng, Roger
    Zhang, Pengfei
    BMC NEUROLOGY, 2025, 25 (01)
  • [35] Leveraging on large language model to classify sentences: a case study applying STAGES scoring methodology for sentence completion test on ego development
    Bronlet, Xavier
    FRONTIERS IN PSYCHOLOGY, 2025, 16
  • [36] TOMGPT: Reliable Text-Only Training Approach for Cost-Effective Multi-modal Large Language Model
    Chen, Yunkai
    Wang, Qimeng
    Wu, Shiwei
    Gao, Yan
    Xu, Tong
    Hu, Yao
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2024, 18 (07)
  • [37] Large Language Model Approach for Zero-Shot Information Extraction and Clustering of Japanese Radiology Reports: Algorithm Development and Validation
    Yamagishi, Yosuke
    Nakamura, Yuta
    Hanaoka, Shouhei
    Abe, Osamu
    JMIR CANCER, 2025, 11
  • [38] Claude 2.0 large language model: Tackling a real-world classification problem with a new iterative prompt engineering approach
    Caruccio, Loredana
    Cirillo, Stefano
    Polese, Giuseppe
    Solimando, Giandomenico
    Sundaramurthy, Shanmugam
    Tortora, Genoveffa
    INTELLIGENT SYSTEMS WITH APPLICATIONS, 2024, 21
  • [39] Revisiting time-varying dynamics in stock market forecasting: A multi-source sentiment analysis approach with large language model
    Shao, Zhiqi
    Yao, Xusheng
    Chen, Feng
    Wang, Ze
    Gao, Junbin
    DECISION SUPPORT SYSTEMS, 2025, 190