Improving the Reporting of Threats to Construct Validity

被引:4
作者
Sjoberg, Dag I. K. [1 ]
Bergersen, Gunnar R. [1 ]
机构
[1] Univ Oslo, Oslo, Norway
来源
27TH INTERNATIONAL CONFERENCE ON EVALUATION AND ASSESSMENT IN SOFTWARE ENGINEERING, EASE 2023 | 2023年
关键词
measurement; research quality; empirical research;
D O I
10.1145/3593434.3593449
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Background: Construct validity concerns the use of indicators to measure a concept that is not directly measurable. Aim: This study intends to identify, categorize, assess and quantify discussions of threats to construct validity in empirical software engineering literature and use the findings to suggest ways to improve the reporting of construct validity issues. Method: We analyzed 83 articles that report human-centric experiments published in five top-tier software engineering journals from 2015 to 2019. The articles' text concerning threats to construct validity was divided into segments (the unit of analysis) based on predefined categories. The segments were then evaluated regarding whether they clearly discussed a threat and a construct. Results: Three-fifths of the segments were associated with topics not related to construct validity. Two-thirds of the articles discussed construct validity without using the definition of construct validity given in the article. The threats were clearly described in more than four-fifths of the segments, but the construct in question was clearly described in only two-thirds of the segments. The construct was unclear when the discussion was not related to construct validity but to other types of validity. Conclusions: The results show potential for improving the understanding of construct validity in software engineering. Recommendations addressing the identified weaknesses are given to improve the awareness and reporting of CV.
引用
收藏
页码:205 / 209
页数:5
相关论文
共 14 条
[1]  
[Anonymous], 1969, Artifacts in behavioral research
[2]  
Cook T. D., 1979, Reproduced in part in research in organizations: Issues and controversies
[3]   CONSTRUCT VALIDITY IN PSYCHOLOGICAL TESTS [J].
CRONBACH, LJ ;
MEEHL, PE .
PSYCHOLOGICAL BULLETIN, 1955, 52 (04) :281-302
[4]  
Jedlitschka Andreas, 2008, Reporting experiments in software engineering, P201, DOI [DOI 10.1007/978-1-84800-044-5_8, DOI 10.1007/978-1-84800-044-58]
[5]   Evaluating guidelines for reporting empirical software engineering studies [J].
Kitchenham, Barbara ;
Al-Khilidar, Hiyam ;
Babar, Muhammed Ali ;
Berry, Mike ;
Cox, Karl ;
Keung, Jacky ;
Kurniawati, Felicia ;
Staples, Mark ;
Zhang, He ;
Zhu, Liming .
EMPIRICAL SOFTWARE ENGINEERING, 2008, 13 (01) :97-121
[6]   Trends in the Quality of Human-Centric Software Engineering Experiments-A Quasi-Experiment [J].
Kitchenham, Barbara ;
Sjoberg, Dag I. K. ;
Dyba, Tore ;
Brereton, O. Pearl ;
Budgen, David ;
Host, Martin ;
Runeson, Per .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2013, 39 (07) :1002-1017
[7]  
Messick S., 1989, Educational Measurement, V3rd, P13, DOI DOI 10.7203/RELIEVE.22.1.8248
[8]  
Ralph Paul, 2018, P 22 INT C EVALUATIO, P13, DOI DOI 10.1145/3210459.3210461
[9]   Guidelines for conducting and reporting case study research in software engineering [J].
Runeson, Per ;
Hoest, Martin .
EMPIRICAL SOFTWARE ENGINEERING, 2009, 14 (02) :131-164
[10]  
Shadish W. R., 2002, Experimental and quasi-experimental designs for generalised causal inference