Discontinuous Distribution of Test Statistics Around Significance Thresholds in Empirical Accounting Studies

被引:10
作者
Chang, Xin [1 ]
Gao, Huasheng [2 ]
Li, Wei [3 ]
机构
[1] Nanyang Technol Univ, Nanyang Business Sch, Singapore, Singapore
[2] Fudan Univ, Int Sch Finance, Shanghai, Peoples R China
[3] City Univ Hong Kong, Dept Accountancy, Hong Kong, Peoples R China
关键词
P-value discontinuity; experimental accounting; archival accounting; researcher degrees of freedom; P-VALUES; PREANALYSIS PLANS; PUBLICATION BIAS; ECONOMICS; REPRODUCIBILITY; PSYCHOLOGY; PREVALENCE;
D O I
10.1111/1475-679X.12579
中图分类号
F8 [财政、金融];
学科分类号
0202 ;
摘要
Examining test statistics from articles in six leading accounting journals, we detect discontinuities in their distributions around conventional significance thresholds (p-values of 0.05 and 0.01) and find an unusual abundance of test statistics that are just significant. Further analysis reveals that these discontinuities are more prominent in studies with smaller samples and are more salient in experimental than in archival studies. The discontinuity discrepancy between experimental and archival studies relates to several proxies for researcher degrees of freedom. Nevertheless, this evidence does not imply that experimental research is more prone to questionable research practices than archival studies. Overall, our findings speak to the concern of whether accounting researchers could exercise undisclosed discretion to obtain and report statistically significant results. Based on our results, a healthy skepticism of some just-significant test statistics is warranted.
引用
收藏
页码:165 / 206
页数:42
相关论文
共 55 条
  • [1] ABDELKHA.AR, 1974, ACCOUNT REV, V49, P743
  • [2] P-hacking in clinical trials and how incentives shape the distribution of results across phases
    Adda, Jerome
    Decker, Christian
    Ottaviani, Marco
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2020, 117 (24) : 13386 - 13392
  • [3] Predicting the replicability of social science lab experiments
    Altmejd, Adam
    Dreber, Anna
    Forsell, Eskil
    Huber, Juergen
    Imai, Taisuke
    Johannesson, Magnus
    Kirchler, Michael
    Nave, Gideon
    Camerer, Colin
    [J]. PLOS ONE, 2019, 14 (12):
  • [4] Identification of and Correction for Publication Bias
    Andrews, Isaiah
    Kasy, Maximilian
    [J]. AMERICAN ECONOMIC REVIEW, 2019, 109 (08) : 2766 - 2794
  • [5] [Anonymous], 2016, Working Paper
  • [6] Banerjee A., 2020, Technical Report
  • [7] Redefine statistical significance
    Benjamin, Daniel J.
    Berger, James O.
    Johannesson, Magnus
    Nosek, Brian A.
    Wagenmakers, E. -J.
    Berk, Richard
    Bollen, Kenneth A.
    Brembs, Bjoern
    Brown, Lawrence
    Camerer, Colin
    Cesarini, David
    Chambers, Christopher D.
    Clyde, Merlise
    Cook, Thomas D.
    De Boeck, Paul
    Dienes, Zoltan
    Dreber, Anna
    Easwaran, Kenny
    Efferson, Charles
    Fehr, Ernst
    Fidler, Fiona
    Field, Andy P.
    Forster, Malcolm
    George, Edward I.
    Gonzalez, Richard
    Goodman, Steven
    Green, Edwin
    Green, Donald P.
    Greenwald, Anthony
    Hadfield, Jarrod D.
    Hedges, Larry V.
    Held, Leonhard
    Ho, Teck Hua
    Hoijtink, Herbert
    Hruschka, Daniel J.
    Imai, Kosuke
    Imbens, Guido
    Ioannidis, John P. A.
    Jeon, Minjeong
    Jones, James Holland
    Kirchler, Michael
    Laibson, David
    List, John
    Little, Roderick
    Lupia, Arthur
    Machery, Edouard
    Maxwell, Scott E.
    McCarthy, Michael
    Moore, Don
    Morgan, Stephen L.
    [J]. NATURE HUMAN BEHAVIOUR, 2018, 2 (01): : 6 - 10
  • [8] Problems in using p-curve analysis and text-mining to detect rate of p-hacking and evidential value
    Bishop, Dorothy V. M.
    Thompson, Paul A.
    [J]. PEERJ, 2016, 4
  • [9] No System Is Perfect: Understanding How Registration-Based Editorial Processes Affect Reproducibility and Investment in Research Quality
    Bloomfield, Robert
    Rennekamp, Kristina
    Steenhoven, Blake
    [J]. JOURNAL OF ACCOUNTING RESEARCH, 2018, 56 (02) : 313 - 362
  • [10] Brodeur A., 2022, We need to talk about mechanical turk: What 22,989 hypothesis tests tell us about publication bias and p-hacking in online experiments