Addressing the "Replication Crisis": Using Original Studies to Design Replication Studies with Appropriate Statistical Power

被引:80
作者
Anderson, Samantha F. [1 ]
Maxwell, Scott E. [1 ]
机构
[1] Univ Notre Dame, Dept Psychol, 118 Haggar Hall, Notre Dame, IN 46556 USA
关键词
Effect size; power; replication; sample size planning; EFFECT SIZE; EMERGENCE; ACCURACY; BIAS;
D O I
10.1080/00273171.2017.1289361
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Psychology is undergoing a replication crisis. The discussion surrounding this crisis has centered on mistrust of previous findings. Researchers planning replication studies often use the original study sample effect size as the basis for sample size planning. However, this strategy ignores uncertainty and publication bias in estimated effect sizes, resulting in overly optimistic calculations. A psychologist who intends to obtain power of .80 in the replication study, and performs calculations accordingly, may have an actual power lower than .80. We performed simulations to reveal the magnitude of the difference between actual and intended power based on common sample size planning strategies and assessed the performance of methods that aim to correct for effect size uncertainty and/or bias. Our results imply that even if original studies reflect actual phenomena and were conducted in the absence of questionable research practices, popular approaches to designing replication studies may result in a low success rate, especially if the original study is underpowered. Methods correcting for bias and/or uncertainty generally had higher actual power, but were not a panacea for an underpowered original study. Thus, it becomes imperative that 1) original studies are adequately powered and 2) replication studies are designed with methods that are more likely to yield the intended level of power.
引用
收藏
页码:305 / 324
页数:20
相关论文
共 42 条
  • [1] Registered Replication Report: Schooler and Engstler-Schooler (1990)
    Alogna, V. K.
    Attaya, M. K.
    Aucoin, P.
    Bahnik, S.
    Birch, S.
    Birt, A. R.
    Bornstein, B. H.
    Bouwmeester, S.
    Brandimonte, M. A.
    Brown, C.
    Buswell, K.
    Carlson, C.
    Carlson, M.
    Chu, S.
    Cislak, A.
    Colarusso, M.
    Colloff, M. F.
    Dellapaolera, K. S.
    Delvenne, J.-F.
    Di Domenico, A.
    Drummond, A.
    Echterhoff, G.
    Edlund, J. E.
    Eggleston, C. M.
    Fairfield, B.
    Franco, G.
    Gabbert, F.
    Gamblin, B. W.
    Garry, M.
    Gentry, R.
    Gilbert, E. A.
    Greenberg, D. L.
    Halberstadt, J.
    Hall, L.
    Hancock, P. J. B.
    Hirsch, D.
    Holt, G.
    Jackson, J. C.
    Jong, J.
    Kehn, A.
    Koch, C.
    Kopietz, R.
    Koerner, U.
    Kunar, M. A.
    Lai, C. K.
    Langton, S. R. H.
    Leite, F. P.
    Mammarella, N.
    Marsh, J. E.
    McConnaughy, K. A.
    [J]. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2014, 9 (05) : 556 - 578
  • [2] There's More Than One Way to Conduct a Replication Study: Beyond Statistical Significance
    Anderson, Samantha F.
    Maxwell, Scott E.
    [J]. PSYCHOLOGICAL METHODS, 2016, 21 (01) : 1 - 12
  • [3] The Rules of the Game Called Psychological Science
    Bakker, Marjan
    van Dijk, Annette
    Wicherts, Jelte M.
    [J]. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2012, 7 (06) : 543 - 554
  • [4] Emergence of Large Treatment Effects From Small Trials
    Batterham, Alan M.
    Hopkins, William G.
    [J]. JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2013, 309 (08): : 768 - 768
  • [5] Accuracy of effect size estimates from published psychological research
    Brand, Andrew
    Bradley, Michael T.
    Best, Lisa A.
    Stoica, George
    [J]. PERCEPTUAL AND MOTOR SKILLS, 2008, 106 (02) : 645 - 649
  • [6] The Replication Recipe: What makes for a convincing replication?
    Brandt, Mark J.
    IJzerman, Hans
    Dijksterhuis, Ap
    Farach, Frank J.
    Geller, Jason
    Giner-Sorolla, Roger
    Grange, James A.
    Perugini, Marco
    Spies, Jeffrey R.
    van 't Veer, Anna
    [J]. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY, 2014, 50 : 217 - 224
  • [7] Power failure: why small sample size undermines the reliability of neuroscience
    Button, Katherine S.
    Ioannidis, John P. A.
    Mokrysz, Claire
    Nosek, Brian A.
    Flint, Jonathan
    Robinson, Emma S. J.
    Munafo, Marcus R.
    [J]. NATURE REVIEWS NEUROSCIENCE, 2013, 14 (05) : 365 - 376
  • [8] Cohen J, 2013, Statistical power analysis for the behavioral sciences, DOI [10.4324/9780203771587, DOI 10.4324/9780203771587]
  • [9] Three conditions under which experiments and observational studies produce comparable causal estimates: New findings from within-study comparisons
    Cook, Thomas D.
    Shadish, William R.
    Wong, Vivian C.
    [J]. JOURNAL OF POLICY ANALYSIS AND MANAGEMENT, 2008, 27 (04) : 724 - 750
  • [10] The perils with the misuse of predictive power
    Dallow, Nigel
    Fina, Paolo
    [J]. PHARMACEUTICAL STATISTICS, 2011, 10 (04) : 311 - 317