Invited Commentary: The Need for Cognitive Science in Methodology

被引:120
作者
Greenland, Sander [1 ,2 ]
机构
[1] Univ Calif Los Angeles, Fielding Sch Publ Hlth, Dept Epidemiol, Los Angeles, CA USA
[2] Univ Calif Los Angeles, Coll Letters & Sci, Dept Stat, Los Angeles, CA USA
关键词
behavioral economics; bias analysis; cognitive bias; motivated reasoning; nullism; overconfidence; sensitivity analysis; significance testing; P-VALUES; CONFIDENCE-INTERVALS; MISINTERPRETATION; INFERENCE; CAUSAL;
D O I
10.1093/aje/kwx259
中图分类号
R1 [预防医学、卫生学];
学科分类号
1004 ; 120402 ;
摘要
There is no complete solution for the problem of abuse of statistics, but methodological training needs to cover cognitive biases and other psychosocial factors affecting inferences. The present paper discusses 3 common cognitive distortions: 1) dichotomania, the compulsion to perceive quantities as dichotomous even when dichotomization is unnecessary and misleading, as in inferences based on whether a P value is "statistically significant"; 2) nullism, the tendency to privilege the hypothesis of no difference or no effect when there is no scientific basis for doing so, as when testing only the null hypothesis; and 3) statistical reification, treating hypothetical data distributions and statistical models as if they reflect known physical laws rather than speculative assumptions for thought experiments. As commonly misused, null-hypothesis significance testing combines these cognitive problems to produce highly distorted interpretation and reporting of study results. Interval estimation has so far proven to be an inadequate solution because it involves dichotomization, an avenue for nullism. Sensitivity and bias analyses have been proposed to address reproducibility problems (Am J Epidemiol. 2017;186(6):646-647); these methods can indeed address reification, but they can also introduce new distortions via misleading specifications for bias parameters. P values can be reframed to lessen distortions by presenting them without reference to a cutoff, providing them for relevant alternatives to the null, and recognizing their dependence on all assumptions used in their computation; they nonetheless require rescaling for measuring evidence. I conclude that methodological development and training should go beyond coverage of mechanistic biases (e.g., confounding, selection bias, measurement error) to cover distortions of conclusions produced by statistical methods and psychosocial forces.
引用
收藏
页码:639 / 645
页数:7
相关论文
共 63 条
[1]   The earth is flat (p > 0.05): significance thresholds and the crisis of unreplicable research [J].
Amrhein, Valentin ;
Korner-Nievergelt, Franzi ;
Roth, Tobias .
PEERJ, 2017, 5
[2]  
[Anonymous], 2008, Modern Epidemiology
[3]  
[Anonymous], 1987, STAT SCI, DOI DOI 10.1214/SS/1177013243
[4]  
[Anonymous], 2007, Thinking and Deciding
[5]  
[Anonymous], 2004, CONT DEBATES PHILOS
[6]  
Birnbaum A., 1961, ANN MATH STAT, V32, P112
[7]  
Broman K., 2017, Recommendations to Funding Agencies for Supporting Reproducible Research, V2, P1
[8]   RECONCILING BAYESIAN AND FREQUENTIST EVIDENCE IN THE ONE-SIDED TESTING PROBLEM [J].
CASELLA, G ;
BERGER, RL .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1987, 82 (397) :106-111
[9]   SOME PROBLEMS CONNECTED WITH STATISTICAL-INFERENCE [J].
COX, DR .
ANNALS OF MATHEMATICAL STATISTICS, 1958, 29 (02) :357-372
[10]   The Long Way From α-Error Control to Validity Proper: Problems With a Short-Sighted False-Positive Debate [J].
Fiedler, Klaus ;
Kutzner, Florian ;
Krueger, Joachim I. .
PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2012, 7 (06) :661-669