2018
DOI: 10.7287/peerj.preprints.3411v2
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Manipulating the alpha level cannot cure significance testing

Abstract: We argue that making accept/reject decisions on scientific hypotheses, including a recent call for changing the canonical alpha level from p= .05 to .005, is deleterious for the finding of new discoveries and the progress of science. Given that blanket and variable alpha levels both are problematic, it is sensible to dispense with significance testing altogether. There are alternatives that address study design and sample size much more directly than significance testing does; but none of the statistical tools… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 13 publications
(23 citation statements)
references
References 0 publications
0
23
0
Order By: Relevance
“…Still, it makes sense to ask if contributions from the statistics community have been helpful. The dust is perhaps still settling, but, based on the expert commentary supplementing the ASA statement, the follow-up by Ionides et al (2017), the recent proposal of Benjamin et al (2017) to simply change the default "p < 0.05" cutoff to "p < 0.005," and the plethora of criticism that followed (e.g., Crane 2018b; Lakens et al 2018;McShane et al 2018;Trafimow et al 2018), it seems that statisticians' views are as diverse as ever.…”
Section: Beyond P and Pmentioning
confidence: 99%
“…Still, it makes sense to ask if contributions from the statistics community have been helpful. The dust is perhaps still settling, but, based on the expert commentary supplementing the ASA statement, the follow-up by Ionides et al (2017), the recent proposal of Benjamin et al (2017) to simply change the default "p < 0.05" cutoff to "p < 0.005," and the plethora of criticism that followed (e.g., Crane 2018b; Lakens et al 2018;McShane et al 2018;Trafimow et al 2018), it seems that statisticians' views are as diverse as ever.…”
Section: Beyond P and Pmentioning
confidence: 99%
“…The American Economic Review and Econometrica furthermore request authors to explicitly report effect sizes and display standard errors (or even confidence intervals) instead of p-values in results tables, but they do not explicitly ban p-values. This is consistent with a suggestion put forward by many critical voices in the recent debateto demote p-values from their pedestal and consider them as a tool amongst many that help make appropriate inferences (cf., e.g., Amrhein et al 2017;McShane et al 2017;Trafimow et al 2018). The fact that some leading economics journals, which are widely considered as beacons for best practice, have initiated modest but sensible changes with regard to the use of the p-value is a promising signal.…”
Section: Reforms Under Way and Outlookmentioning
confidence: 99%
“…Dichotomization in conjunction with misleading terminology propagate cognitive biases that seduce researchers to make logically inconsistent and overconfident inferences, both when p is below and when it is above the "significance" threshold. The following errors seem to be particularly widespread: 1 1) use of p-values when there is neither random sampling nor randomization 2) confusion of statistical and practical significance or complete neglect of effect size 3) unwarranted binary statements of there being an effect as opposed to no effect, coming along with -misinterpretations of p-values below 0.05 as posterior probabilities of the null hypothesis -mixing up of estimating and testing and misinterpretation of "significant" results as evidence confirming the coefficients/effect sizes estimated from a single sample treatment of "statistically non-significant" effects as being zero (confirmation of the null) 4) inflation of evidence caused by unconsidered multiple comparisons and p-hacking 5) inflation of effect sizes caused by considering "significant" results only 1 See, for example, McCloskey and Ziliak (1996), Sellke et al (2001), Ioannidis (2005), Ziliak and McCloskey (2008), Krämer (2011), Ioannidis and Doucouliagos (2013), Kline (2013), Colquhoun (2014), Gelman and Loken (2014), Motulsky (2014), Vogt et al (2014), Gigerenzer and Marewski (2015), Greenland et al (2016), Hirschauer et al (2016;2018), Wasserstein and Lazar (2016), Ziliak (2016), Amrhein et al (2017), and Trafimow et al (2018). This list contains but a small selection of the literature on p-value misconceptions from the last 20 years.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations