2014
DOI: 10.1073/pnas.1323051111
|View full text |Cite
|
Sign up to set email alerts
|

Reproducibility issues in science, is P value really the only answer?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(14 citation statements)
references
References 5 publications
0
14
0
Order By: Relevance
“…Cumulative evidence often builds up from several studies with larger p-values that only when combined show clear evidence against the null hypothesis (Greenland et al 2016, p. 343). Very possibly, more stringent thresholds would lead to even more results being left unpublished, enhancing publication bias (Gaudart et al 2014;Gelman & Robert 2014). What we call winner's curse, truth inflation or inflated effect sizes will become even more severe with more stringent thresholds (Button et al 2013b).…”
Section: 'We Need More Stringent Decision Rules'mentioning
confidence: 99%
“…Cumulative evidence often builds up from several studies with larger p-values that only when combined show clear evidence against the null hypothesis (Greenland et al 2016, p. 343). Very possibly, more stringent thresholds would lead to even more results being left unpublished, enhancing publication bias (Gaudart et al 2014;Gelman & Robert 2014). What we call winner's curse, truth inflation or inflated effect sizes will become even more severe with more stringent thresholds (Button et al 2013b).…”
Section: 'We Need More Stringent Decision Rules'mentioning
confidence: 99%
“…For many years, however, and more so recently, the use of the term "significant" for findings that cross this somewhat arbitrary p-threshold is pointed out as detrimental to research and reliable findings [47,59]. Some authors have suggested a more stringent alpha level-for example, .005 instead of the traditional .05 (e.g., References [6,44])-yet this suggestion leads others to be concerned about inflated Type II errors [33]. At any rate, using the word "significant" can be misleading, and some scholars recommend to refrain from using this term altogether [41].…”
Section: P-values Significance and Effect Sizesmentioning
confidence: 99%
“…Conversely, a researcher might obtain a result with low significance (high p-value) with a large effect size measure (or an important/practical result). There is a tendency to overvalue significance and to ignore effect size (see early comment in archaeology by Thomas, 1978:233; see a recent discussion relevant to science by Gaudart et al, 2014), which relates to singular focus on the role of the p-value in hypothesis testing (a multitude of examples could be cited here, but to do so would target case studies and authors; see Wolverton (2005) and Wolverton et al (2008) for self-critical examples). We return to effect size and its interpretation in the Analyse effect size to determine 'practical significance' section later.…”
Section: Statistical Hypothesis Testingmentioning
confidence: 99%