2012
DOI: 10.2139/ssrn.2089580
|View full text |Cite
|
Sign up to set email alerts
|

Star Wars: The Empirics Strike Back

Abstract: jected tests. Our interpretation is that researchers might be tempted to inflate the value of those just-rejected tests by choosing a "significant" specification. We propose a method to measure this residual and describe how it varies by article and author characteristics.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
57
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 48 publications
(58 citation statements)
references
References 31 publications
(17 reference statements)
0
57
0
Order By: Relevance
“…Our first yardstick will be the number of variables which are 'robust', those where the posterior inclusion probability exceeds the prior inclusion probability; These numbers are reported in Table 4. 12 We report two sets of results, one based on the 10 This perspective acknowledges that, in the growth setting, the true data generating process is unlikely to be among the models considered, even when the number of candidate models is very large. This raises the question of whether AIC-based methods for model averaging are worth exploring; see Anderson (2002, 2004) and Claeskens and Hjort (2008).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our first yardstick will be the number of variables which are 'robust', those where the posterior inclusion probability exceeds the prior inclusion probability; These numbers are reported in Table 4. 12 We report two sets of results, one based on the 10 This perspective acknowledges that, in the growth setting, the true data generating process is unlikely to be among the models considered, even when the number of candidate models is very large. This raises the question of whether AIC-based methods for model averaging are worth exploring; see Anderson (2002, 2004) and Claeskens and Hjort (2008).…”
Section: Resultsmentioning
confidence: 99%
“…11 In much the same way, panel data estimation of the Solow model with country fixed effects (for example, Islam 1995) uses within-country variation to identify the same structural parameters as the cross-section study of Mankiw et al (1992). 12 These and the other BMA results in the paper were obtained using the BMS software due to Feldkircher and Zeugner. Taken together, these results point clearly in one direction.…”
Section: Resultsmentioning
confidence: 99%
“…These include selectively removing outliers and trying different testing procedures until results are significant (Abelson, 1995, p.55). Such practices go by various names such as p-hacking, torturing data, data dredging, or researcher degrees of freedom (Nuzzo, 2014;Lakens et al, 2014;Simmons et al, 2011;Brodeur et al, 2012;Gelman and Loken, 2013). They differ from the legitimate practice of exploratory data analysis (Tukey, 1980) because their goal is to obtain the results one wishes for, not to learn or to inform.…”
Section: Issues Regarding P-hackingmentioning
confidence: 99%
“…Second, biases can enter into the research process. For example, specification searching or publication bias can lead to an over-representation of positive effects (Brodeur et al, 2016;Vivalt, 2017a), while the "file-drawer problem" means that non-significant results often fail to see the light of day (Rosenthal, 1979). The third reason for low reproducibility -and the focus of this paperis statistical in nature.…”
Section: Introductionmentioning
confidence: 97%