2018
DOI: 10.1177/0081175018777988
|View full text |Cite
|
Sign up to set email alerts
|

We Ran 9 Billion Regressions: Eliminating False Positives through Computational Model Robustness

Abstract: False positive findings are a growing problem in many research literatures. We argue that excessive false positives often stem from model uncertainty. There are many plausible ways of specifying a regression model, but researchers typically report only a few preferred estimates. This raises the concern that such research reveals only a small fraction of the possible results and may easily lead to nonrobust, false positive conclusions. It is often unclear how much the results are driven by model specification a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
59
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 59 publications
(59 citation statements)
references
References 55 publications
0
59
0
Order By: Relevance
“…for example, leading to omitted variable bias in estimation. Similarly, with many inputs, one generally runs the risk of model misspecification (Belloni et al, 2014;Ho et al, 2007;King & Nielsen, 2016;Raftery, 1995;Young & Holsteen, 2017;Muñoz & Young, 2018). Athey & Imbens (2015) develop a measure of sensitivity to misspecification.…”
Section: Causal Inferencementioning
confidence: 99%
“…for example, leading to omitted variable bias in estimation. Similarly, with many inputs, one generally runs the risk of model misspecification (Belloni et al, 2014;Ho et al, 2007;King & Nielsen, 2016;Raftery, 1995;Young & Holsteen, 2017;Muñoz & Young, 2018). Athey & Imbens (2015) develop a measure of sensitivity to misspecification.…”
Section: Causal Inferencementioning
confidence: 99%
“…Common examples of this strategy include extreme bounds analysis in econometrics [80], multimodel ensembles [34] in hydrology and climatology and sensitivity analyses, which are used across many disciplines. More recent approaches to report uncertainty include the "vibration of effects" approach [28], "specification curve analysis" [29], "multiverse analysis" [30] "multimodel analysis" [32] and "compu-tational robustness analysis" [33], as discussed previously. Silberzahn et al [20] go a step further and propose the reporting of the results of different teams of researchers analyzing the same research question on the same data set.…”
Section: Report Uncertaintymentioning
confidence: 99%
“…There are for instance a number of recently proposed approaches which assess the robustness of research findings to alternative analytical pathways by reporting the results of a large number of analysis strategies: the "vibration of effects" approach in epidemiology [28], "specification curve analysis" [29] and "multiverse analysis" in psychology [30], a "measure of robustness to misspecification" in economics [31] or "multimodel analysis" [32] and "computational robustness analysis" [33] in sociology. In other disciplines, including climatology, ecology and risk analysis, there is a long-standing tradition of addressing the robustness to alternative analysis strategies through sensitivity analyses, multimodel ensembles [34] and Bayesian model averaging [35,36].…”
mentioning
confidence: 99%
“…Even using the same data and same models we may observe researcher variability. Although some argue for the utility of millions of models (Muñoz and Young 2018), in crowdsourced research the set of models are not a simulation but a set of real models derived from practicing social scientists. No matter how strong the statistical methods, data itself cannot produce the underlying causal model, this requires rational construction of theory, exclusion restrictions, attention to confounding and counterfactuals (Pearl 2010) 5 .…”
Section: The Crisis Of Social Science and Macro-comparative Secondary Dmentioning
confidence: 99%