Abstract:An alternative to null hypothesis significance testing is presented and discussed. This approach, referred to as observation-oriented modeling, is centered on model building in an effort to explicate the structures and processes believed to generate a set of observations. In terms of analysis, this novel approach complements traditional methods based on means, variances, and covariances with methods of pattern detection and analysis. Using data from a previously published study by Shoda et al., the basic tenet… Show more
“…By contrast, even the simple integrated models presented above provided the means for differentiating between various cognitive processes and distinct types of causes. Additional examples of such models can be found in the work of Grice and his colleagues [ 11 , 46 , 47 , 48 ], Powers [ 49 , 50 ], and Cevasco and Marmolejo-Ramos [ 51 ].…”
Four data sets from studies included in the Reproducibility Project were re-analyzed to demonstrate a number of flawed research practices (i.e., “bad habits”) of modern psychology. Three of the four studies were successfully replicated, but re-analysis showed that in one study most of the participants responded in a manner inconsistent with the researchers’ theoretical model. In the second study, the replicated effect was shown to be an experimental confound, and in the third study the replicated statistical effect was shown to be entirely trivial. The fourth study was an unsuccessful replication, yet re-analysis of the data showed that questioning the common assumptions of modern psychological measurement can lead to novel techniques of data analysis and potentially interesting findings missed by traditional methods of analysis. Considered together, these new analyses show that while it is true replication is a key feature of science, causal inference, modeling, and measurement are equally important and perhaps more fundamental to obtaining truly scientific knowledge of the natural world. It would therefore be prudent for psychologists to confront the limitations and flaws in their current analytical methods and research practices.
“…By contrast, even the simple integrated models presented above provided the means for differentiating between various cognitive processes and distinct types of causes. Additional examples of such models can be found in the work of Grice and his colleagues [ 11 , 46 , 47 , 48 ], Powers [ 49 , 50 ], and Cevasco and Marmolejo-Ramos [ 51 ].…”
Four data sets from studies included in the Reproducibility Project were re-analyzed to demonstrate a number of flawed research practices (i.e., “bad habits”) of modern psychology. Three of the four studies were successfully replicated, but re-analysis showed that in one study most of the participants responded in a manner inconsistent with the researchers’ theoretical model. In the second study, the replicated effect was shown to be an experimental confound, and in the third study the replicated statistical effect was shown to be entirely trivial. The fourth study was an unsuccessful replication, yet re-analysis of the data showed that questioning the common assumptions of modern psychological measurement can lead to novel techniques of data analysis and potentially interesting findings missed by traditional methods of analysis. Considered together, these new analyses show that while it is true replication is a key feature of science, causal inference, modeling, and measurement are equally important and perhaps more fundamental to obtaining truly scientific knowledge of the natural world. It would therefore be prudent for psychologists to confront the limitations and flaws in their current analytical methods and research practices.
“…This proposal is far from being new (e.g., see already Lamiell, 1981 , 2013 , 2014 ). However, there is now an increasing awareness of the necessity to study individuals repeatedly over time in order to adequately describe, explain, and predict the psychological processes underlying behavior (e.g., Roe, 2008 , 2014 ; Grice, 2015 ; Grice et al, in press ). This, together with the availability of new technology (e.g., apps and mobile devices) and statistical advances that enable researchers to collect and model extensive repeated measurement data more efficiently and effectively will allow researchers to make greater progress.…”
Section: Approaches To Between- and Within-person Integrationmentioning
“…To complement the hypothesis tests with linear mixed models, we also rely on observation-oriented modeling. This framework describes a novel perspective on data analysis that goes beyond traditional statistics by focusing on the description, detection, and interpretation of theoretically relevant patterns in the raw data (Grice et al, 2012(Grice et al, , 2017Grice, 2011). Instead of drawing statistical inferences based on hypothesis testing within standard models such as ANOVA or regression modeling, observation-oriented modeling focuses on descriptive measures such as Percent Classification Correct (PCC) which is defined as the relative frequency of observations conforming to a specific pattern predicted by a hypothesis.…”
Section: Robustness Analysis Using Observation-oriented Modelingmentioning
confidence: 99%
“…Moreover, it has a direct intuitive interpretation because it is always between 0% and 100% (with larger values indicating that the data better fit a predicted pattern) and can be used as an effect-size measure tailored to a specific hypothesis. Methods of observation-oriented modeling offer the additional benefit of not making distributional assumptions and are thus robust against outliers (Grice et al, 2012(Grice et al, , 2017. Given that observation-oriented modeling is robust to outliers, this allows us to test the robustness of our results by computing the PCC score twice using (1) the data excluding outliers similar as above and (2) the complete data including all outliers.…”
Section: Robustness Analysis Using Observation-oriented Modelingmentioning
On the internet, people often collaborate to generate extensive knowledge bases such as Wikipedia for semantic information or OpenStreetMap for geographic information. When contributing to such online projects, individual judgments follow a sequential process in which one contributor creates an entry and other contributors have the possibility to modify, extend, and correct the entry by making incremental changes. We refer to this way of working together as sequential collaboration because it is characterized by dependent judgments that are based on the latest judgment available. Since the process of correcting each other in sequential collaboration has not yet been studied systematically, we compare the accuracy of sequential collaboration and wisdom of crowds, the aggregation of a set of independent judgments. In three experiments with groups of four or six individuals, accuracy for answering general knowledge questions increased within sequences of judgments in which participants had the possibility to correct the judgment of the previous contributor. Moreover, the final individual judgments in sequential collaboration were slightly more accurate than the averaged judgments in wisdom of crowds. This shows that collaboration can benefit from the dependency of individual judgments, thus explaining why large collaborative online projects often provide data of high quality.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.