Neuroimaging faces the daunting challenge of multiple testing -an instance of multiplicity -that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approach of massively univariate modeling. In dealing with multiplicity, the general strategy employed in the field is the same regardless of the specifics: trust the local "unbiased" effect estimates while adjusting the extent of statistical evidence at the global level. However, in this approach, modeling efficiency is compromised because each spatial unit (e.g., voxel, region, matrix element) is treated as an isolated and independent entity during massively univariate modeling. In addition, the required step of multiple testing "correction" by taking into consideration spatial relatedness, or neighborhood leverage, Global calibration is accomplished via a Gaussian distribution for the cross-region effects whose properties are not a priori specified, but a posteriori determined by the data at hand through the BML model. Our framework therefore incorporates multiplicity as integral to the modeling structure, not a separate correction step. By turning multiplicity into a strength, we aim to achieve five goals: 1) improve model efficiency with higher predictive accuracy, 2) control the error of incorrect magnitude and incorrect sign, 3) validate each model relative to competing candidates, 4) reduce the reliance and sensitivity on the choice of data space, and 5) encourage full results reporting. Our modeling proposal reverberates with recent proposals to eliminate the dichotomization of statistical evidence ("significant" vs. "non-significant"), to improve the interpretability of study findings, as well as to promote reporting the full gamut of results (not only "significant" ones), thereby enhancing research transparency and reproducibility. * Corresponding author. E-mail address: gangchen@mail.nih.gov 1 The two steps discussed here should not be confused with the "two steps" of first estimating regression coefficients at the individual-participant level and then employing those at a second step of statistical testing across participants.to rethink her approach of controlling for the overall false positive rate (FPR) under the null hypothesis significance testing (NHST) framework. One is not simply worried about errors at a single location but at all locations of the dataset at the same time -hence the need for a second step. This multiple testing problem has led to beautiful and creative solutions by the neuroimaging community in the past quarter of century, including random field theory (Worsley et al., 1992), Monte Carlo simulations (Forman et al., 1995), and permutation testing (Nichols and Holmes, 2001;Smith and Nichols, 2009).However, we believe that the current approaches have an important shortcoming: excessive penalties from correction for multiplicity due to modeling inefficiency.Consider the two-step procedure. In t...