Given the growing popularity of nonprobability samples as a cost- and time-efficient alternative to probability sampling, a variety of adjustment approaches have been proposed to correct for self-selection bias in nonrandom samples. Popular methods such as inverse propensity-score weighting (IPSW) and propensity-score (PS) adjustment by subclassification (PSAS) utilize a probability sample as a reference to estimate pseudo-weights for the nonprobability sample based on PSs. A recent contribution, kernel weighting (KW), has been shown to be able to improve over IPSW and PSAS with respect to mean squared error. However, the effectiveness of these methods for reducing bias critically depends on the ability of the underlying propensity model to reflect the true (self-)selection process, which is a challenging task with parametric regression. In this study, we propose a set of pseudo-weights construction methods, KW-ML, utilizing both machine learning (ML) methods (to estimate PSs) and KW (to construct pseudo-weights based on the ML-estimated PSs), which provides added flexibility over logistic regression-based methods. We compare the proposed KW-ML pseudo-weights that are based on model-based recursive partitioning, conditional random forests, gradient tree boosting, and model-based boosting, with KW pseudo-weights based on parametric logistic regression in population mean estimation via simulations and a real data example. Our results indicate that particularly boosting methods represent promising alternatives to logistic regression and result in KW estimates with lower bias in a variety of settings, without increasing variance.
Reconstructing spectral functions from Euclidean Green's functions is an important inverse problem in physics. The prior knowledge for specific physical systems routinely offers essential regularization schemes for solving the ill-posed problem approximately. Aiming at this point, we propose an automatic differentiation framework as a generic tool for the reconstruction from observable data. We represent the spectra by neural networks and set chi-square as loss function to optimize the parameters with backward automatic differentiation unsupervisedly. In the training process, there is no explicit physical prior embedding into neural networks except the positive-definite form. The reconstruction accuracy is assessed through Kullback-Leibler(KL) divergence and mean square error(MSE) at multiple noise levels. It should be noted that the automatic differential framework and the freedom of introducing regularization are inherent advantages of the present approach and may lead to improvements of solving inverse problem in the future.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.