2022
DOI: 10.1007/s00362-022-01296-x
|View full text |Cite
|
Sign up to set email alerts
|

Variable selection in Propensity Score Adjustment to mitigate selection bias in online surveys

Abstract: The development of new survey data collection methods such as online surveys has been particularly advantageous for social studies in terms of reduced costs, immediacy and enhanced questionnaire possibilities. However, many such methods are strongly affected by selection bias, leading to unreliable estimates. Calibration and Propensity Score Adjustment (PSA) have been proposed as methods to remove selection bias in online nonprobability surveys. Calibration requires population totals to be known for the auxili… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 56 publications
0
4
0
Order By: Relevance
“…The inclusion of these response types could potentially be achieved through various types of hierarchical modelling to account for pseudo‐replication, propensity score weighting to calibrate for nonprobability (Ferri‐García & Rueda, 2022), or multilevel regressions with poststratification adjustments (Mercer et al., 2017). However, such techniques may only be viable with larger data sets, may require extensive knowledge of model assumptions, are imperfect in reducing bias and still risk inflating covariance‐based estimates even with small proportions of invalid respondents (Copas et al., 2020; Dever et al., 2008; Guo et al., 2020; King et al., 2018).…”
Section: Results From Suspicion Variable Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…The inclusion of these response types could potentially be achieved through various types of hierarchical modelling to account for pseudo‐replication, propensity score weighting to calibrate for nonprobability (Ferri‐García & Rueda, 2022), or multilevel regressions with poststratification adjustments (Mercer et al., 2017). However, such techniques may only be viable with larger data sets, may require extensive knowledge of model assumptions, are imperfect in reducing bias and still risk inflating covariance‐based estimates even with small proportions of invalid respondents (Copas et al., 2020; Dever et al., 2008; Guo et al., 2020; King et al., 2018).…”
Section: Results From Suspicion Variable Analysismentioning
confidence: 99%
“…Although the elimination of possible careless responders would result in the loss of some valuable information from our study, we deemed it appropriate to eliminate all possible fraudulent responses due to the relatively small sample size and implications that their inclusion could have on correlational statistics. Various statistical modelling techniques can assist researchers wanting to include careless responders or other fraudulent responses (Copas et al, 2020;Dever et al, 2008;Ferri-García & Rueda, 2022;Mercer et al, 2017); however, a similar level of concern should be granted to analyses considering the use of statistical adjustments. Furthermore, researchers who use these techniques should still report the presence of fraud in their surveys and the extent to which the modelling adjustments may differ with the exclusion of that potential fraud.…”
Section: Removing Survey Responsesmentioning
confidence: 99%
“…24 Groups were compared to assess significance of differences before and after propensity score based IPTW, using measure of associations with Cramer's Phi (V) for categorical variables and Rsquared for continuous variables. 25 Mann-Whitney tests were used for continuous variables, whereas Pearson χ 2 tests were used for categorical variables, including endpoints.…”
Section: Discussionmentioning
confidence: 99%
“…Therefore, although weighting within classes is a commonly used procedure for non-response cross-sectional and longitudinal weighting in panels, a more pragmatic alternative is to use a regression-based approach, all the more so when numerous auxiliary variables are available [18]. For this we are going to use the popular Propensity Score Adjustment (PSA) method [20,29,30] to model the probability that a unit k of the new theoretical sample s (j) responds to M j , where j = 1, ..., t , or that another unit k of the effective sample s (i) r responds to M j , where i = 1, ..., j − 1 , j = 2, ..., t , and i < j.…”
Section: Weight Adjustment Based On Propensitiesmentioning
confidence: 99%
“…• Learning rate ∈ [0.001, 0.9] : the weight shrinkage applied after each boosting step. • Maximum depth ∈ [1,30] : the maximum number of splits that each tree can contain. • Minimum child weight ∈ [0, 10] : the minimum total of instance weights needed to consider a new partition.…”
Section: Modelling Non-responsementioning
confidence: 99%