2012
DOI: 10.1002/pam.21660
|View full text |Cite
|
Sign up to set email alerts
|

External Validity in Policy Evaluations That Choose Sites Purposively

Abstract: Evaluations of the impact of social programs are often carried out in multiple “sites,” such as school districts, housing authorities, local TANF offices, or One-Stop Career Centers. Most evaluations select sites purposively following a process that is nonrandom. Unfortunately, purposive site selection can produce a sample of sites that is not representative of the population of interest for the program. In this paper, we propose a conceptual model of purposive site selection. We begin with the proposition tha… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
121
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
8
1
1

Relationship

4
6

Authors

Journals

citations
Cited by 105 publications
(124 citation statements)
references
References 19 publications
1
121
0
Order By: Relevance
“…IES also sponsors large-scale evaluations through contracts to major research firms (e.g., Abt Associates, Mathematica Policy Research, MDRC). These evaluations usually select a sample designed to cover all regions of the country, but sites are selected purposively to reduce costs and sometimes with other objectives in mind (e.g., to test the intervention in sites where it will produce the greatest “contrast” between the treatment and control conditions, suggesting that it may have the greatest impact); they are rarely selected randomly to be formally representative of any broader population of potential interest to policymakers (Olsen et al, 2013). …”
Section: Background and Overviewmentioning
confidence: 99%
“…IES also sponsors large-scale evaluations through contracts to major research firms (e.g., Abt Associates, Mathematica Policy Research, MDRC). These evaluations usually select a sample designed to cover all regions of the country, but sites are selected purposively to reduce costs and sometimes with other objectives in mind (e.g., to test the intervention in sites where it will produce the greatest “contrast” between the treatment and control conditions, suggesting that it may have the greatest impact); they are rarely selected randomly to be formally representative of any broader population of potential interest to policymakers (Olsen et al, 2013). …”
Section: Background and Overviewmentioning
confidence: 99%
“…In fact, experiments are frequently conducted in only one or two localities that cannot claim to be representative of the nation, state, or other jurisdiction for which policy is made. (See Olsen et al (2013) for a derivation of the bias that may occur when experiments are conducted with nonrepresentative populations. )…”
Section: Internal Validity and External Validity Of Experimental Estimentioning
confidence: 99%
“…The average treatment effect is typically estimated using the n units in the sample and a multilevel model that accounts for the study design (e.g., random block design or cluster randomized design; Raudenbush & Bryk, 2002), and the impact of the experiment is evaluated by comparing this estimate to its standard error. Imai, King, and Stuart (2008) and Olsen et al (2012) have shown that when the sample is not representative of the population and when site-specific treatment effects vary, the sample based estimate of is biased. Our goal here, therefore, is to develop a strategy for selecting the n units in the sample S so that the sample is compositionally similar to the N units in the inference population P, thereby leading to a less biased and more precise estimate of the population average treatment effect.…”
Section: Definitions Assumptions and Goalsmentioning
confidence: 97%