Documentos de Académico
Documentos de Profesional
Documentos de Cultura
Institute for
Public Policy
110 Fifth Avenue Southeast, Suite 214 • PO Box 40999 • Olympia, WA 98504-0999 • (360) 586-2677 • FAX (360) 586-2793 • www.wsipp.wa.gov
September 2006
2
questions, but for this review, we are focused on Exhibit 2
the question: What works to improve academic Initial List of K–12 Educational Programs,
outcomes? Policies, and Services Included in the
Literature Review
The programs, services, and policies included in
the Institute’s initial literature review are listed in
Exhibit 2. Over time, this list of topics can be 9 Alternative learning environments (such as
expanded to meet Washington’s policy distance learning and alternative schools)
information needs. The legislative direction to the 9 Block scheduling
Institute was to “begin the development of a 9 Career and technical education
repository of research and evaluations” on what 9 Charter schools
works—the programs, services, and policies listed 9 Class size
in Exhibit 2 is our initial list for the March 2007
9 Comprehensive school reform
report.
9 Dropout prevention
The specific types of outcomes that we will meta- 9 English language learner instruction
analyze depends on the measures used in 9 Extended learning options (including summer
existing K–12 evaluation studies. Again, the school, before/after school and Saturday
focus is on programs, services, and policies that programs, and a longer school day)
aim to improve students’ academic outcomes. 9 Full-day kindergarten
These outcome measures include, but are not 9 Gifted/talented student programs
limited to, the following: 9 Instructional curricula
3
To do this, we have developed, and are Appendix A: Meta-Analysis Techniques
continuing to refine, techniques to measure costs
and benefits associated with the programs, This technical appendix describes the study coding
policies, and services listed in Exhibit 2. We will criteria, procedures for calculating effect sizes, and
identify programs whose benefits outweigh their adjustments to effect sizes that will be used in the
costs and also estimate the costs and benefits of Institute’s meta-analysis of K–12 educational programs
and services.
programs that do not break even. In
systematically examining the state’s investment in
K–12 education, it is just as important to know Meta-Analysis Coding Criteria
which strategies do not produce positive returns
for the taxpayer’s dollar as it is to identify those The following are key coding criteria for our meta-analysis
with proven positive returns. of evaluations of K–12 educational programs and services.
As in our previous cost-benefit analyses, we will 1. Study Search and Identification Procedures. We
estimate costs and benefits in two ways: first, we are currently searching for all K–12 evaluation studies
estimate the benefits that accrue directly to completed since 1970 that are written in English. We
are using three primary sources: a) study lists in other
program participants (in this case, the students),
reviews of the K–12 research literature; b) citations in
and second, we estimate the benefits that accrue individual evaluation studies; and c) research
to non-participants. For example, a student who databases/search engines such as Google, Proquest,
graduates from high school enjoys the benefit of Ebsco, ERIC, and SAGE.
greater earning potential compared with students
2. Peer-Reviewed and Other Studies. Many K–12
who do not graduate. Non-participants benefit
evaluation studies are published in peer-reviewed
from the higher taxes paid based on those academic journals, while others are from
increased earnings. Additionally, evidence exists government or other reports. It is important to
for a causal link between high school graduation include non-peer reviewed studies, because it has
and subsequent reduced crime.7 Examining costs been suggested that peer-reviewed publications
and benefits from both the participant and may be biased toward positive program effects.
taxpayer perspectives provides a more Therefore, our meta-analysis will include studies
comprehensive description of cost-effectiveness. regardless of their source.
3. Control and Comparison Group Studies. We
The methodology for estimating costs and only include studies in our analysis if they have a
benefits associated with educational outcomes of control or comparison group. We do not include
K–12 programs and services is described in studies with a single-group, pre-post research
Appendix B. design. We believe that it is only through rigorous
comparison group studies that average treatment
effects can be reliably estimated.8
4. Random Assignment and Quasi-Experiments.
Random assignment studies are preferred for
inclusion in our review, but we also include studies
with non-randomly assigned groups if sufficient
information is provided to demonstrate
comparability between the treatment and
comparison groups.
5. Exclusion of Studies of Program Completers
Only. We do not include a study in our meta-
analytic review if the treatment group is made up
solely of program completers. We believe there
are too many significant unobserved self-selection
factors that distinguish a program completer from a
8
See: Identifying and implementing education practices
supported by rigorous evidence: A user friendly guide (2003,
December) Coalition for Evidence-Based Policy, U.S.
Department of Education, Institute of Education Sciences,
National Center for Education Evaluation and Regional
7
L. Lochner and E. Moretti (2004) The effect of education on Assistance. Available at:
crime: Evidence from prison inmates, arrests, and self- http://www.evidencebasedpolicy.org/docs/Identifying_and_
reports. American Economic Review 94(1): 155-189. Implementing_Educational_Practices.pdf
4
program dropout, and these factors are likely to meta-analytic reviews we have completed,
bias estimated treatment effects. We do, however, however, we anticipate that some studies will
retain such a study if sufficient information is report some, but not all, of the information required.
provided to allow us to reconstruct an intent-to- The rules we follow for these situations are as
treat group that includes both completers and non- follows:
completers, or if the demonstrated rate of program
a. Two-Tail P-Values. Sometimes, studies only
non-completion is very small (e.g. under 10
report p-values for significance testing of
percent).
program outcomes. If the study reports a one-
6. Enough information to Calculate an Effect Size. tail p-value, we will convert it to a two-tail test.
Following the statistical procedures in Lipsey and
b. Declaration of Significance by Category.
Wilson,9 a study must provide the necessary
Some studies report results of statistical
statistical information to calculate an effect size. If
significance tests in terms of categories of p-
such information is not provided, the study cannot be
values, such as p<=.01, p<=.05, or “not
included in our review.
significant at the p=.05 level.” We calculate
7. Mean Difference Effect Sizes. For this study we effect sizes in these cases by using the highest
are coding mean difference effect sizes following p-value in the category; e.g., if a study reports
the procedures in Lipsey and Wilson (2001). For significance at “p<=.05,” we calculate the effect
dichotomous measures, we use the arcsine size at p=.05. This is the most conservative
transformation to approximate the mean difference strategy. If the study simply states a result was
effect size, again following Lipsey and Wilson. “not significant,” we compute the effect size
assuming a p-value of .50 (i.e. p=.50).
8. Unit of Analysis. Our unit of analysis is an
independent test of treatment at a particular site.
Some studies report outcome evaluation
information for multiple sites; we include each site
Procedures for Calculating Effect Sizes
as an independent observation if a unique
Effect sizes measure the degree to which a program has
comparison group is also used at each site.
been shown to change an outcome for program
9. Multivariate Results Preferred. Some studies participants relative to a comparison group. There are
present two types of analyses: raw outcomes that several methods used by meta-analysts to calculate
are not adjusted for covariates, such as family effect sizes, as described in Lipsey and Wilson (2001).
income and ethnicity; and those that are adjusted In this meta-analysis, we are using statistical procedures
with multivariate statistical methods. In these to calculate the mean difference effect sizes of programs.
situations, we code the multivariate outcomes. We are not using the odds-ratio effect size because
many outcomes measured in K–12 program evaluations
10. Dichotomous Measures Preferred Over
are continuously measured. Thus, the mean difference
Continuous Measures. Some studies include two
effect size is a natural choice.
types of measures for the same outcome: a
dichotomous (yes/no) outcome and a continuous
Many of the outcomes we are analyzing, however, are
(mean number) measure. In these situations, we
measured as dichotomies. For these yes/no outcomes,
code an effect size for the dichotomous measure,
Lipsey and Wilson (2001) show that the mean difference
because in small sample studies, continuous
effect size calculation can be approximated using the
measures of education outcomes can be unduly
arcsine transformation of the difference between
influenced by a small number of outliers, while
proportions.10
dichotomous measures can avoid this problem. Of
course, if a study only presents a continuous
measure, then we will code that measure. (A1) ESm( p ) = 2 × arcsin Pe − 2 × arcsin Pc
12. Some Special Coding Rules for Effect Sizes. A second effect size calculation involves continuous
Most studies that meet the criteria for inclusion in data where the differences are in the means of an
our review will have sufficient information to code outcome. When an evaluation reports this type of
exact mean difference effect sizes. Based on other
9
M. Lipsey and D. Wilson (2001). Practical meta-analysis.
10
Thousand Oaks, CA: Sage. Ibid., Table B10, formula (22).
5
information, we use the standard mean difference In equation (A4), ne and nc are the number of
effect size statistic.11 participants in the experimental and control groups and
ES'm is from equation (A3).
Me − Mc
(A2) ESm =
Next, the inverse variance weight wm is computed for
SDe2 + SDc2
each mean effect size with:15
2
1
In this formula, ESm is the estimated effect size for the (A5) wm =
SEm2
difference between means from the research information;
Me is the mean number of an outcome for the
experimental group; Mc is the mean number of an The weighted mean effect size for a group of studies in
outcome for the control group; SDe is the standard program area i is then computed with:16
deviation of the mean number for the experimental
group; and SDc is the standard deviation of the mean (A6) ES = ∑ (w ES ′
mi mi )
number for the control group. ∑w mi
Often, research studies report the mean values needed Confidence intervals around this mean are then
to compute ESm in (A2) but fail to report the standard computed by first calculating the standard error of the
deviations. Sometimes, however, the research will mean with:17
report information about statistical tests or confidence
intervals that can then allow the pooled standard
(A7) 1
deviation to be estimated. These procedures are also SEES =
described in Lipsey and Wilson (2001). ∑ wmi
Adjusting Effect Sizes for Small Sample Sizes. Since Next, the lower, ESL, and upper limits, ESU, of the
some studies have very small sample sizes, we follow confidence interval are computed with:18
the recommendation of many meta-analysts and adjust
for this. Small sample sizes have been shown to
upwardly bias effect sizes, especially when samples are (A8) ES L = ES − z(1−α ) ( SE ES )
less than 20. Following Hedges,12 Lipsey and Wilson
(2001)13 report the “Hedges correction factor,” which we (A9) ESU = ES + z(1−α ) ( SE ES )
use to adjust all mean difference effect sizes (N is the
total sample size of the combined treatment and
comparison groups): In equations (A8) and (A9), z(1-α) is the critical value for
the z-distribution (1.96 for α = .05).
(A3)
⎡
ES′m = ⎢1 −
3 ⎤
[
⎥ × ES m , or , ES m ( p ) ] The test for homogeneity, which provides a measure of
⎣ 4N − 9 ⎦
the dispersion of the effect sizes around their mean, is
given by:19
Computing Weighted Average Effect Sizes,
Confidence Intervals, and Homogeneity Tests. Once
(∑ wi ESi ) 2
effect sizes are calculated for each program effect, the (A10) Qi = (∑ wi ESi2 ) −
individual measures are summed to produce a weighted ∑ wi
average effect size for a program area. We calculate the
inverse variance weight for each program effect, and The Q-test is distributed as a chi-square with k-1 degrees of
these weights are used to compute the average. These freedom (where k is the number of effect sizes).
calculations involve three steps. First, the standard
error, SEm of each mean effect size is computed with:14 Computing Random Effects Weighted Average Effect
Sizes and Confidence Intervals. When the p-value on
ne + nc ( ESm′ )2 the Q-test indicates significance at values of p less than
(A4) SEm = + or equal to .05, a random effects model is performed to
ne nc 2( ne + nc )
calculate the weighted average effect size. This is
accomplished by first calculating the random effects
variance component, v.20
11
Ibid., Table B10, formula (1).
12 15
L. Hedges (1981). Distribution theory for Glass’s estimator Ibid., 49, equation 3.24.
16
of effect size and related estimators. Journal of Educational Ibid., 114.
17
Statistics 6: 107-128. Ibid.
13 18
Lipsey and Wilson, Practical meta-analysis, 49, formula Ibid.
19
3.22. Ibid., 116.
14 20
Ibid., 49, equation 3.23. Ibid., 134.
6
Qi − (k − 1) • A “5” is assigned to an evaluation with well-
(A11) v=
∑ wi − (∑ wsqi ∑ wi ) implemented random assignment of subjects to a
treatment group and a control group that does not
This random variance factor is then added to the receive the treatment/program. A good random
variance of each effect size and then all inverse assignment study should also indicate how well
variance weights are recomputed, as are the other the random assignment actually occurred by
meta-analytic test statistics. reporting values for pre-existing characteristics of
the program and control groups.
• A “4” is assigned to a study that employs a
Adjustments to Effect Sizes rigorous quasi-experimental research design with
a program and matched comparison group,
In the Institute’s meta-analytic reviews, we make three controlling with statistical methods for self-
types of adjustments to effect sizes that we believe are selection bias that might otherwise influence
necessary to better estimate the results each program is outcomes. These quasi-experimental methods
likely to actually achieve in real-world settings. We make may include estimates made with an instrumental
adjustments for: variables modeling approach, or a Heckman
approach to modeling self-selection.22 A level 4
• Methodological quality; study may also be used to “downgrade” an
experimental random assignment design that had
• Relevance or quality of the outcome measure(s) problems in implementation, perhaps with
used; and significant attrition rates.
• Degree to which the researcher(s) who conducted • A “3” indicates a non-experimental evaluation
the study was invested in the program’s design where the program and comparison groups
and implementation. are reasonably well matched on pre-existing
differences in key variables. There must be
Methodological Quality. Not all research is of equal evidence presented in the evaluation that
quality, and this, we believe, greatly influences the indicates few, if any, significant differences
confidence that can be placed in study results. Some observed in these salient pre-existing
studies are well designed and implemented, and the variables. Alternatively, if an evaluation
results can be viewed as accurate representations of employs sound multivariate statistical
whether the program worked. Other studies are not techniques (e.g. logistic regression) to control
designed as well and less confidence can be placed in for pre-existing differences, then a study with
reported outcomes. In particular, studies of inferior some differences in pre-existing variables can
research design cannot completely control for sample qualify as a level 3.
selection bias or other unobserved threats to the validity
• A “2” involves a study with a program and
of reported research results. This does not mean that
matched comparison group where the two
results from these studies are of no value, but it does
groups lack comparability on pre-existing
mean that less confidence can be placed in any cause-
variables and no attempt was made to control
and-effect conclusions.
for these differences in the study.
To account for differences in the quality of research • A “1” involves a study where no comparison
designs, we use a 5-point scale to adjust the reported group is utilized. Instead, the relationship
results. The scale is based closely on the 5-point scale between a program and an outcome, e.g., grade
developed by researchers at the University of point average, is analyzed before and after the
Maryland.21 On this 5-point scale, a rating of “5” reflects program.
an evaluation in which the most confidence can be
placed. As the evaluation ranking gets lower, less We do not use the results from program evaluations rated
confidence can be placed in any reported differences (or as a “1” on this scale, because they do not include a
lack of differences) between the program and comparison group, and we believe that there is no context
comparison/control groups. to judge program effectiveness. We also regard
evaluations with a rating of “2” as highly problematic and,
On the 5-point scale, as interpreted by the Institute, each as a result, we do not consider their findings in the meta-
study is rated with the following numerical ratings. analysis. In this study, we only consider evaluations that
rate at least a 3 on this scale.
21 22
L. Sherman, D. Gottfredson, D. MacKenzie, J. Eck, For a discussion of these methods, see W. Rhodes,
P. Reuter, and S. Bushway (1998). Preventing crime: What B. Pelissier, G. Gaes, W. Saylor, S. Camp, and S. Wallace
works, what doesn't, what's promising. Prepared for the (2001). Alternative solutions to the problem of selection bias
National Institute of Justice. Department of Criminology and in an analysis of federal residential drug treatment programs.
Criminal Justice, University of Maryland. Chapter 2. Evaluation Review 25(3): 331-369.
7
An explicit adjustment factor is assigned to each effect Adjusting Effect Sizes for Researcher Involvement in
size based on the Institute’s judgment concerning Program Design and Implementation. The purpose of
research design quality. We believe this adjustment is the Institute’s work is to identify programs that can make
critical and is the only practical way to combine the results cost-beneficial improvements to Washington’s actual K–12
of a high quality study with those of lesser design quality. service delivery system. There is some evidence that
programs that are closely controlled by researchers or
• A level 5 study carries a factor of 1.0 (that is, program developers have better results than those in “real
there is no discounting of the study’s world” settings.24 Therefore, we make an adjustment to
evaluation outcomes). effect sizes ESm to reflect this distinction. As a parameter
• In our previous meta-analytic studies of other for all studies deemed not to be “real world” trials, the
areas such as criminal justice, a level 4 study Institute discounts ES'm by .5, although this can be
carried a factor of .75 (effect sizes discounted modified on a study-by-study basis.
by 25 percent). We will re-evaluate the
magnitude of this adjustment for the K–12
literature. Appendix B: Estimating Costs and
• In our previous meta-analytic studies of other Benefits of Educational Outcomes
areas such as criminal justice, a level 3 study
carried a factor of .50 (effect sizes discounted This technical appendix describes how the Institute has
by 50 percent). We will re-evaluate the previously estimated the costs and benefits associated
magnitude of this adjustment for the K–12 with various educational outcomes.25 These methods
literature. will be re-examined and refined for this study.
• We do not include level 2 and level 1 studies in
our analyses. Valuation of Education Outcomes
These factors are subjective to a degree; they are Many K–12 educational outcomes are measures of
based on our researchers’ general impressions of their human capital: graduation from high school, number of
confidence in the predictive power of studies of years of schooling completed, and achievement test
different quality. We also rely on evidence of the scores earned during the K–12 years. Other often-
degree and direction of selection bias in non-random measured educational outcomes relate to the use of
assignment studies to establish these adjustments.23 certain K–12 resources: years of special education and
grade retention. The benefits associated with each of
The effect of the adjustment is to multiply the effect size these possible outcomes are discussed in this section.
for any study, ES'm, in equation (A3), by the appropriate Exhibit 3 lists the equations involved in the procedures
research design factor. For example, if a study has an to calculate economic values for these outcomes.
effect size of -.20 and it is deemed a level 4 study, then
the -.20 effect size would be multiplied by .75 to Human Capital Outcomes. Our approach estimates the
produce a -.15 adjusted effect size for use in the cost- value of changes in high school graduation rates, years
benefit analysis. of education completed, and achievement test scores
during the K–12 years by estimating the expected
Adjusting Effect Sizes for Relevance or Quality of change in lifetime earnings caused by a change in the
Outcome Measures. Our focus in this analysis is whether human capital measure. Measuring the earnings
K–12 educational programs and services improve implications of these human capital variables is a
academic outcomes. We prefer measures such as test commonly used approach in economics.26
scores or grades and avoid measures such as tardiness or
self-esteem, since these may or may not be related to
long-run academic outcomes.
24
Ibid. Lipsey found that, for juvenile delinquency evaluations,
In addition, we require that all studies have at least a six- programs in routine practice (i.e., “real world” programs)
month follow-up period. For those studies that have a produced effect sizes only 61 percent as large as
follow-up period of less than 12 months but more than six research/demonstration projects. See also: A. Petrosino and
months, and for those studies that only use weak H. Soydan (2005). The impact of program developers as
measures, we reduce effects sizes by 25 percent. This evaluators on criminal recidivism: Results from meta-analyses of
adjustment multiplies the effect size for any study with a experimental and quasi-experimental research. Journal of
Experimental Criminology 1(4): 435-450.
short follow-up period or weak measure by .75. 25
Aos, et al. (2004). Benefits and costs of prevention and early
intervention programs for youth.
26
See, for example, A. Krueger (2003) Economic considerations
23
M. Lipsey (2003). Those confounded moderators in meta- and class size. The Economic Journal 113(485): F34-F63,
analysis: Good, bad, and ugly. The Annals of the American accessed from the author’s website:
Academy of Political and Social Science 587(1): 69-81. http://edpro.stanford.edu/eah/eah.htm; and E. Hanushek
Lipsey found that, for juvenile delinquency evaluations, (2003, October) Some Simple Analytics of School Quality,
random assignment studies produced effect sizes only 56 accessed from the author’s website:
percent as large as nonrandom assignment studies. http://edpro.stanford.edu/Hanushek/files_det.asp?FileId=139
8
In this analysis, all human capital earnings estimates For the human capital number of years of education
derive from a common dataset. The estimates are taken outcome, the process is exactly the same. The CPS
from the latest U.S. Census Bureau’s March Supplement money earnings data, by age, are taken as a weighted
to the Current Population Survey, which provides cross- average of those with a high school diploma and those
sectional data for earnings by age and by educational with some college but no degree, Earnhsgradplus. This
status.27 To these data we apply different measures of stream of earnings is multiplied by an estimated rate of
the net advantage gained through increases in each return to earnings per extra year of formal education,
human capital outcome. EdyearsROR. The remaining calculations in (B6) follow
those discussed for equation (B2).
For the human capital high school graduation outcome,
the CPS money earnings data, by age, are differenced Some K–12 programs, policies, and services we will
between those who graduate from high school (with no evaluate include more than one of these human capital
further degree), Earnhsgrad, and those with less than a variables. For example, some K–12 evaluations
high school diploma, Earnnonhsgrad. This differenced produce effect sizes for high school graduation and for
series is then present valued to age 18 by applying the K–12 test scores. In these cases, we only include one
general real discount rate used in the overall analysis, of the human capital variables, and we use the
Dis, and any assumed real rate of growth in wages, outcome that produces the highest economic return.
Earnesc. We use age 65 as the cut-off point for
earnings. K–12 Resource Outcomes. The model can also
calculate the value of two other often measured K–12
These earnings in equation (B1) are then present valued educational outcomes: years of special education and
further to the age of the person in the program, progage. grade retention. The present value costs of a year of
The values are also converted to the base year dollars special education is estimated by discounting the cost
chosen for the overall cost-benefit analysis, IPDbase, of a year in special education, SpecEdCostYear, for the
relative to the year in which the CPS data are estimated average number of years that special
denominated, IPDcps. A fringe benefit rate is applied to education is used, conditional on entering special
the earnings, Fringe. As mentioned, the model can education, specedyears. These years are assumed to
accommodate a rough estimate of any non-market (i.e., be consecutive. The present value is to the age when
non-earnings) outcomes that may be causally related to special education is assumed to first be used, start. In
education outcomes; this is modeled with the NonMarket equation (B8), this sum is further present valued to the
parameter in equation (B2). Additionally, since the age of the youth in a program, progage, and the cost is
observed difference between the wages of these two expressed in the dollars used for the overall cost
groups may not be all due to the causal factor of earning benefit analysis, IPDbase, relative to the year in which
a high school diploma, a multiplicative the special education costs per year are denominated,
causation/correlation factor, HSgradCC (with a value IPDspecedcostyear.
greater than or equal to zero, or less than or equal to
one), can be applied to the present value to provide an The present value cost of an extra year of K–12
estimate of the causal effect.28 education is estimated for those retained for an extra
year. This is modeled by assuming that the cost of the
For the human capital achievement test score outcome, extra year of K–12 education, EdCostYear, after
a similar process is used. The CPS money earnings adjusting the dollars to be denominated in the base
data, by age, are taken as a weighted average of those year dollars used in the overall analysis, would be
with a high school diploma and those with some college, borne when the youth is approximately 18 years old.
Earnhsgradplus, but not a college degree. This stream Since there is a chance that the youth will not finish
of earnings is multiplied by an estimated rate of return to high school and, therefore, that the cost of this year will
earnings per one standard deviation increase in never be incurred, this present valued sum is multiplied
achievement test scores, TestScoreROR.29 We calculate by the probability of high school completion,
a present value to age 18 by applying the general real Hsgradprob.
discount rate used in the overall analysis, Dis, and any
assumed real rate of growth in wages, Earnesc. We use
age 65 as the cut-off point for earnings. The remaining
calculations in equation (B4) follow the procedures
discussed for equation (B2).
27
The data are from the March Supplement to the CPS, PINC-04.
Educational Attainment—People 18 Years Old and Over by Total
Money Earnings, Age, Race, Hispanic Origin, and Sex.
28
These types of causation/correlation adjustments have also
been made in other cost-benefit analyses to avoid overstating
benefits due to some unobserved selection bias. See, for
example, M. Cohen (1998) The monetary value of saving a high
risk youth. Journal of Quantitative Criminology 14(1): 5-33.
29
Hanushek (2003). Some Simple Analytics of School Quality.
9
Other Outcomes Linked to Human Capital Outcomes.
Research literature has also focused attention on several
types of non-market benefits associated, perhaps
causally, with the human capital outcomes evaluated in
this analysis. A listing of possible non-market benefits to
education appears in the work of Wolfe and Haveman.30
In our current cost-benefit model, we do not estimate
these non-earnings values explicitly, with one exception
(discussed below). Rather, we provide a simple
multiplicative parameter that can be applied to the
estimated earnings effects so that the non-market
benefits can be roughly modeled. Since some research
indicates that these non-market benefits of human capital
outcomes can be considerable, future refinements to our
cost-benefit model will attempt to analyze these possible
non-wage benefits explicitly.
30
B. Wolfe and R. Haveman (2002) “Social and nonmarket
benefits from education in an advanced economy.” Proceedings
from the Federal Reserve Bank of Boston's 47th economic
conference, Education in the 21st Century: Meeting the
Challenges of a Changing World, accessed from:
http://www.bos.frb.org/economic/conf/conf47/index.htm. See
also a collection of articles on the topic published in J. Behrman
and N. Stacey, eds. (1997). The social benefits of education.
Ann Arbor: The University of Michigan Press.
31
L. Lochner and E. Moretti (2004). The effect of education on
crime.
10
Exhibit 3
Equations Used to Calculate the Present Valued
Costs of Education Outcomes
IPDbase
PVEarn18 × × (1 + Earnesc)18− progage × Fringe × (1 + NonMarket ) × HSgradCC
IPDcps
(B2) PVEarn progage =
(1 + Dis )18− progage
Test Scores
65 ( Earnhsgradplus y × TestScoreROR y ) × (1 + Earnesc) y −17
(B3) PVEarn18 = ∑ (1 + Dis ) y −17
y =18
IPDbase
PVEarn18 × × (1 + Earnesc)18− progage × Fringe × (1 + NonMarket ) × TestScoreCC
IPDcps
(B4) PVEarn progage =
(1 + Dis )18− progage
Years of Education
65 ( Earnhsgradplus y × EdyearsROR y ) × (1 + Earnesc) y −17
(B5) PVEarn18 = ∑ (1 + Dis ) y −17
y =18
IPDbase
PVEarn18 × × (1 + Earnesc)18− progage × Fringe × (1 + NonMarket ) × EdyearsCC
IPDcps
(B6) PVEarn progage =
(1 + Dis )18− progage
Earnhsgrad = The annual CPS earnings of high school graduates. Annual money earnings of an individual in year y, taken
from the U.S. Census Bureau’s March Current Population Survey, Annual Demographic Supplement, Table:
PINC-04. Educational Attainment—People 18 Years Old and Over, by Total Money Earnings, Age, Race,
Hispanic Origin, and Sex.
Earnnonhsgrad = The annual CPS earnings of non high school graduates, same source as above.
Earnhsgradplus = The annual CPS earnings of high school graduates plus those with some college but no degree, same source as
above.
Earnesc = An estimated long-run annual growth rate in real earnings.
IPDbase, IPDcps = The implicit price deflator for the year chosen as the base year for the overall analysis and for the year in which
the current population survey is based.
NonMarket = An estimate of the non-earnings benefits of education expressed as a percentage of the earnings effect.
HSgradCC = A causation-correlation factor for high school graduation to adjust the cross-sectional CPS data.
TestScoreCC = A causation-correlation factor for test scores to adjust the cross-sectional CPS data.
EdyearsCC = A causation-correlation factor for the number of years of education to adjust the cross-sectional CPS data.
Fringe = The fringe benefit rate used in the analysis.
Taxrate = The tax rate used in the analysis.
TestScoreROR = The annual rate of return for a one standard deviation increase in achievement test scores.
EdyearsROR = The annual rate of return for an extra year of education.
progage = The average age of a youth in a program.
Dis = The real discount rate.
11
Exhibit 3, Continued
Equations Used to Calculate the Present Valued
Costs of Education Outcomes
IPDbase
PVspecedstart ×
IPDspeced cos tyear
(B8) PVspeced progage =
(1 + Dis) start − progage
Washington State
Institute for Public Policy
The Washington Legislature created the Washington State Institute for Public Policy in 1983. A Board of Directors—representing
the legislature, the governor, and public universities—governs the Institute and guides the development of all activities. The
Institute's mission is to carry out practical research, at legislative direction, on issues of importance to Washington State.