- Paul T. von Hippel
Search for EdWorkingPapers here by author, title, or keywords.
Paul T. von Hippel
At least sixteen US states have taken steps toward holding teacher preparation programs (TPPs) accountable for teacher value-added to student test scores. Yet it is unclear whether teacher quality differences between TPPs are large enough to make an accountability system worthwhile. Several statistical practices can make differences between TPPs appear larger and more significant than they are. We reanalyze TPP evaluations from 6 states—New York, Louisiana, Missouri, Washington, Texas, and Florida—using appropriate methods implemented by our new caterpillar command for Stata. Our results show that teacher quality differences between most TPPs are negligible—.01-.03 standard deviations in student test scores—even in states where larger differences were reported previously. While ranking all a state’s TPPs may not be possible or desirable, in some states and subjects we can find a single TPP whose teachers stand out as significantly above or below average. Such exceptional TPPs may reward further study.
Many kindergarten teachers place students in higher and lower “ability groups” to learn math and reading. Ability group placement should depend on student achievement, but critics charge that placement is biased by socioeconomic status (SES), gender, and race/ethnicity. We predict group placement in the Early Childhood Longitudinal Study of the Kindergarten class of 2010-11, using linear and ordinal regression models with classroom fixed effects. The best predictors of group placement are test scores, but girls, high-SES students, and Asian Americans receive higher placements than their test scores alone would predict. One third of students move groups during kindergarten, and some movement is predicted by changes in test scores, but high-SES students move up more than score gains would predict, and Hispanic children move up less. Net of SES and test scores, there is no bias in the placement of African American children. Differences in teacher-reported behaviors explain the higher placement of girls, but do little to explain the higher or lower placement of other groups. Although achievement is the best predictor of ability group placement, there are signs of bias.
Enrollment in higher education has risen dramatically in Latin America, especially in Chile. Yet graduation and persistence rates remain low. One way to improve graduation and persistence is to use data and analytics to identify students at risk of dropout, target interventions, and evaluate interventions’ effectiveness at improving student success. We illustrate the potential of this approach using data from eight Chilean universities. Results show that data available at matriculation are only weakly predictive of persistence, while prediction improves dramatically once data on university grades become available. Some predictors of persistence are under policy control. Financial aid predicts higher persistence, and being denied a first-choice major predicts lower persistence. Student success programs are ineffective at some universities; they are more effective at others, but when effective they often fail to target the highest risk students. Universities should use data regularly and systematically to identify high-risk students, target them with interventions, and evaluate those interventions’ effectiveness.
Year-round school calendars take the usual 175-180 instruction days of the school year and redistribute them, replacing the usual schedule – nine months on, three months off – with a more “balanced” schedule of short instruction periods alternating with shorter breaks across all four seasons of the year. Over the past three decades, the number of schools using year-round calendars has increased ninefold, from 410 in 1985 to 3,700 in 2011-12 (Skinner, 2014). Over 2 million children now attend year-round schools – as many as attend charter schools – yet year-round schools have attracted relatively little attention from researchers and the public.
In this chapter, I review the evidence for the effects of year-round calendars on test scores. Once thought to be positive, these effects now appear to be neutral at best. Although year-round calendars do increase summer learning, they reduce learning at other times of year, so that the total amount learned over a 12-month period is no greater under a year-round calendar than under a nine-month calendar. I also review evidence that year-round calendars make it harder to recruit and retain experienced teachers, make it harder for mothers to work outside the home, and reduce property values. When students' schedules are staggered, year-round calendars do offer a way to reduce school crowding – an alternative to busing or portable classrooms, and a low-cost alternative to new school construction.
Evidence-based policy is the practice of basing policy decisions on rigorous research evidence, such as randomized experiments. But it is unclear how often evidence-based decisions produce more effective policy. We evaluate an evidence-based policy implemented in 1989-93, after the state of Tennessee completed the famous Project STAR randomized experiment, which showed that reducing average class sizes from 23 to 15 could raise test scores by nearly 0.2 standard deviations (SD). After Project STAR, the state launched Project Challenge, which tried to achieve similar score gains by earmarking $5 million to reduce class sizes in the state’s 17 poorest districts.
We evaluate the effects of Project Challenge by applying regression discontinuity and difference in differences analysis to data from district report cards. Our analysis offers no evidence that Project Challenge districts raised test scores, and even raises questions about whether districts reduced class sizes. After Project Challenge, Tennessee’s Basic Education Plan did reduce class sizes, but only by a token amount, from 26 to 25. In this example, it seems that a successful randomized experiment did not lead to successful policy.
Debates in education policy draw on different theories about how to raise children’s achievement. The school competition theory holds that achievement rises when families can choose among competing schools. The school resource theory holds that achievement rises with school spending and resources that spending can buy. The family resources theory holds that children’s achievement rises with parental education and income. We test all three theories in Chile between 2002 and 2013, when reading and math scores rose by 0.2-0.3 standard deviations, while school competition, school resources, and family resources all increased. In a difference in differences analysis, we ask which Chilean municipalities saw the greatest increases in test scores. Test scores did not rise faster in municipalities with greater increases in competition, but did rise faster in municipalities with greater increases in school resources (teachers per student) and especially family resources (parental education, not income). Student grade point averages show similar patterns. Results contradict the school competition theory but fit the family resource theory and, to a lesser extent, the school resource theory.