The publication of rankings of schools
Transcrição
The publication of rankings of schools
The publication of school rankings: a step toward increased accountability? Ana Balcão Reis*,t Carmo Seabra*,t Luis C. Nunes*,t *Nova School of Business and Economics, Universidade Nova de Lisboa, Campus de Campolide 1099-032 LISBOA Portugal October 2013 Abstract This paper contributes to the discussion of the effects of the publication of school rankings. We study the effectiveness of this (low-stakes) accountability mechanism. Our results suggest that the publication of rankings has strong effects upon families and schools. After the rankings publication, students move out of badly classified schools and the probability of closing for these schools increases. These effects are stronger for private schools. Keywords: school accountability, school rankings JEL: I21, I28, H42 * Email: [email protected] , [email protected] , [email protected] We would like to thank João Miguel Silva and Adriana Ferro for excellent research assistance. Financial support from Fundação para a Ciência e Tecnologia is gratefully acknowledged. t 0 The publication of school rankings: a step toward increased accountability? October 2013 Abstract This paper contributes to the discussion of the effects of the publication of school rankings. We study the effectiveness of this (low-stakes) accountability mechanism. Our results suggest that the publication of rankings has strong effects upon families and schools. After the rankings publication, students move out of badly classified schools and the probability of closing for these schools increases. These effects are stronger for private schools. Keywords: school accountability, school rankings JEL: I21, I28, H42 1 1. Introduction During the last decade, many countries and/or states have introduced some form of accountability in schools in an attempt to increase their performance. This movement was at least partly generated by the frustration felt by many countries with the results of their heavily financed public education systems. The weaknesses of these results became clearer as the realization of internationally comparable exams, like PISA and TIMMS, was generalized. Accountability policies vary considerably, from the so-called low-stakes policies (disclosing information regarding the performance of schools) to high-stakes or consequential systems, whereby the financing conditions of the schools and/or the payments to teachers are adjusted according to the performance of students. The effectiveness of these accountability systems on the learning conditions of students has been studied extensively in the literature. Using different data sets and methodological tools, several authors have shown that accountability policies, even low-stakes ones, such as the one considered in this paper, have significant and positive effects on students’ behavior. Figlio and Lucas (2004) showed that lowstakes accountability policies have strong effects on parents’ choices and are reflected in changes in housing prices. Using data from Florida, where a hard-stakes policy was launched, Figlio and Rouse (2006) concluded that the improvements of poorly performing schools were, in fact, very large; Chakrabarti (2008) examined the impact of different incentive schemes and showed that the 1999 Florida accountability program unambiguously improved schools’ performance. Chiang (2009) also found persistent gains in performance due to Florida’s highstakes accountability policy; Hanushek and Raymond (2004) reported increased scores for all students, following the introduction of accountability policies in the US, although they found a much weaker impact of low-stakes policies relative to the effects of sequential or high-stakes measures. Using data from PISA 2003 for 27 OECD countries, Schutz et al. (2007) also found a positive relationship between accountability and academic outcomes. 1 In this paper we analyze the impacts of the public disclosure of school rankings, a low-stakes accountability policy. We consider the case of Portugal, where every year several newspapers publish rankings of every high school (public and private) based on the average scores obtained by students on national exams. Secondary school rankings were first published in 2001, were not published in 2002, and then published every year thereafter. We study the effect of the publication of the rankings in terms of the ability of schools to attract students and of the increased probability of closing of badly classified ones. To identify the effect of the public disclosure of school results, we compare the effects of the rankings before and after their publication. We distinguish between private and public schools, as we expect that there is more flexibility and freedom of choice in the subset of private schools, allowing for a stronger effect of the rankings on these schools. 1 Positive effects of several accountability systems have been found for several other geographical and institutional settings (see, for instance, Koning and van der Wiel (2012) for the Netherlands, Anghel et al. (2012) for Spain, Rockoff and Turner (2010) for New York City). 2 The organization of the paper is the following: Section 2 briefly describes the institutional setting of the education system in Portugal for the period under analysis. Section 3 describes the data. In Section 4 we present our methodology and our results for the effect of rankings publication on students’ relocations. In this section we also study the impact of the published ranking on the probability that schools close. These analyses help to understand the impact and the vehicle through which the publication of rankings affects schools. Section 5 discusses the results and concludes. 2. The Institutional setting in the period 1998 to 2005 The Portuguese educational system did not undergo significant changes from 1998 to 2005. Mandatory schooling starts at the age of six and lasts 9 years.2 These 9 years are divided into three cycles with durations of 4, 2, and 3 years, respectively. Tracking in general starts at the end of mandatory schooling 3, with the choice between 3 additional years of academic studies aimed at the pursuit of studies at the university level, or professionally oriented studies. Vocational study tracks can consist of either 2 or 3 years, and transition to academic tracks or to the university is possible. The net rate of enrollment in secondary education was around 60% in the period analyzed. Below we describe only academic studies 4, as these are the subject of this paper. To finish academic secondary school, students have to take final exams, carried out at the national level. Final scores per subject are computed as an average of the score attributed by the student’s school (internal score) and the score obtained at the final exam. These national exams perform two roles in the educational system. Besides being required to graduate from secondary school, they also determine the conditions for admission to the university. Portuguese Public Universities have fixed slots set for each course by the Ministry, that fall short of demand for places in the most prestigious degrees. Candidates for each degree are selected (centrally, by the government) according to their order number in the candidacy grade. This grade is a weighted average of the score obtained at the end of secondary studies, calculated as described above, and the scores obtained in one or two of the mandatory national exams 5. These specific exams, thus, enter twice in the definition of the candidacy grade. These rules place a strong weight on the scores obtained in secondary final exams. Public and Private schools co-exist in the Portuguese educational system. The level of autonomy of Portuguese schools is very limited, with teaching contents and learning methods being decided by the government. Private schools can choose teachers and students, and set 2 After our period of analysis, some changes occurred, namely the extension of mandatory schooling by 3 years. The first cohort of students affected by this measure entered secondary schools only in 2010. 3 There are some professionally oriented paths initiated at the age of 14, but these tracks remained exceptional, accounting for only 4 to 5% of the schooling population. 4 The net rate of enrollment in secondary education ran from 59.1% in 1998 to 59.8% in 2005; out of these, the percentage of students in academic studies decreased during the period, from 65% in 1998 to 55% in 2005. 5 The choice of the specific subjects that are considered twice on the candidacy grade for each course is left to the course authorities. This choice has been kept stable since 1998. 3 tuition fees. 6 They are allowed to allocate teaching time to the different subjects, but only according to strict requirements set by the government. They are also subject to a set of rules concerning the premises, the number of students allowed per class, and so forth. Public schools are strictly restricted in their educational supply, and neither choose nor influence the choice of teachers. The allocation of students to public schools and to publicly financed private school is based largely on geographical criteria. For the years analyzed, there were legal restrictions on the specific school where students could enroll, according to the geographical proximity, first, to the home of the family, and second, to the work place of the mother or father. Nevertheless, the large movements of students in and out of schools in our data reveal that those legal restrictions were not being strictly enforced. 3. Data Description The publication of school rankings in Portugal occurred for the first time in 2001. The first ranking was published by the newspaper Público, following a court decision on a law suit by this newspaper, claiming its right to disclose the data that were being denied by the Government. In the following year, 2002, the Ministry of Education decided to publish its own ranking, taking into account some socio-economic characteristics of the school region. This ranking was strongly criticized, and from 2003 on the Ministry started disclosing the raw data to the media that treat them and publish the rankings. The published ranking is the simple ordering that results from calculating the average score on national exams for each school. This average is obtained taking into account a selection of exams that correspond to the courses with more students at the national level (Mathematics, Portuguese, etc.). The choice of which exams to include in this ranking is a decision of the newspaper. 7 Our dataset consists of these same data disclosed to the media and includes scores for all student-exam pairs for all of the 12th grade national exams since 1998. The exam scores have been published online since 2003 by the Ministry of Education, and may be freely downloaded. 8 The data for the period before the beginning of publication, covering years 1998 to 2001, became publicly available only at the end of 2012. Our main variable of interest is the ranking of each school, RANK. We replicated the school rankings published by the newspaper Público, given that our purpose was to analyze the impact of published rankings. For the period before publication we built equivalent rankings based on the students’ scores in each school, following the criteria adopted by Público in its publication. The variable RANK takes the value 1 for the school with higher average and increases as the school average decreases. 6 Some private schools are publicly financed and neither choose their students nor set tuition fees. After this several newspapers started publishing their own rankings, but Público remained the principal source on this issue. 8 http://www.dgidc.min-edu.pt/jurinacionalexames/ 7 4 To determine whether the publication of school rakings affects students' choices we first analyze whether the percent change in the number of students in each school depends on previous rankings. As we have no access to the number of students enrolled in each school, we extracted from the dataset described above the number of exams taken at each school in each year. This is our measure of the size of the school, DIMS. The percent change in the number of exams taken at each school in the period under analysis is our dependent variable, ΔN. We then look at the closing of schools. We define a dummy variable, CLOSE, that takes the value 1 if the school has students taking exams in the initial year of the period under analysis and no students taking exams in the final and subsequent years, and 0 otherwise. The dataset also includes information about school characteristics that we use as additional controls: public, PUB, versus private, PRIV, status and geographical localization. We use this last information to build several variables concerning the availability of schools in each municipality, and thus, the effective competitive pressure and freedom of choice of school by the students. We considered three alternatives: (i) CHOICE, a dummy that equals 1 if there is more than one school in the municipality, and thus, the possibility of choice, and 0 otherwise; (ii) NSMUN, the number of schools in the municipality, and (iii) HH, the Herfindahl-Hirschman Index of concentration of schools at the municipal level. To take into account that it is probably easier to move to a school in the nearby region, we have also calculated regional rankings. These would be the most relevant rankings if each school competes mainly with the schools in the same region. We considered that for each school the relevant region would be the one resulting from considering only the neighboring municipalities. The dataset does not include any information on the socio-economic characteristics of the students, and for the period under analysis this information is not available at the school level, either. 4. The impact of published rankings Our goal is to study the effect of the publication of the rankings. The gold standard for evaluating the impact of a treatment (in our case the publication of the rankings) is the randomized experiment in which units (in our case the schools) are assigned into two groups at random. One of the groups receives the treatment while the other, the control group, does not. In such a setting, one can simply take the difference between the outcomes of the two groups to measure the impact of the treatment. Randomized experiments are relatively rare in practice. For instance, assignment into the two groups may not be random, leading to differences in characteristics in the two groups which confound the impact of the treatment. If data before and after the treatment are available, the impact of the differences in characteristics can be eliminated by analyzing the difference-in-differences of the outcome variable. For this approach to be valid it requires that in the absence of the treatment, the two groups would have followed parallel trends. When the probability of receiving the treatment is a discontinuous function of one or more variables, an alternative approach consists in a regression discontinuity design. The idea is that by using only observations lying close to either 5 side of the threshold it is possible to obtain two groups with similar characteristics, except that one received the treatment and the other did not. In our study, the rankings were published for all schools. If the form of publication consisted of assigning specific labels to each school based on student performance on the exams, it would be possible to divide the schools into different groups and to follow a difference-in-differences approach (since we have data before and after the publication) or a regression discontinuity design (if the assignment of the labels was a discontinuous function of student performance). However, the form of publication consisted in disclosing only an ordered list of schools. Basically, each school is associated with a particular order so that there are no explicit control and treatment groups. As a result, we have to follow a different strategy. To capture the impact of the publication of the ranking, and because we cannot divide the schools into treatment and control groups, we estimate linear regression and probit models for the outcome variables with the ranking variable entering as an explanatory (continuous) variable. We include additional controls to capture other differences in the characteristics of the schools that may affect the outcome variables. Since we have data before and after the publication of the rankings, we estimate the models for the two periods and check for differences in how the ranking variable affects the outcome variable. If the impact of the ranking is the same in the two periods, we conclude that the publication had no impact. For this interpretation to be valid, the main assumption required is that, apart from the control variables included in the estimated models, the only difference between the two periods is the publication. In fact, as discussed above, between 1998 and 2005 there were no major changes in the educational system. To define the two estimation periods, before and after the publication, it is important to understand the timing of rankings’ publication and of students’ enrollments. Rankings are published at the end of summer, after the enrollment of students for the following academic year, which happens in June and July. So, the 2001 rankings are expected to have affected enrollments for the academic years of 2002/2003 and following. Moreover, although some students change school in the transition to the 12th grade, most of the movements between schools occur in the transition to the 10th grade. The exams are taken at the end of each academic year, in June. Thus, the impact of the 2001 rankings is expected to appear in the number of students taking exams in 2003 and, more strongly, in 2005, when secondary school exams are taken by those students who enrolled in the 10th grade in June 2002. We define our periods of analysis taking these timings into account. For the impact of publication, we look at the percent variation in the number of students taking exams in each school between 2001 and 2005. We regress this variation on the rank in 2001 and school characteristics at the beginning of the period. We consider the number of exams in 2001 to be an appropriate point of departure for the analysis as we may expect it to be exogenous relative to the ranking published in 2001, which was the first published ranking. For the period before publication we look at the percent variation in the number of students between 1998 and 2002. As before, we regress this variation on the rank in 1998 and school characteristics at the beginning of the period. 6 Table 1: Descriptive statistics Period before publication (schools with more than 50 exams in 1998): 569 schools Variable Description Mean St deviation Max Min ∆N percent change in the number of exams between 1998 and 2002 -11.19 36.98 352.59 -100.00 RANK Rank in 1998 299.43 171.59 607 2 PUB Dummy=1 for public schools 0.83 0.38 1 0 DIMS Number of exams in the school in 1998 375.88 236.48 1531 53 NSMUN Number of schools in the municipality in 1998 11.28 18.12 62 1 0.73 0.44 1 0 0.46 0.37 1 0.02 0.01 0.11 1 0 CHOICE HH CLOSE Dummy=1 if there is another school in the municipality in1998 Herfindahl-Hirschman Index of Concentration for 1998 Dummy=1 if the school has no exams from 2002 onwards Period after publication (schools with more than 50 exams in 2001): 583 schools Variable Description Mean St deviation Max Min ∆N percent change in the number of exams between 2001 and 2005 1.70 44.11 307.35 -100.00 RANK Rank in 2001 309.93 175.37 620 2 PUB Dummy=1 for public schools 0.82 0.38 1 0 DIMS Number of exams in the school in 2001 339.52 221.01 1648 51 NSMUN Number of schools in the municipality in 2001 10.52 16.86 60 1 0.72 0.45 1 0 0.47 0.37 1 0.02 0.04 0.19 1 0 CHOICE HH CLOSE Dummy=1 if there is another school in the municipality in 2001 Herfindahl-Hirschman Index of Concentration for 2001 Dummy=1 if the school has no exams from 2005 onwards We restricted our analysis to schools that had in the initial year of analysis more than 50 exams. 9 This eliminates extreme cases with very large percent variations due to the low initial number of exams. Table 1 lists the main variables and presents some descriptive statistics for the two periods of analysis, 1998-2002 and 2001-2005. Except for ∆N and CLOSE, there are no noticeable differences in the distribution of the variables between the two periods. In 1998 there were 569 secondary schools with more than 50 students applying to 12th grade exams, 99 private and 470 public. Of these, 7 closed before 2002. In 2001, there were 583 schools, 104 private and 479 public. Of these, 22 closed before 2005. 9 The variable RANK corresponds to the ranking published by Público which includes all the schools, 609 in 1998 and 622 schools in 2001. 7 4.1. The impact of published rankings on the Percent Change in the Number of Students In order to evaluate whether the publication of the rankings affects students’ decisions about the school that they choose to enroll in, we regressed the percent change in the number of exams performed in each high school on the rank of the school for the periods before and after the first publication, and to test whether there are systematic differences between those two periods. For the period before publication the explanatory variable is the rank implied by school scores in 1998 (which was not public knowledge at that time), and for the period after publication the independent variable is the rank published in 2001. We controlled for the private versus public nature of the school. In order to test whether the impact of rankings was weaker for public schools, we also considered the cross-effect between the rank and the public character of the school, by including the interaction variable PUB×RANK and PRIV×RANK, where the variable PRIV is a dummy that identifies private schools. We also controlled for the size of the school and for the degree of competition among schools in each region, as stated above. To test if the impact of published rankings was greater for schools subject to stronger competition, we also considered the cross-effect between the rank and the competition measure. Our benchmark specification is the following: ∆N i = α + β1 PUBi + β 2 PUBi × RANK i + β 3 PRIVi × RANK i + β 4 DIMS i + β 5CHOICEi + β 6 CHOICEi × RANK i + ε i Models were estimated by ordinary least squares. We also included in all the regressions dummy variables for the administrative regions (18 plus the islands) to control for differences in the demographic evolution across regions and other regional characteristics. Estimation results are shown in Tables 2.a and 2.b. 8 Table 2.a: Estimation Results of the Linear Regression Model Dependent variable: ∆N, 1998-2002 1 -5.94 CONST 2 1.71 5 28.52*** 6 21.00** 7 43.36*** 8 41.74*** 6.57 8.5 7.6 9.95 9.07 9.98 11.72 13.67 0.37 0.84 0.61 0.52 0.00 0.04 0.00 0.00 -0.02** -0.04** -0.07*** -0.05** s.e. 0.01 0.02 0.01 0.02 Prob 0.03 0.03 0.00 0.01 -1.63 -1.52 -5.30 -7.03 -5.04 -5.07 -28.95*** -28.64** s.e. 5.27 5.27 7.3 7.42 6.23 6.25 11.68 11.89 Prob 0.76 0.77 0.47 0.34 0.42 0.42 0.01 0.02 -0.03*** -0.03*** -0.03*** -0.03*** -0.02** -0.02** -0.01* -0.01 s.e. 0.01 0.01 0.01 0.01 0.01 0.01 0.01 0.01 Prob 0.00 0.00 0.00 0.00 0.05 0.05 0.1 0.1 -0.47 -9.33 -0.07 -10.58 -11.76*** -2.76 -10.45** -8.77 s.e. 4.11 7.59 4.13 7.64 4.05 -8.11 4.03 8.14 Prob 0.91 0.22 0.99 0.17 0.00 0.73 0.01 0.28 DIMS CHOICE 0.02 0.03 -0.03 -0.005 s.e. 0.02 0.02 0.02 0.02 Prob 0.29 0.21 0.25 0.83 CHOICE×RANK RANK×PUB s.e. Prob -0.02* -0.04** -0.05*** -0.05*** 0.01 0.02 0.01 0.02 0.08 0.04 0.00 0.01 -0.03 -0.06* -0.13*** -0.12*** s.e. 0.02 0.03 0.03 0.03 Prob 0.19 0.07 0.00 0.00 RANK×PRIV District Dummies 2 2 0.09 0.10 0.09 0.10 0.15 0.15 0.17 0.17 0.06 0.06 0.05 0.05 0.12 0.12 0.13 0.13 -0.02 -0.08*** s.e. 0.01 0.01 Prob 0.14 0.00 β RANK +β CHOICE×RANK 0.01 0.02 0.08*** 0.08*** s.e. 0.03 0.03 0.03 0.03 Prob 0.63 0.48 0.01 0.01 569 569 583 583 β RANK×PUB -β RANK×PRIV # Observations 4 6.36 s.e. PUB ADJUSTED R 3 -3.93 Prob RANK R Dependent variable: ∆N, 2001-2005 569 569 583 583 Notes: s.e stands for robust standard-errors *** significant at the 1 percent level, ** significant at the 5 percent level, * significant at the 10 percent level. 9 Table 2.b: Estimation Results of the Linear Regression Model Dependent variable: ∆N , 1998-2002 2 3 4 5 6 7 8 10.32 8.88 11.56 9.16 39.61*** 34.62*** 55.69*** 50.93*** s.e. 7.08 7.13 8.32 9 9.66 9.56 12.18 12.92 Prob 0.15 0.21 0.17 0.31 0.00 0.00 0.00 0.00 CONST RANK -0.03*** -0.02** -0.08*** -0.06*** s.e. 0.01 0.01 0.01 0.01 Prob 0.01 0.04 0.00 0.00 -3.75 -3.94 -5.81 -4.31 -5.31 -5.41 -29.99*** -27.27** s.e. 5.31 5.27 7.41 8.08 6.16 6.16 11.72 12.24 Prob 0.48 0.46 0.43 0.59 0.39 0.38 0.01 0.03 PUB DIMS -0.03*** -0.03*** -0.03*** -0.03*** -0.03*** -0.03*** -0.02*** -0.02*** s.e. 0.01 0.01 0.01 0.01 0.01 0.01 0.01 0.01 Prob 0.00 0.00 0.00 0.00 0.00 0.00 0.01 0.00 0.38*** -0.30** -0.38*** -0.30** -0.60*** -0.31 -0.58*** -0.42** s.e. 0.1 0.13 0.1 0.13 0.17 0.21 0.17 0.22 Prob 0.00 0.02 0.00 0.02 0.00 0.14 0.00 0.05 NSMUN NSMUN×RANK -0.00 -0.00 -0.00** -0.00 s.e. 0.00 0.00 0.00 0.00 Prob 0.42 0.48 0.04 0.28 RANK×PUB -0.03** -0.02** -0.05*** -0.04*** s.e. 0.01 0.01 0.01 0.01 Prob 0.01 0.04 0.00 0.00 -0.03 -0.02 -0.13*** -0.11*** 0.02 0.03 0.03 0.03 RANK×PRIV s.e. Prob District Dummies 2 R ADJUSTED R β RANK*PUB β RANK×PRIV Dependent variable: ∆N , 2001-2005 1 2 0.18 0.42 0.00 0.00 0.11 0.11 0.11 0.11 0.17 0.18 0.19 0.19 0.07 0.07 0.07 0.07 0.13 0.14 0.15 0.15 0.01 0.00 0.08*** 0.07** s.e. 0.03 0.03 0.03 0.03 Prob 0.79 0.96 0.01 0.03 569 569 583 583 # Observations 569 569 583 583 Notes: s.e stands for robust standard-errors *** significant at the 1 percent level, ** significant at the 5 percent level, * significant at the 10 percent level. We discuss in detail the results shown in Table 2.a. where the control variable CHOICE is considered. Similar conclusions are derived from Table 2.b. considering instead the control variable NSMUN. 10 For both periods and all models, the size of the school, DIMS, has a negative impact on the percent variation in the number of exams. The parameters are in general significant at the 1 percent level. The number of schools available in the region also has a negative and significant impact on the percent variation in the number of exams. 10 For the period before publication, according to model 1 (column 1 in Table 2.a) the effect of the ranking is negative and significant. In model 2 we include the interaction between choice and ranking to capture the possibility that the impact of the ranking is amplified when there are alternative schools in the municipality. However, this interaction is not significant. In model 3 we check whether the impact of the rank is different for public and private schools. We find no significant differences, as can be checked in the last row of Table 2.a. Finally, in model 4 we allow for both the interaction between choice and the rank and the difference between the impact of the rank in public and private schools and obtain that none is significant. Thus, model 1 appears as the one that best fits the data for the period before publication. For the period after publication we estimated the same models, and obtain a negative and significant effect of the rank in model 5. We obtain again, in models 6 and 8, that the interaction between rank and choice is never significant. In model 7 we obtain that now the impact of the rank is significantly different for private and public schools. Hence, for this period model 7 is the preferred specification. The above findings suggest that the impact of the ranking changed due to the publication of the ranking. This is also confirmed by the results of the tests for differences in the relevant parameters comparing models 1 versus 5 and 3 versus 7 shown in Table 3. Table 3: Tests for Equality of the Parameters before and after publication (t-statistics) RANK RANK×PUB RANK×PRIV Models 1 versus 5 3 versus 7 -3.21*** -1.84* -2.65*** *** significant at the 1 percent level, * significant at the 10 percent level Our main finding is that the direct impact of the ranking on the number of exams has the expected negative sign and is significantly stronger for the period after publication. An increase in the rank of the school (corresponding to a deterioration of quality) leads to a decrease in the number of exams carried out in that school and this effect is much stronger when the information about rankings is published, and thus, available to everyone, as shown in Table 3. This Table also shows that the impact of the rank in both public and private schools changed 10 We do not show the results with the Herfindahl-Hirschman Index of concentration, as they are qualitatively identical to those shown. 11 significantly with the publication of the rankings. Furthermore, private schools were more strongly affected by publication. 4.2. The impact of published rankings on the probability of closing As an important part of the effect of published rankings may be the closing of badly classified schools, we now look at the event of closing. In our analysis a closed school is one that presents no students to the national exams in the final year of the period analyzed. We estimated the probability that a school i closes as a function of published rankings, considering the same controls as in the previous regressions. We assumed a probit function for the probability to close. 11 Prob(CLOSEi = 1) = Φ (α + γ 1 PUBi + γ 2 PUBi × RANK i + γ 3 PRIVi × RANK i + γ 4 DIMSi + γ 5 NSMUN i + γ 6 NSMUN i × RANK i ) Again we look separately at the periods before and after publication and compare the results. The estimation results are shown in Table 4. The table shows marginal effects. 12 For instance, model 6 obtains that an increase of 100 places in the ranking increases the probability that a school will close by 3 percentage points. Table 4: Estimation results of the probit model for the probability of closing - Marginal Effects Dependent variable: CLOSE, 1998-2002 1 RANK 2 3 4 Prob -0.00003 0.18 0.0004 0.14 Prob -0.003 0.69 -0.009 0.34 0.05 0.24 Prob -0.0002** 0.04 -0.0001** 0.04 Prob 0.0005*** 0.00 0.004 0.13 PUB DIMS NSMUN NSMUN×RANK Prob 5 6 7 8 0.0001*** 0.004 0.0003*** 0.01 0.10 0.21 0.01 0.60 0.01 0.52 0.09 0.14 0.11* 0.07 -0.0001** 0.04 -0.0001** 0.04 -0.0002** 0.01 -0.0002*** 0.01 -0.0002*** 0.01 -0.0002*** 0.004 0.0004** 0.01 0.004* 0.10 0.001*** 0.00 -0.000*** 0.001 0.001*** 0.00 -0.000*** 0.001 -0.0000 0.15 RANK×PUB Dependent variable: CLOSE, 2001-2005 -0.0000 0.10 -0.0000* 0.06 -0.0000** 0.04 Prob -0.0000 0.74 0.0004 0.14 0.00007 0.22 0.0002** 0.04 Prob 0.0001 0.15 0.0005* 0.08 0.0003** 0.03 0.0004*** 0.002 RANK×PRIV McFadden R2 0.39 0.46 0.43 0.52 0.33 0.35 0.35 0.38 # Observations 569 569 569 569 583 583 583 583 Notes: the p-values correspond to the estimated coefficients of the Probit regression. *** significant at the 1 percent level, ** significant at the 5 percent level, * significant at the 10 percent level. 11 The variable CHOICE does not have enough variability in the sample to allow its inclusion in this regression. Thus, in this analysis we consider only the variable NSMUN as a measure of competition. 12 These are marginal effects averaged across schools. 12 For both periods and all models, the size of the school has a significant negative impact on the probability of closing. The number of schools available in the region also has a significant impact in almost all the models: the larger the number of alternatives, the greater the probability of closing. For the period before publication, the effect of the ranking is not significant in models 1 and 2. If we distinguish between private and public schools, in models 3 and 4, the effect of the rank is not significantly different between them. In models 2 and 4 we include the interaction between NSMUN and RANK to capture the possibility that the impact of the ranking is amplified when there are alternative schools in the municipality. However, this interaction is not significant. For the period after publication we estimated the same models. We obtain a positive and significant effect of the rank in models 5 and 6. In models 7 and 8 we allow for the impact of the rank to be different for private and public schools and find, in model 8, that the impact is significantly stronger for private schools. Moreover, the effect of the rank is amplified by the existence of alternatives, as shown in models 6 and 8. As in the previous analysis, our results show that the publication of the rank had a strong impact. We estimated both the linear regressions and the probit models considering the regional rankings described at the end of Section 3 and obtained results similar to those obtained with the national rankings. 13 This corroborates our previous findings. 5. Discussion and Conclusions Our findings suggest that the publication of rankings has strong effects upon families and schools. Students react by moving in and out of the schools, according to their published rank. This result is in line with the empirical literature about the effects of low-stakes accountability policies on the performance of schools measured by scores on national exams. This result is especially interesting given that, at least for public schools, there are typically legal restrictions on enrollment decisions according to the geographic location of the student’s home. Even so, students exert some choice over which school to enroll in. We find that after the rankings publication, badly classified schools lose students and their probability of closing increases. We also find that the effect for private schools is stronger as would be expected, given that the freedom of choice between private schools is greater than between public ones. As several authors have already stressed, school quality measures such as the ones studied in this paper may be very imprecise measures of school quality. There are several reasons for these weaknesses: (i) rankings of schools are too volatile and imprecise (Kane and Staiger (2002), Chay et al. (2005)); (ii) rankings reflect factors that are outside the control of schools, especially the socioeconomic background of students’ families, and not the better quality of well performing schools (Mizala et al. (2007)); (iii) public disclosure of schools‘ results biases 13 This may be justified by the small size of Portugal. 13 schools’ and teachers’ efforts toward an excessive investment on test-specific skills, easily observable, at the cost of other school outputs, not so easily measurable (Jacob (2005), Reback (2009), Neal and Schanzenbach, (2010)); (iv) accountability systems can be “cheated on” in several ways, so that any reported beneficial effects of accountability may be merely the result of schools gaming or cheating on the system, with no real content as far as effective quality evaluation is concerned (Cullen and Reback (2006), Figlio and Getzler (2002)). Moreover, the public disclosure of bad outcomes might have perverse consequences over equity due to cream-skimming of both students and teachers, or to motivational effects (Clotfelter et al. (2004), Figlio and Lucas (2004), Schutz et al.(2007)). The position emphasizing the possible negative effects of the publication of school rankings is shared by politicians in many countries, and calls for the prohibition of public disclosure of school rankings that is seen in Spain and some other European countries (see Anghel et al., 2012). Our point here is not to state that the rankings are a good instrument of policy, but rather to analyze whether their publication has consequences over the education system. Our conclusion is that it has, and that the effects are strong. Kane and Staiger (2002) regret that, for the US, “school accountability systems have led to little reallocation of students across schools” as they believe this would be an important mechanism to improve school quality. We show that in Portugal people have reacted to school rankings by moving to better schools. Our results reinforce the need to verify the type and quality of the information implicit in those rankings. Also, as mentioned above, the disclosure of this information is independent from any government intervention. Given that, as illustrated by Mizala et al. (2007), it may be very hard to produce a ranking of schools that is not too volatile and simultaneously does not just reflect students’ socioeconomic status, the issue of the existence of potential benefits stemming from the public disclosure of complementary information gains relevance. References Anghel, Brindusa, Antonio Cabrales, Jorge Sainz, and Ismael Sanz, 2012, “Publicizing the results of standardized external tests: Does it have an effect on school outcomes?”, Working-Paper 1234, Economic Series, Universidad Carlos III de Madrid. Chakrabarti Rajashri, 2008,” Impact of Voucher Design on Public School Performance: Evidence from Florida and Milwaukee Voucher Programs”, Federal Reserve Bank of New York Staff Report no. 315. Chay, Kenneth, Patrick MacEwan, and Miguel Urquiola, 2005, ''The Central Role of Noise in Evaluating Interventions that Use Test Scores to Rank Schools'', The American Economic Review, 95 (4): 1237-1258. Chiang, Hanley, 2009, “How accountability pressure on failing schools affects student achievement”, Journal of Public Economics, 93, 1045-1057. 14 Clotfelter, Charles T. Helen F. Ladd , Cob L. Vigdor, and Aliaga Diaz, 2004, “Do School Accountability systems Make It More Difficult for Low-Performing Schools to Attract and Retain High-Quality Teachers?” Journal of Policy Analysis and Management, Vol. 23, ( 2), 251–271. Cullen, Julie Berry and Randall Reback, 2006, “Tinkering toward accolades: school gaming under a performance accountability system” in T. Gronberg and D. Jansen, eds., Improving School Accountability: Check-Ups or Choice, Advances in Applied Microeconomics 14 (Amsterdam: Elsevier Science): 1-34. Elstad, Eyvind, 2009, "Schools that are Named, Shamed and Blamed by the Media: School Accountability in Norway", Educational Assessment, Evaluation and Accountability, 21(2): 173189. Figlio, David N. and Lawrence S. Getzler, 2002, “Accountability, ability and disability: gaming the system”, National Bureau of Economic Research Working Paper 9307, at http://www.nber.org/papers/w9307 . Figlio, David N. and M. Lucas, 2004, “What’s in a grade? School Report Cards and the Housing Market,” American Economic Review, 94 (3): 591-604. Figlio, David N. and C. Rouse, 2006, “Do accountability and voucher threats improve lowperforming schools?”, Journal of Public Economics, 90. Hanushek, Eric and M. Raymond, 2005, “Does school accountability lead to improved student performance”, Journal of Policy Analysis and Management, 24, (2): 297-325. Jacob, Brian, 2005, “Accountability, Incentives and behavior: the impact of high-stakes testing in the Chicago Public Schools”, Journal of Public Economics, 89: 761-896. Kane, J. Thomas and Douglas O. Staiger, 2002, “The Promise and Pitfalls of Using Imprecise School Accountability”, Journal of Economic Perspectives, 16 (4), pages 91-114. Koning, Pierre and Karen van der Wiel, 2012, “School Responsiveness to Quality Rankings: An Empirical Analysis of Secondary Education in the Netherlands”, De Economist, vol 160 (4), pages 339-355. Mizala, Alejandra, Pilar Romaguera, and Miguel Urquiola, 2007, “Socioeconomic status or noise? Tradeoffs in the generation of school quality information”, Journal of Development Economics, 84, pages 61-75. Neal, D. and D. Schanzenbach, 2010, “Left behind by design: proficiency counts and test-based accountability”, The Review of Economics and Statistics, 92 (2): 263-283. Reback, R., 2008, Teaching to the rating: School accountability and the distribution of student achievement, Journal of Public Economics, 92: 1394-1415. Schutz, Gabriela, Martin R. West, and Ludger Wobmann, 2007, “School Accountability, Autonomy, Choice, and the Equity of Student Achievement: International Evidence from PISA 2003”, OECD Education Working Papers, No. 14OECD. West, Martin R. and Paul E. Peterson, 2006, “The Efficacy of Choice Threats Within School Accountability Systems”, The Economic Journal, 116, ( 510): C46-C62. 15