Closing the achievement gap in mathematics: evidence from a remedial program in Mexico City
Abstract
This paper evaluates the impact of an intervention targeted at marginalized low-performance students in public secondary schools in Mexico City. The program consisted in offering free additional math courses, taught by undergraduate students from some of the most prestigious Mexican universities, to the lowest performance students in a set of marginalized schools in Mexico City. We exploit the information available in all students’ (treated and not treated by the program) transcripts enrolled in participating and non-participating schools. Before the implementation of the program, participating students lagged behind non-participating ones by more than a half base point in their GPA (over 10). Using a difference-in-differences approach, we find that students participating in the program observed a higher increase in their school grades after the implementation of the program, and that the difference in grades between the two groups decreases over time. By the end of the school year (when the free extra courses had been offered, on average, for 10 weeks), participating students’ grades were not significantly lower than non-participating students’ grades. These results provide some evidence that short and low-cost interventions can have important effects on student achievement.
Keywords
Impact evaluation Remedial education programs Difference-in-differences estimatorsJEL Classification
C23 I21 I28 O151 Introduction
While the evidence on the efficacy of different programs aimed at increasing a population’s schooling accumulates, the evidence on the effectiveness of programs targeted at improving school quality is not as vast. Existing evidence has measured the impact of conditional transfer programs (Behrman et al. 2011), information provision (Jensen 2010), school construction (Duflo 2001), and other school resources, such as teachers (Banerjee et al. 2004) and text books (Glewwe et al. 2002) on school enrollment or schooling years.
However, studies finding that particular interventions increase school attendance or schooling do not always find evidence that these programs improve students’ performance on standardized tests. In other words, “students often seem not to learn anything in the additional days that they spend at school” (Banerjee et al. 2008). Such is the case of Mexico, where despite successful efforts to increase schooling (e.g., Oportunidades), children’s cognitive skills have arguably increased by very little or nothing (Behrman et al. 2011). In addition, interventions aimed at improving school inputs, such as text books and number of teachers, find very little or no effects on students’ performance (Banerjee et al. 2004; Glewwe et al. 2002).
Recently, more attention has been put into the design and evaluation of programs aimed at improving learning. Examples of such interventions include Conditional Cash Transfers programs like the one studied by (Barham et al. 2013) in Nicaragua, and pedagogical strategies aimed at better matching teaching to students’ learning needs, like the one studied in (Banerjee et al. 2008) and in this paper, specifically targeted at the worst performing students within classrooms. However, their effectiveness is likely to depend on their design and the setting in which they are put in place.
This paper evaluates the impact on students’ grades of a low-cost intervention in public secondary schools in Mexico City, which consisted in offering free additional math courses to students lagging behind their peers in marginalized low-income schools in Mexico City.
We exploit the information available in all students’ (treated and not treated by the program) transcripts enrolled in participating and non-participating schools. Before the implementation of the program, participating students lagged behind non-participating ones by more than a half base point of their GPA (over 10). Using a difference-in-differences approach, we find that students participating in the program observed a higher increase in their school grades after the implementation of the program, and that the difference in grades between the two groups decreased over time. By the end of the school year (when the free extra courses had been offered, on average, for 10 weeks), participating students’ grades were not significantly lower than non-participating students’ grades.
After accounting for the presence of (i) mean reversion, and (ii) differences in testing and grading among teachers, our results suggest that the impact on math grades associated with the program is positive and significant in the order of 0.21 and 0.26 standard deviations in the fourth and fifth partial exams.^{1} The estimated impact is similar in magnitude to other studies’ findings. For example, Banerjee et al. (2008) find that a similar intervention, the Balsakhi Program in India, increased participating students’ grades by 0.14 standard deviations in the first year.^{2}
The paper is organized as follows. Section 2 reviews the recent literature that analyzes similar interventions, describing their design and findings in detail. Section 3 describes the setting in which the program evaluated in this paper was put in place, and motivates the need for its evaluation. Section 4 describes the data and the empirical strategy used in this paper. Results are presented in Sect. 5. Section 6 concludes.
2 Literature
Despite the growing number of programs targeting underperforming students in different countries, there exists very little evidence on their effectiveness. Two recent exceptions, which analyze similar interventions to the one studied here, are Banerjee et al. (2008), and Lavy and Schlosser (2005).
Banerjee et al. (2008) evaluate a remedial education program that hired women to teach third and fourth grade students in India lagging behind their peers in basic literacy and math skills in small groups. The evaluation design for this intervention used a sample of 15,000 students in two Indian cities, to which the treatment was randomly assigned. The treatment consisted in offering additional courses taught by an instructor (typically a young woman who received 2 weeks of training for this purpose) to a subset of lower performing students from each of the treated classrooms. These additional courses lasted for 2 h a day, took place during school hours, and were taught to groups of 15–20 children. The courses covered basic material that children were supposed to have learned in first and second grade. Finally, as the one studied in this paper, the cost of this intervention was relatively low, since teachers were local personnel trained for 2 weeks and paid 15 dollars per month.
The authors look at the impact of the program on learning levels, measuring learning using annual pre-intervention tests administered during the first few weeks of the school year and post-intervention tests, administered at the end. They found that it increased average test scores of all children in treatment schools by 0.14 standard deviations, mostly due to large gains experienced by children at the bottom of the test-score distribution. Their results suggest that the remedial course program was more cost effective than hiring new teachers^{3} and than a computer-assisted learning program implemented at the same time in a similar geographic area.^{4}
Lavy and Schlosser (2005) evaluate the short-term effects of a remedial education program, which provided additional instruction to underperforming high school students in Israel. The program targeted tenth to twelfth graders in need of additional help to pass the matriculation exams. As a control, they use a comparison group of schools with similar characteristics to those treated (schools that enrolled later in the program), and apply a difference-in-differences empirical strategy.
The intervention consisted in individualized instruction in small study groups of up to five tenth, eleventh and twelfth graders. The main goal of these study groups was, through individualized instruction based on underperforming students’ needs, to increase their matriculation rate and enhance their scholastic and cognitive abilities, self-image, and leadership aptitudes. The participants were chosen by their teachers based on their perceived likelihood that each student could pass his or her matriculation exam.
The evaluation focuses on the first year of implementation of the program. 4,100 students were affected by the intervention (one-fifth of all students in treated schools).
The authors look for the effects of the program on matriculation status, which is a comparable outcome for 12th graders. They find that the program raised the school mean matriculation rate by 3.3% points, mainly through its impact on targeted participants, rejecting the existence of externalities on their untreated peers.
In conclusion, as summarized in Kremer et al. (2013), programs that reduce the costs of attending school and improve students’ health and the availability of information may have large impacts on school attendance, although not necessarily impacting their performance. Moreover, interventions that increase existing school inputs, such as teachers and textbooks, have been shown to be generally ineffective. However, programs that are tailored to each individual student’s learning needs, such as remedial programs, not only have been shown to have large impacts on students’ performance, but are also cost effective. Nevertheless, the effectiveness of programs of this kind relies on a correct design and the context of their implementation.
3 Mexico’s education system
Pre-college education in Mexico is divided into three main stages: primary or elementary school, secondary school, and high school. Private schools must cover the same curriculum as schools in the public system, and the content of this curriculum is exclusively designed by the Ministry of Public Education (Secretaría de Educación Pública, SEP) for the first nine grades (all of primary and secondary school). SEP shares the responsibility of designing the curriculum and regulating private education at the high school level with the public university system. Primary school lasts for 6 academic years, and is typically attended by children aged 6–12. Secondary school and high school last three academic years each. Unlike the United States, Mexico does not regulate the age at which children can legally drop out of school. Instead, although the law is not enforced, it is compulsory for all Mexicans to graduate from secondary school.
There exists a variety of different programs within both primary and secondary public education in Mexico. Primary schools can be “general” (the most common), indigenous (where courses are taught in indigenous languages), and community schools (which target the most isolated communities and where students from different grades generally share a classroom). In addition to general and community schools, secondary schools can also be technical, “tele-secundarias” (which teach the general program but through television and pre-recorded classes), and secondary schools for adults (offered to individuals aged 15 or older who have completed their primary education). After secondary school, there are 3 years of high school which must be completed to attend college.
Mexico’s education system. Descriptive Statistics (2010)
Percentage enrolled in school | ||
---|---|---|
All of Mexico | Mexico City | |
Children aged 6–11 | 100 | 100 |
Children aged 12–14 | 90 | 100 |
Children aged 15–17 | 60.3 | 86.7 |
Continuation rate* | ||
Primary to secondary | 97 | 99 |
Secondary to high school | 100+ | 100+ |
It is then clear that the largest drop in enrollment rates, particularly in Mexico City, takes place in the transition from secondary school to high school. However, perhaps surprisingly, from all those students graduating from the last grade of secondary school, nearly all of them enroll in high school, particularly in Mexico City. The drop in enrollment seems then primarily a consequence of students’ low performance in secondary school, which does not allow them to graduate and continue with their education. 16 % of secondary school students do not pass the grade in which they are enrolled. According to the standardized test applied by INEE, ENLACE, in 2010, 40 % of secondary school students had insufficient verbal skills and another 40 % just reached basic verbal skills’ levels; 53 % of secondary school students had an insufficient math skills and another 34.7 % just reached basic skills in math. According to the PISA 2009 test, 51 % of 15-year-old students in Mexico’s educational system performed below level 2 in mathematics (55 % in PISA 2012), and just 5 % ranked in the highest level (5). It seems then crucial to design and evaluate interventions aimed at improving students’ performance in secondary schools in Mexico, if increasing educational attainment is a desired goal.
4 The program
The program evaluated in this paper offers a remedial math course to low-performance students enrolled in the last grade of secondary school from marginalized schools in Mexico.
It was put in place by the representation of the SEP in Mexico City in collaboration with the Laboratory of Initiatives for Development (LID), a local NGO. It ran during the second half of the academic year 2009–2010, from April to June 2010 in 33 schools in 11 different delegaciones.^{5}
The remedial course was taught by a group of undergraduate students. There were, in total, 55 advisors from three of the most prestigious universities in the country: UNAM, ITAM, and UP. UNAM is a public university, while the other two are private. These advisors fulfilled their “social service” requirement by participating in this program. 480 h of social service activities (understood as activities beneficial to society, the State or a university), after covering 70 % of the undergraduate program credits, is a legal requirement for students to obtain a college degree in Mexico. There is a wide range of activities that undergraduates can do to fulfill this requirement, which go from being a research assistant, participating in reforestation campaigns, or taking a job for a social organization or government agency, with the latter being a common case. Employers are not legally obliged to offer any remuneration for students during their social service. However, generally, students do receive payments to cover their transportation needs (as was the case for the students participating in the remedial program). The program seems then easily replicable and, the extent to which it will be displacing social service activities, its opportunity cost seems potentially low.
Administrative data
Female | Male | Total | |
---|---|---|---|
Advisors | |||
Number of advisors | 29 | 26 | 55 |
UNAM | 10 | 7 | 17 |
ITAM | 14 | 16 | 30 |
UP | 5 | 3 | 8 |
Group size | |||
Initial size | 13.1 | 13.7 | 13.4 |
Final size | 12.0 | 13.0 | 12.4 |
Sessions | 11.2 | 7.7 | 9.9 |
Total hours | 27.6 | 31.2 | 29.2 |
Schools | |||
Type of secondary school | |||
General | 34 | ||
Technical | 21 | ||
Shift | |||
Morning school | 18 | ||
Afternoon school | 29 | ||
Both | 8 | ||
Geographical location | |||
North | 14 | ||
South | 7 | ||
East | 30 | ||
Center | 4 |
Prior to the intervention, the advisors were required to attend a brief but useful training session of 4 h in total given by the Mexican Academy of Science, where they had access to the mathematics syllabus at the secondary school level, a previous final exam, and were instructed to look at children’s notebooks and continuously ask the students for specific questions to regularly adapt the course’s content to the group’s needs. The intervention in each of the participating schools typically consisted of a meeting of two advisors with a group of up to 20 students, 2 days a week for 2 h, after school. This extra course focused on helping children develop the mathematical skills needed to improve their grades to pass the course.
The assignment of advisors to schools occurred as follows: (a) each of the volunteers stated the neighborhood to which they would prefer to be assigned. (b) coordinators located three schools in the said neighborhood, choosing the worst ranked in the ENLACE 2009 test who were willing to receive the program, (c) the students ranked them according to their preferences and finally, taking into account this information, coordinators assigned advisors to a single school.
Table 2 presents descriptive statistics for the outcome of the assignment of advisors to schools. 34 advisors were assigned to general secondary schools and 21 to technical secondary schools; 18 taught the remedial courses to morning shift students, and 29 to students in the afternoon shift. 8 advisors taught to students in both shifts. Finally, from a total of 55 advisors, 14 volunteered at schools located in northern Mexico City, 7 in the south, 30 in the east and 4 in the center.^{6}
The assignment of students to the remedial course was decided by the schools’ principals. The general guidelines suggested by the coordinators of the program were to identify students at risk of failing the school year, based on their performance in the first two partial exams that they had already taken during that academic year. The third partial exam had not yet been administered in any of the schools at the time of the assignment of students to the remedial course.
The intervention began after the third partial exam on a date that varies from school to school but in all cases before the fourth partial exam had been administered, and lasted until the end of the school year.
Summary statistics on math approval rates of 9th graders
Within treated schools | Between schools | All the schools in the sample | |||||
---|---|---|---|---|---|---|---|
Treatment | Control | Simple differences (1–2) | All in treated schools | All in non-treated schools | Simple differences (1–5) | ||
(1) | (2) | (3) | (4) | (5) | (6) | (7) | |
Pre-treatment | 0.559 [0.497] | 0.784 [0.412] | −0.225*** [0.017] | 0.758 [0.428] | 0.763 [0.425] | −0.209*** [0.016] | 0.762 [0.426] |
Post-treatment | 0.806 [0.396] | 0.830 [0.376] | −0.024 [0.015] | 0.827 [0.378] | 0.840 [0.366] | −0.032* [0.014] | 0.837 [0.370] |
Observations | 689 | 5,258 | – | 5,947 | 16,278 | – | 22,225 |
The second panel in Table 2 shows that dropout rates from the remedial course were low. At the beginning of the program, on average, 13.4 students attended the course. The average remedial course class size by the end of the school year was 12.4 students.
5 Evaluation design
For the evaluation of this program, we collected detailed information on the math grades obtained by a sample of participating and non-participating students in the five partial exams administered during the academic year. As the program only started after the third partial exam, we then have information on students’ performance before and after its implementation. There are two different groups of non-participating students in our sample: non-participating students in schools in which the program was put in place, and all students in a set of 60 schools (similar in observable characteristics to the treated ones), in which the remedial course was not offered.
In contrast with other studies listed above, the scores observed for each student correspond to their grades on exams designed and graded by their specific teachers. On one hand, this implies the possibility of high subjectivity on our performance measure. However, as long as this subjectivity is teacher specific and not correlated with the treatment within classrooms, we can evaluate to which extent treated students caught up with their untreated peers.
Summary statistics on average scores of 9th graders in math
Within treated schools | Between schools | All the schools in the sample | |||||
---|---|---|---|---|---|---|---|
Treatment | Control | Simple differences (1–2) | All in treated schools | All in non-treated schools | Simple differences (1–5) | ||
(1) | (2) | (3) | (4) | (5) | (6) | (7) | |
Pre-treatment | |||||||
Partial 1 | 6.296 [1.273] | 6.856 [1.424] | −0.560*** [0.057] | 6.797 [1.420] | 6.923 [1.533] | −0.619*** [0.058] | 6.904 [1.505] |
Partial 2 | 6.261 [1.190] | 6.918 [1.746] | −0.6572*** [0.069] | 6.866 [1.671] | 6.932 [1.756] | −0.647*** [0.068] | 6.911 [1.733] |
Partial 3 | 6.222 [1.213] | 6.635 [2.108] | −0.413*** [0.082] | 6.588 [2.029] | 6.807 [2.032] | −0.556*** [0.079] | 6.761 [2.034] |
Average pre-treatment | 6.260 [1.047] | 6.803 [1.480] | −0.544** [0.058] | 6.758 [1.427] | 6.908 [1.550] | −0.607*** [0.059] | 6.868 [1.506] |
Post-treatment | |||||||
Partial 4 | 6.511 [1.270] | 6.645 [2.426] | −0.134 [0.094] | 6.630 [2.322] | 6.751 [2.322] | −0.204** [0.090] | 6.728 [2.322] |
Partial 5 | 6.718 [1.560] | 6.695 [2.630] | 0.024 [0.103] | 6.698 [2.530] | 6.897 [2.516] | −0.139 [0.098] | 6.863 [2.522] |
Average post-treatment | 6.615 [1.209] | 6.670 [2.436] | −0.056 [0.094] | 6.694 [2.296] | 6.823 [2.304] | −0.171* [0.089] | 6.817 [2.311] |
Observations | 689 | 5,258 | – | 5,947 | 16,278 | – | 22,225 |
Column 4 shows the average score for all the 5,947 students in treatment schools, and Column 5 shows the average scores for all the 16,278 students from the 60 non-treated schools in our sample. Two facts are worth highlighting. First, non-treated schools show higher average scores than treated schools, but the difference between both does not seem to change considerably over time. As for the differences between treated and non-treated students within treated schools, when comparing treated students with all non-treated students in our sample, treated students reduce the distance in average grades from non-treated ones after the implementation of the program (Column 6).
These descriptive statistics suggest that treated students observe an important increase in their exam grades after the implementation of the program, which allows them to catch up with their peers by the last partial exam. However, they also show important differences in levels for average grades between treated and non-treated students. Determining if the closing of the grade gap is indeed a result of the intervention requires a more refined analysis, which we describe in what follows.
6 Estimation strategy
The coefficients estimated for the dummies for each partial exam will measure the average scores for the non-treated students in each of the five exams. The coefficients estimated for the effect of the interactions between the partial exams and the treatment variable measure the average differential in grades between students in the treatment and the control groups in each period.
Treated and non-treated students are likely to be different in terms of their scholastic achievement. The treatment was only introduced after the third period. Given this, if our identification strategy is correctly measuring the causal impact of the treatment on students’ scores, we would not expect to see any statistical difference in the coefficients for the interaction between the treatment and the period variables for the first three periods. As the constant term is excluded from the regressions, the coefficient on these three variables will simply capture the differences in grades between the treatment and control groups in the absence of the remedial courses.
The measured impact of the remedial courses will then consist of comparing the coefficient for the interaction between the treatment and period variables for the last two periods, with those of the first three periods. This estimation strategy allows us then to identify if the trends in exam grades were similar before the implementation of the remedial program for the treatment and control groups, and also evaluate if its effects increase or decrease between the fourth and fifth periods.
6.1 Mean reversion correction
As described above, the assignment of students to the remedial course was decided by the school’s principal, who followed guidelines to identify students at risk of failing the school year based on their performance on the first two partial exams. The problem with evaluating interventions that select the treatment and control groups based on previous test scores is that a single pre-program test scores represent noisy measures of students’ performance, due to error variance.
The reason is that there may be one-time events occurring during the exam such as a simple flu, variation in the ingested amount of sugar, or other distractions that may alter students’ performance. Hence, a student placing at the bottom or top of the class distribution may do so due to a transitory testing noise, and may thus not be indicative of her true performance.
If this is the case, as Chay et al. (2005) point out, if some students are assigned to the treatment due to strong negative shocks to their performance in the first and second partial exams, their pre-program grades will contain a strongly negative error. Unless errors are perfectly correlated over time, one would expect scores in subsequent partial exams to rise, even in the absence of the intervention. Thus, the measured test score gains from a difference-in-differences analysis, as the specification in Eq. (1) suggests, will reflect a combination of the true program effect and spurious mean reversion.
To correctly identify the impact of the program, we need to eliminate sources of spurious correlation between the change in partial exams’ grades and the grades obtained before the intervention.
If our estimation strategy is correct, given that the program only started after the third midterm exam, if the student in the treatment group experienced a temporary negative shock during this first partial exam, we would expect the coefficients for the interaction of the treatment dummy with period 2 and 3 dummies to reflect the real gap among groups (being statistically of the same magnitude for both periods) if there is any. And so, if the coefficients for the interaction of the treatment with period 4 and 5 dummies get reduced in comparison to that associated with the pre-program situation, then this reduction could be attributed to the program. This specification concretely allows us to relax the implicit assumption of Eq. (1), \( E\left({{u_{ijt}}} \right) = 0, \forall i, \forall j, \forall t \), and allows for potential transitory shocks suffered in period 1 leading to mean reversion.
6.2 Ability of the students to improve
Nonetheless, concerns related to mean reversion can still remain. The school’s principals assigned students to the treatment with more information than just their performance on the first partial exam. It is possible that their selection rule included in fact the observed trend in grades for the first two evaluations, for example. Therefore, they would tend to select students who worsen in the second exam compared to the first one and leave without treatment those who probably would apparently be able to improve on their own, given their noisy second partial grade. If this was the case, the estimated coefficients for the interaction between the dummies for the second and third partial exams and the treatment dummy are likely to differ from zero.
Now, the assumption would be that treated and non-treated students with similar grades in the first partial exam and similar changes in grades between the first and second exams were equally likely selected to receive the intervention. Moreover, there is no further assumption needed with respect to the existence of noise when measuring the pre-program performance of the students, because this specification controls for that, allowing the occurrence of transitory shocks.
The idea now is that apart from the possibility of having experienced a negative transitory shock during the first partial, it could also be the case that students in the treatment (or in the control) group experienced a negative (or positive) shock during the second exam correlated with the shock in the previous one. In this case, we would observe improvements in grades on further exams even in the absence of the program, caused by mean reversion. Therefore, we need to add to Eq. 2 the interactions between the score in the second partial exam and the dummies of partial exams 3–5.^{7} This specification allows us to further relax the assumption in Eq. (1), \( E\left({{u_{ijt}}} \right) = 0, \forall i, \forall j, \forall t \), and allows for potential transitory shocks suffered in periods 1 and 2 leading to mean reversion .
Also, note that adding \( Score_{i2} \) or \( (Score_{i2} - Score_{i1} ) \) to Eq. (2) solves the concern about mean reversion explained in the previous paragraph, since the difference \( (Score_{i2} - Score_{i1} ) \) is a linear transformation of \( Score_{i2} \). Therefore, the specification suggested by Eq. (3) not just controls for the ability of students to improve, but also for mean reversion in a more satisfactory way.
If this specification is correct, we would expect the coefficients for the interaction of the treatment and the third period dummy to be not significantly different from zero, as the program had not been implemented by that time. And the coefficients for the interaction of the treatment with periods 4 and 5 dummies would reflect the impact associated with the program.
6.3 Differences in testing and grading among teachers
An important difference between this paper and similar studies (listed above) is that the scores observed for each student correspond to their grades on exams designed and graded by their specific teachers and not to standardized tests. In this section, we analyze the advantages and disadvantages of considering this measure of students’ performance.
Standardized tests are uniform in subject matter, format, administration, and grading procedure across all test takers, while a course grade might depend on a particular teacher’s judgment. However, course grades and standardized tests both reflect students’ skills and knowledge, and are thus generally highly correlated (Willingham et al. 2002).
Trying to provide an explanation to why these two evaluation tools are not entirely interchangeable, recent literature has underlined aspects of personality that seem essential to earning strong course grades because of what is required of students to earn them. Almlund et al. (2011) analyze how the contribution of personality to performance varies among these two evaluation methods.
One key point is that despite the power of standardized achievement tests to predict later academic and occupational outcomes (for example in Kuncel and Hezlett (2007); Sackett, et al. (2008)), cumulative high school GPA predicts graduation from college much better than standardized test scores do (as shown by Bowen et al. (2009)). Similarly, high school GPA more powerfully predicts college rank-in-class (Bowen et al. 2009; Geiser and Santelices 2007).
Duckworth et al. (2012) compare the variance explained in standardized test scores and GPA at the end of the school year by self-control and intelligence measured at the beginning of the school year. In a sample of children, they found that fourth graders’ self-control was a stronger predictor of ninth graders’ GPA than fourth graders’ IQ. At the same time, fourth graders’ self-control was a weaker predictor of ninth graders’ standardized test scores than was fourth graders’ GPA. Similarly, Oliver et al. (2007) found that parent and self-reported ratings of distractibility at age 16 predicted high school and college GPA, but not SAT test scores.
Given these arguments, analyzing grades instead of standardized scores seems to be a better idea to measure the impact of the program according to the initial objective. Unfortunately, the possibility of high subjectivity on our performance measure still remains. However, as long as this subjectivity is teacher specific and not correlated with the treatment within classrooms, we can evaluate to what extent treated students caught up with their untreated peers.
Taking advantage of the panel structure of our data, which contains a measure of each student performance in five different periods, we can control by school level fixed effects. In this way, we relax the implicit assumption in every previous specification that \( E(s_{j} ) = 0, \forall j \), in words, we are capturing the differences in students’ scores driven by differences across school characteristics, such as the teacher’s specific preferences for testing and grading.
7 Results
Treatment effects on the average math scores within treated schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 6.863*** [0.028] | |||
Partial 2 | 6.945*** [0.028] | 2.108*** [0.126] | ||
Partial 3 | 6.663*** [0.027] | 2.390*** [0.126] | 0.991*** [0.128] | 0.992*** [0.128] |
Partial 4 | 6.679*** [0.027] | 2.253*** [0.126] | 0.802*** [0.128] | 0.810*** [0.128] |
Partial 5 | 6.729*** [0.028] | 2.505*** [0.126] | 0.945*** [0.128] | 0.952*** [0.128] |
Treatment * Partial 1 | −0.569*** [0.082] | |||
Treatment * Partial 2 | −0.678*** [0.082] | −0.278*** [0.079] | ||
Treatment * Partial 3 | −0.438*** [0.081] | −0.084 [0.079] | 0.100 [0.077] | 0.100 [0.077] |
Treatment * Partial 4 | −0.162** [0.081] | 0.205*** [0.079] | 0.396*** [0.077] | 0.393*** [0.077] |
Treatment * Partial 5 | −0.013 [0.081] | 0.337*** [0.079] | 0.543*** [0.077] | 0.540*** [0.077] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 29,410 | 23,528 | 17,646 | 17,646 |
R-squared | 0.92 | 0.93 | 0.93 | 0.93 |
Treatment effects on the average math scores for all the sample’s schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 6.923*** [0.014] | |||
Partial 2 | 6.931*** [0.039] | 2.026*** [0.059] | ||
Partial 3 | 6.807*** [0.023] | 1.989*** [0.062] | 0.608*** [0.059] | 0.612*** [0.060] |
Partial 4 | 6.751*** [0.013] | 1.809*** [0.059] | 0.457*** [0.059] | 0.463*** [0.060] |
Partial 5 | 6.897*** [0.014] | 2.317*** [0.055] | 0.951*** [0.056] | 0.963*** [0.057] |
Treatment * Partial 1 | −0.628*** [0.079] | |||
Treatment * Partial 2 | −0.664*** [0.078] | −0.219*** [0.073] | ||
Treatment * Partial 3 | −0.582*** [0.078] | −0.145** [0.072] | 0.005 [0.070] | 0.004 [0.070] |
Treatment * Partial 4 | −0.234*** [0.078] | 0.215*** [0.073] | 0.361*** [0.070] | 0.359*** [0.070] |
Treatment * Partial 5 | −0.180** [0.078] | 0.236*** [0.073] | 0.384*** [0.070] | 0.382*** [0.070] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 110,005 | 88,004 | 66,003 | 66,003 |
R-squared | 0.92 | 0.93 | 0.94 | 0.94 |
The results of estimating Eq. (1) are shown in Column 1 of Tables 5 and 6. As can be seen, the coefficients for the interaction between the treatment and partial variables are negative and significantly different from zero for the first three partials both when restricting the sample to all students in treated schools (Table 5) and when including all students in treated and non-treated schools (Table 6). The coefficient for the interaction between the treatment dummy and partial 4 is significantly lower in magnitude than those for the first three partial exams. The grade gap between the treated and non-treated students seems then to decrease after the implementation of the program. Perhaps more interestingly, the coefficient for the interaction between the treatment and period 5 dummies is close to zero and insignificant when the sample is restricted to students in treated schools (suggesting that the program might have completely closed the performance gap between the two groups after 3 months of its implementation). When including students in non-treated schools in the sample, this coefficient remains significantly negative, although still smaller in magnitude than that for partial 4.
F tests of differences on treatment effects across partial exams
Null hypothesis | Treated schools | All sample’s schools | ||
---|---|---|---|---|
F test | P value | F test | P value | |
Partial 1 = Partial 2 | 0.900 | 0.342 | 0.100 | 0.747 |
Partial 2 = Partial 3 | 4.330 | 0.037 | 0.550 | 0.460 |
Partial 3 = Partial 4 | 5.690 | 0.002 | 9.760 | 0.002 |
Partial 3 = Partial 5 | 13.550 | 0.000 | 13.050 | 0.000 |
Partial 4 = Partial 5 | 1.680 | 0.195 | 0.240 | 0.626 |
Observations | 29,400 | 110,005 |
The following three columns in both tables present the regression results for the specifications described in Eqs. (2)–(4), respectively. Column 2 compares scores in partial exams 2–5 of students from the treatment and control groups controlling for the score in the first partial exam. As can be seen, students in the treatment group, by the fifth partial exam, had experienced a 0.33 point increase in their grades relative to their non-treated peers (Table 5) and a 0.24 points increase relative to the whole sample of non-treated students. This change is roughly equivalent to half the grade gap between the groups in their first partial exam. However, it is worth noting that the coefficient on the interaction between the treatment dummy and the second period is, for both samples, negative and statistically different from zero, but a closing in the gap is still observed by the third period, when the program had not yet begun. Adding the interaction between the first score and the period dummies does not seem to fully solve the problem of mean reversion caused by the use of a noisy measure of the students’ pre-treatment performance.
The last two columns in Tables 5 and 6 include further controls measuring students’ performance in the first two partial exams. The results are quantitatively similar when excluding or including (Columns 3 and 4) the interaction between the grade in the first partial and the change in grades between the first and the second exam as controls. Once we include all the observable information that the principals had at the moment of selecting the treated students (the scores in partial 1 and partial 2, and the difference between them), the coefficient of the interaction between the treatment dummy and the score in the third partial exam is insignificantly different from zero. Students’ grades are then uncorrelated with the treatment before the implementation of the program. With this specification, the difference in grades between treated students and their non-treated peers (Table 5) decreases by 0.54 points by the fifth partial exam. This magnitude is very close to the grade gap in the first partial exam, suggesting that the grade gap between treated students and their peers was fully closed by the fifth partial exam. The same estimate using all non-treated students in our sample as the control group (Table 6) suggests that the grade gap was closed by 0.38 points by the fifth partial exam.
Treatment effects on the average math scores within treated schools, FE schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 6.859*** [0.208] | |||
Partial 2 | 6.941*** [0.165] | 1.726*** [0.273] | ||
Partial 3 | 6.658*** [0.076] | 2.007*** [0.348] | 0.638*** [0.204] | 0.637** [0.215] |
Partial 4 | 6.674*** [0.067] | 1.870*** [0.356 | 0.448*** [0.203] | 0.456** [0.134] |
Partial 5 | 6.724*** [0.101] | 2.123*** [0.360] | 0.591** [0.292] | 0.597** [0.202] |
Treatment * Partial 1 | −0.527** [0.205] | |||
Treatment * Partial 2 | −0.637*** [0.178] | −0.095 [0.108] | ||
Treatment * Partial 3 | −0.396* [0.203] | 0.099 [0.146] | 0.216 [0.115] | 0.215 [0.116] |
Treatment * Partial 4 | −0.121 [0.205] | 0.387*** [0.158] | 0.511*** [0.123] | 0.507*** [0.123] |
Treatment * Partial 5 | 0.029, 0.208 | 0.520*** [0.173] | 0.658*** [0.142] | 0.655*** [0.143] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends & grade in Partial 1 | No | No | No | Yes |
Observations | 29,410 | 23,528 | 17,646 | 17,646 |
R-squared | 0.92 | 0.93 | 0.94 | 0.94 |
Treatment effects on the average math scores for all the sample’s schools, FE schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 6.919*** [0.036] | |||
Partial 2 | 6.927*** [0.033] | 1.929*** [0.117] | ||
Partial 3 | 6.803*** [0.046] | 1.891*** [0.122] | 0.151*** [0.071] | 0.587** [0.071] |
Partial 4 | 6.747*** [0.058] | 1.711*** [0.141] | 0.435*** [0.094] | 0.438*** [0.093] |
Partial 5 | 6.893*** [0.065] | 2.220*** [0.170] | 0.494*** [0.103] | 0.937*** [0.103] |
Treatment * Partial 1 | −0.501** [0.212] | |||
Treatment * Partial 2 | −0.537*** [0.178] | −0.008 [0.137] | ||
Treatment * Partial 3 | −0.455** [0.187] | 0.067 [0.134] | 0.221* [0.123] | 0.218* [0.123] |
Treatment * Partial 4 | −0.107 [0.189] | 0.426*** [0.136] | 0.577*** [0.121] | 0.574*** [0.121] |
Treatment * Partial 5 | −0.053 [0.183] | 0.447*** [0.155] | 0.599*** [0.143] | 0.596*** [0.143] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 110,005 | 88,004 | 66,003 | 66,003 |
R-squared | 0.92 | 0.93 | 0.94 | 0.94 |
Treatment effects on the average math scores of boys within treated schools, FE schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 6.691** [0.191] | |||
Partial 2 | 6.723*** [0.097] | 1.628*** [0.338] | ||
Partial 3 | 6.412*** [0.165] | 1.854*** [0.273] | 0.476** [0.205] | 0.487** [0.194] |
Partial 4 | 6.396*** [0.168] | 1.451*** [0.158] | 0.272** [0.166] | 0.295** [0.167] |
Partial 5 | 6.413*** [0.261] | 1.535*** [0.267] | 0.235*** [0.271] | 0.253*** [0.269] |
Treatment * Partial 1 | −0.567*** [0.196] | |||
Treatment * Partial 2 | −0.590*** [0.189] | −0.025 [0.144] | ||
Treatment * Partial 3 | −0.298** [0.064] | 0.219 [0.191] | 0.246 [0.144] | 0.242 [0.143] |
Treatment * Partial 4 | 0.04 [0.225] | 0.572** [0.214] | 0.602*** [0.173] | 0.596*** [0.171] |
Treatment * Partial 5 | 0.167 [0.210] | 0.693*** [0.188] | 0.735*** [0.155] | 0.730*** [0.156] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 15,042 | 12,034 | 9,026 | 9,026 |
R-squared | 0.564 | 0.568 | 0.81 | 0.81 |
Treatment effects on the average math scores of girls within treated schools, FE schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 7.033*** [0.208] | 2.214 | ||
Partial 2 | 7.167*** [0.162] | 1.904*** [0.393] | ||
Partial 3 | 6.912*** [0.086] | 2.286*** [0.277] | 0.856*** [0.265] | 0.878*** [0.267] |
Partial 4 | 6.963*** [0.068] | 2.214*** [0.250] | 0.703*** [0.225] | 0.703** [0.226] |
Partial 5 | 7.044** [0.105] | 2.684*** [0.472] | 1.085*** [0.367] | 1.085** [0.367] |
Treatment * Partial 1 | −0.472** [0.224] | |||
Treatment * Partial 2 | −0.678*** [0.182] | −0.177 [0.095] | ||
Treatment * Partial 3 | −0.496** [0.226] | −0.038 [0.138] | 0.167 [0.125] | 0.169 [0.126] |
Treatment * Partial 4 | −0.291 [0.228] | 0.175* [0.089] | 0.392*** [0.115] | 0.391*** [0.116] |
Treatment * Partial 5 | −0.106 [0.239] | 0.333** [0.104] | 0.563*** [0.176] | 0.563*** [0.177] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 14,368 | 11,494 | 8,620 | 8,620 |
R-squared | 0.672 | 0.676 | 0.894 | 0.894 |
When including further controls and school fixed effects, the difference in grades between treated students and their non-treated peers decreases by 0.73 points by the fifth partial exam for boys, and by 0.56 points for girls.
Treatment effects on average math standardized scores within treated schools, FE schools
(1) | (2) | (3) | (4) | |
---|---|---|---|---|
Partial 1 | 0.000 [0.088] | |||
Partial 2 | 0.004 [0.070] | 0.001 [0.064] | ||
Partial 3 | 0.003 [0.092] | 0.000 [0.021] | −0.001 [0.030] | 0.001 [0.038] |
Partial 4 | 0.003 [0.084] | 0.001 [0.024] | 0.000 [0.021] | −0.002 [0.031] |
Partial 5 | 0.002 [0.024] | −0.000 [0.002] | 0.000 [0.030] | 0.000 [0.017] |
Treatment * Partial 1 | −0.379*** [0.134] | |||
Treatment * Partial 2 | −0.367*** [0.099] | −0.073 [0.055] | ||
Treatment * Partial 3 | −0.186* [0.097] | 0.047 [0.069] | 0.099* [0.052] | 0.099* [0.053] |
Treatment * Partial 4 | −0.046 [0.086] | 0.171** [0.067] | 0.214*** [0.050] | 0.213*** [0.050] |
Treatment * Partial 5 | 0.019 [0.081] | 0.218*** [0.068] | 0.260*** [0.054] | 0.260*** [0.055] |
Controls by Partial 1 grade | No | Yes | Yes | Yes |
Controls by trends in grades before the treatment | No | No | Yes | Yes |
Controls by the interaction of trends and grade in Partial 1 | No | No | No | Yes |
Observations | 29,625 | 23,700 | 17,775 | 17,775 |
R-squared | 0.059 | 0.291 | 0.420 | 0.420 |
Cost effectiveness analysis of educational programs
Treatment | Place | Grades | Duration | Impact | Cost per s.d. | Authors |
---|---|---|---|---|---|---|
Cash transfers^{a} | Mexico | 10th- | 5 years | 0 | ∞ | Berhman et al. (2007) |
Teacher incentives | Mexico | 3rd–6th | 1 year | 0.08 | 108 | McEwan and Santibañez (2005) |
Computers in classrooms^{a} | Colombia | 3rd–11th | 2 years | 0 | ∞ | Barrera-Osorio and Linden (2009) |
CARES vouchers | Colombia | 6th | 3 years | 0.2 | 40 | Angrist et al. (2002) |
Math tutoring^{a} | Mexico | 9th | 3 month | 0.16 | 13.3 | Gutiérrez and Rodrigo (this paper) |
Reading tutoring^{b} | Chile | 4th | 3 month | 0.15 | 21.4 | Cabezas et al. (2011) |
Textbooks | Kenya | 3rd–8th | 2 years | 0 | ∞ | Glewwe et al. (2002) |
Remedial sessions^{b} | India | 3rd–4th | 1 year | 0.14 | 3.3 | Banerjee et al. (2004) |
Camera monitoring^{b} | India | 1st–4rd | 3 years | 0.17 | 11.7 | Duflo et al. (2012) |
8 Conclusions
This paper presents the results of the evaluation of a low-cost intervention in public secondary schools in Mexico City, which consisted in offering free additional math courses to students lagging behind their peers in marginalized low-income schools in Mexico City.
We exploit the information available in all students’ (treated and not treated by the program) transcripts enrolled in participating and non-participating schools. Before the implementation of the program, participating students lagged behind non-participating ones by more than half a base point in their GPA (over 10). As the program was not randomly assigned, we suggest a difference-in-differences strategy and, increasing the number of controls used, we discuss the validity of our method. Regardless of the control variables included in the analysis, we find that students participating in the program observed a higher increase in their school grades after the implementation of the program, and that the difference in grades between the two groups decreases over time. By the end of the school year (when the free extra courses had been offered, on average, for 10 weeks), participating students’ grades were not significantly lower than non-participating students’ grades. Nonetheless, without any controls, the closing of the grade gap between treated and untreated students seems to start before the program’s implementation. The differences in scores before the implementation of the program between treated and non-treated students remain significant when controlling for their grade in the first partial exam. However, these differences disappear when including the information on students’ grades in the first two exams.
We then conclude that this paper shows evidence that interventions of this kind can, at a relatively low cost, contribute to increase underperforming students’ exam grades in the short run.
Footnotes
- 1.
In Mexico, grading throughout the school year is based on five evaluations (one every 2 months), called “exámenes parciales”. The final GPA is calculated as the simple average of these five exams’ grades. Throughout the paper, we call each of these evaluations a “partial exam”.
- 2.
The Balsakhi program provided schools with a teacher (local personnel paid 15 dollars a month) to work for 2 h a day during the school year, with groups of 15–20 children in the third and fourth grades identified as falling behind their peers. Banerjee et al. (2008) estimate improvements in average test scores of 0.14 standard deviations in the first year, and 0.28 in the second.
- 3.
A program reducing class size appears to have had little or no impact on test scores, and the remedial course program costs 2.25 dollars per student per year.
- 4.
The computer-assisted learning program costs approximately 15.18 dollars per student per year, including the cost of computers, assuming a 5-year depreciation.
- 5.
Mexico City is divided into 16 delegaciones, which are the smallest administrative entities in the city.
- 6.
The odd numbers reported in each school category are due to a few cases in which one single advisor was assigned to a class, or to more than one school.
- 7.Specifically, we would need to estimate the following equation:$$ \begin{array}{l} Scor{e_{it}} = \mathop \sum \limits_{t = 2}^5 {\emptyset_t}Partia{l_t} + \mathop \sum \limits_{t = 2}^5 {\beta_t}Treatmen{t_i}*Partia{l_t}\\ + \mathop \sum \limits_{t = 2}^5 {\gamma_t}Scor{e_{i1}}*Partia{l_t} + \mathop \sum \limits_{t = 2}^5 {\varphi_t}Scor{e_{i2}}*Partia{l_t} + {e_{it}} \end{array} $$
- 8.
This exercise was made at the school level and not at class level as was suggested in the Evaluation Design section, due to data limitations (we count with school identifier but not class identifier). Nevertheless, according with the educational authorities, it is always the same teacher who teaches the same subject to all classes in a school (but rare exceptions). In which case, school or class fixed effects estimations would equally correct for differences in testing and grading among teachers.
Notes
Acknowledgments
We thank the Ministry of Education (SEP) in Mexico City for providing the data and the Laboratory of Initiatives for Development (LID) for the commitment with the evaluation. We are also grateful to Jere Behrman, Kensuke Teshima, and participants of the XVI meeting of the LACEA/IADB/WB/UNDP Research Network on Inequality and Poverty (NIP) Conference for valuable comments and suggestions. Emilio Gutierrez is grateful for support from the Asociación Mexicana de Cultura.
References
- Almlund M, Lee A, Heckman J (2011) Personality psychology and economics. Handbook of economics of education, chapter 1, vol 4Google Scholar
- Angrist J, Bettinger E, Bloom E, King E, Kremer M (2002) Vouchers for private schooling in colombia: evidence from a randomized natural experiment. Am Econ Rev 92(5):1535–1558CrossRefGoogle Scholar
- Banerjee A, Jacob S, Kremer M (2004) Promoting school participation in rural Rajasthan: results from some prospective trials, MIT Department of Economics, Working paperGoogle Scholar
- Banerjee A, Cole S, Duflo E, Linden L (2008) Remedying education: evidence from two randomized experiments in India. Quart J Econ, MIT Press 122(3):1235–1264CrossRefGoogle Scholar
- Barham, Tania, Karen Macours, and John A. Maluccio (2013) “More Schooling and More Learnings? Effects of a Three-Year Conditional Cash Transfer Program in Nicaragua after 10 years”. IDB Working Paper Series No. IDB-WP-432Google Scholar
- Barrera-Osorio F, Linden LL (2009) The use and misuse of computers in education: evidence from a randomized experiment in Colombia. World bank policy research working paper series, vol 2009Google Scholar
- Behrman JR, Parker SW, Todd PE (2011) Do conditional cash transfers for schooling generate lasting benefits? A five-year followup of PROGRESA/Oportunidades. J Hum Resour 46(1):93–122CrossRefGoogle Scholar
- Bowen William G, Chingos Mathew, McPherson Michael (2009) Test scores and high school grades as predictors. Crossing the Finish Line: Completing College at America’s Public Universities”. Princeton University Press, Princeton, NJ, pp 112–133Google Scholar
- Cabezas, Verónica, José I. Cuesta, and Francisco A. Gallego (2011) “Effects of Short Term Tutoring on Cognitive and Non Cognitive Skills: Evidence from a Randomized Evaluation in Chile”, Instituto de Economía, P. Universidad Católica de Chile, Working paperGoogle Scholar
- Chay Kenneth Y, McEwan Patrick J, Urquiola Miguel (2005) The central Role of Noise in Evaluating Interventions that Use Test Scores to Rank Schools. Am Econ Rev 95(4):1237–1258CrossRefGoogle Scholar
- Duckworth AL, Quinn PD, Tsukayama E (2012) What no child left behind leaves behind: the roles of IQ and self-control in predicting standardized achievement test scores and report card grades. J Educ Psycol 104(2):439–451CrossRefGoogle Scholar
- Duflo Esther (2001) Schooling and labor market consequences of school construction in Indonesia: evidence from an unusual policy experiment. Am Econ Rev Am Econ Assoc 91(4):795–813CrossRefGoogle Scholar
- Duflo E, Hanna R, Rya SP (2012) Incentives work: getting teachers to come to school. Am Econ Rev 102(4):1241–1278Google Scholar
- Geiser S, Santelices MV (2007) Validity of high school grades in predicting student success beyond the freshman year: high-school record VS. standardized tests as indicators of four-year college outcomes. Center for Studies in Higher Education at the University of California, Berkeley CSHEGoogle Scholar
- Glewwe P, Kremer M, Moulin S (2002) Textbooks and test scores: evidence from a prospective evaluation in Kenya. BREAD Working Paper, Cambridge, MAGoogle Scholar
- INEE (2012) Panorama Educativo de México 2012 Indicadores del Sistema Educativo Nacional. México, Instituto Nacional para la Evaluación de la Educación (INEE)Google Scholar
- Jensen R (2010) The perceived returns to education and the demand for schooling. Q J Econ 125(2):515–548CrossRefGoogle Scholar
- Kremer M, Brannen C, Glennerster R (2013) The challenge of education and learning in the developing world. Science 340:297–300CrossRefGoogle Scholar
- Kuncel NR, Hezlett SA (2007) Standardized tests predict graduate students’ success. Science 315:1080–1081CrossRefGoogle Scholar
- Lavy V, Schlosser A (2005) Targeted remedial education for underperforming teenagers: cost and benefits. J Labor Econ XXIII:839–874Google Scholar
- McEwan PJ, Santibañez L (2005) Teacher incentives and student achievement: evidence from a large-scale reform in Mexico (unpublished paper)Google Scholar
- Oliver PH, Guerin DW, Gottfried AW (2007) Temperamental task orientation: relation to high school and college educational accomplishments. Learn Ind Differ 17(3):220–230CrossRefGoogle Scholar
- Sackett PR, Borneman MJ, Connelly BS (2008) High stakes testing in higher education and employment: appraising the evidence for validity and fairness. Am Psychol 63(4):215–227CrossRefGoogle Scholar
- Willingham WW, Pollack JM, Lewis C (2002) Grades and test scores: accounting for observed differences. J Educ Meas 39(1):1–37CrossRefGoogle Scholar
Copyright information
This article is published under license to BioMed Central Ltd. Open AccessThis article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.