Skip to content

Marc F. Bellemare Posts

Evaluating the Impact of Policies Using Regression Discontinuity Design, Part 2

I had a long post yesterday on regression discontinuity design (RDD), a statistical apparatus that allows identifying causal relationships even in the absence of randomization.

I split my discussion of RDD into two posts so as to respect my self-imposed rule #3 (“anything longer than 500 words, you split into two posts,” which constitutes an example of RDD in itself) but to make a long story short, the assumption made by RDD is that units of observation (e.g., children) immediately above and below some exogenously imposed threshold (e.g., the passing mark on an entrance exam for an elite school) are similar, so that comparing units immediately above and below that threshold allows estimating a causal effect (e.g., the causal effect of going to an elite school).

An RDD design is nice to have when eligibility for some treatment (e.g., going to an elite school) consists of a single threshold. Often, however, there will be multiple thresholds, which are aggregated into a single index without any clear idea as to what weight is given to each variable. So what are we to do in those cases?

Aid Bloggers’ Best Awards

Tom Murphy, who runs A View from the Cave and is one of the partners running the Development and Aid Workers News Service (DAWNS) Digest — which delivers daily news relevant to development straight your inbox every morning for a very small fee — has opened up nominations for the 2011 Aid Bloggers’ Best Awards (ABBAs).

You can vote here. In case you would like to nominate one of my posts for best post of 2011, my top posts since I started the blog in 2011 (which makes this blog eligible for best new blog) were:

  1. Methodological Convergence in the Social Sciences, in which I discussed the increasingly porous boundaries between disciplines in the social sciences.
  2. Seven Billion People on Earth: Enough with the Fear Mongering, in which I expressed a great deal of skepticism at neo-Malthusian ideas, and which was linked to by Andrew Sullivan.

Evaluating the Impact of Policies Using Regression Discontinuity Design, Part 1

Do students in smaller classes perform better than students in larger classes?

The answer might seem obvious. After all, students in smaller classes receive more attention from teachers, and so they should perform better.

We cannot know for sure, however, without looking at actual data on class size and student performance. In order to do so, we could collect data on student performance from various schools whose class sizes vary and look at whether students in smaller classes perform better.

But that wouldn’t be enough to determine whether smaller classes actually cause students to perform better. Correlation is not causation, and it could be the case that high-performing students are assigned to smaller classes composed of similar students. Thus, finding a correlation between class size and student performance would not be an indication that smaller classes cause students to perform better — only that school administrators want to put high-performing students in the same classes.

So how are we to know whether smaller classes actually cause students to perform better? One way could be to create classes of varying sizes (say, classes of 15, 30, 45, and 60 students) and randomly assign students to a given class size at the beginning of the year. Then, we could collect data on student performance on a standardized year-end exam and test whether average student performance is better in smaller than in bigger classes. Unfortunately, such a nice, clean experiment isn’t always feasible.