(It’s Spring Break here this week, so I am taking the week off from blogging to work to revise a few articles and begin working on new research projects. As a result, I am re-posting old posts that some new readers might have missed but which were very popular the first time I posted them. The following was initially posted on May 25, 2011.)
Last weekend, Nicholas Kristof published a column in the New York Times in which he praised the use of randomized controlled trials (RCTs) in development policy. In a fit of econ envy, Kristof even went so far as to confess that if he had to do it all over again, he would major in economics in college instead of political science.
As a result of Kristof’s column, however, the use of RCTs in development policy has come under a considerable amount of scrutiny in the development blogosphere.
On the (Mis)Use of Regression Analysis: Country Music and Suicide
This article assesses the link between country music and metropolitan suicide rates. Country music is hypothesized to nurture a suicidal mood through its concerns with problems common in the suicidal population, such as marital discord, alcohol abuse, and alienation from work. The results of a multiple regression analysis of 49 metropolitan areas show that the greater the airtime devoted to country music, the greater the white suicide rate. The effect is independent of divorce, southernness, poverty, and gun availability. The existence of a country music subculture is thought to reinforce the link between country music and suicide. Our model explains 51 percent of the variance in urban white suicide rates.
That’s the abstract of an article published in Social Forces — a top-10 journal in sociology — in 1992.
Before my snark gets me into trouble: Yes, I do realize that the article was published in 1992, back when most social science researchers only had a flimsy grasp of identification and causality. I also realize it would be foolish to impose on the authors of the above-referenced article the same standards of identification we impose upon ourselves today.
Yet, I cannot help but think that someone with a lesser of understanding of causality than the average reader of this blog is bound to eventually stumble upon the abstract, think “Hey, that totally makes sense!,” and run with it.
I’m sure there are also examples of such findings in other disciplines. If you know of any, please share.
(HT: Friend and former student Norma Padron, who is doing her PhD at Yale and has just launched a nice health economics blog.)