{"id":9906,"date":"2014-02-13T05:00:17","date_gmt":"2014-02-13T10:00:17","guid":{"rendered":"http:\/\/marcfbellemare.com\/wordpress\/?p=9906"},"modified":"2014-02-08T11:21:22","modified_gmt":"2014-02-08T16:21:22","slug":"goodness-of-fit-in-binary-choice-models-technical","status":"publish","type":"post","link":"https:\/\/marcfbellemare.com\/wordpress\/9906","title":{"rendered":"Goodness of Fit in Binary Choice Models [Technical]"},"content":{"rendered":"<p>In econometrics, goodness-of-fit measures tell us what percentage of the variation in a dependent variable is explained by the explanatory variables. If you&#8217;ve ever taken a statistics class, you are almost surely familiar with the R-square measure. In a regression of, say the logarithm of wage on age, gender, and education level, the R-square is simply the fraction of the total variation in wage that is explained by variation in age, gender, and education level.<\/p>\n<p>Given the foregoing, you&#8217;d think R-square is a great measure, right? I mean, it tells you how much of the variation in\u00a0<em>Y<\/em> all of your\u00a0<em>X<\/em>&#8216;s explain! Yeah, no&#8230; R-square is actually not all that interesting, because you can thrown in any variable on the right-hand side &#8212; for example, the color of one&#8217;s underwear in the log wage regression above &#8212; and R-square can only increase, because there is bound to be a (spurious) correlation between the color of one&#8217;s underwear and one&#8217;s wage. Even the adjusted R-square, which corrects for how many variables there are in\u00a0<em>X<\/em>, isn&#8217;t that great, since that correction is somewhat arbitrary.<!--more--><\/p>\n<p>With binary outcomes (i.e., when\u00a0<em>Y<\/em>\u00a0can only be equal to one or zero, such as in questions answered by &#8220;Yes&#8221; or &#8220;No&#8221;), people often like to use the percentage of ones and zeroes correctly predicted, and report that as a measure of goodness-of-fit. Kennedy, in his classic econometric intuition-building treatise, argued that this was not a very good measure:<\/p>\n<blockquote><p>It is tempting to use the percentage of correct predictions as a measure of goodness of fit. This\u00a0temptation should be resisted: a na\u00efve predictor, for example that every y = 1, could do well on\u00a0this criterion. A better measure along these lines is the sum of the fraction of zeros correctly\u00a0predicted plus the fraction of ones correctly predicted, a number which should exceed unity if\u00a0the prediction method is of value. See <a title=\"McIntosh and Dorfman (AJAE, 1992)\" href=\"http:\/\/ajae.oxfordjournals.org\/content\/74\/1\/209.short\" target=\"_blank\">McIntosh and Dorfman (1992)<\/a>.<\/p><\/blockquote>\n<p>This could use a bit of explanation: Suppose we have\u00a0<em>Y<\/em> = (0, 1, 1, 1, 1, 1, 1, 1), and we have a vector of predicted values of\u00a0<em>Y\u00a0<\/em>be (0, 1, 1, 0, 1, 1, 1, 0). The usual percentage-of-correct-predictions measure would be 0.75, since 75% of observations are correctly predicted, or 6 out of 8. But one can do even better by guessing &#8220;all ones.&#8221; Indeed, if I were to guess all ones, I&#8217;d get 87.5% goodness of fit, or 7 out of 8.<\/p>\n<p>What McIntosh and Dorfman (1992) suggested instead was to add up (i) the fraction of correctly predicted zeros (in my example, 100%) and (ii) the fraction of correctly predicted ones (in my example, 50%). In my example, then, the total McIntosh-Dorfman goodness-of-fit measure would be 1.5 which, by McIntosh and Dorfman criterion standards, would be deemed a good fit, since it exceeds 1.<\/p>\n<p>Now, if your reaction to the above was this:<\/p>\n<p><a href=\"http:\/\/marcfbellemare.com\/wordpress\/wp-content\/uploads\/2014\/02\/PeterGriffin.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-medium wp-image-9908\" alt=\"PeterGriffin\" src=\"http:\/\/marcfbellemare.com\/wordpress\/wp-content\/uploads\/2014\/02\/PeterGriffin-580x440.jpg\" width=\"580\" height=\"440\" srcset=\"https:\/\/marcfbellemare.com\/wordpress\/wp-content\/uploads\/2014\/02\/PeterGriffin-580x440.jpg 580w, https:\/\/marcfbellemare.com\/wordpress\/wp-content\/uploads\/2014\/02\/PeterGriffin.jpg 750w\" sizes=\"auto, (max-width: 580px) 100vw, 580px\" \/><\/a><\/p>\n<p>Consider the following example from a referee report I received on <a title=\"Bellemare (WD, 2012)\" href=\"http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0305750X11003111\" target=\"_blank\">my 2012 <em>World Development<\/em> article about the welfare impacts of participation in contract farming<\/a>.<\/p>\n<p>In that referee report, the reviewer was faulting me for low pseudo R-square measures on a probit, and suggested that I report the percentage of correct predictions. Notwithstanding the fact that that pseudo R-square measures are pretty bad (see <a title=\"Estrella (JBES, 1998)\" href=\"http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/07350015.1998.10524753#.UvZYTfldWcU\" target=\"_blank\">Estrella, 1998<\/a> on that point), I responded with the Kennedy quote above, and in the published version of my paper, in table 5, I actually report three measures: the pseudo R-square (0.081), the percentage of correct predictions (0.63), and the Dorfman-McIntosh measure (1.29). Note that although the percentage of correct predictions and the McIntosh-Dorfman measure are consistent with one another (if I assume I predicted 63% of both ones and zeros correctly, I get a McIntosh-Dorfman of 0.126), the pseudo R-square tells me that only 8 percent of the variance in the dependent variable is explained by my left-hand side variable, which does strike me as misleading in this case.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>In econometrics, goodness-of-fit measures tell us what percentage of the variation in a dependent variable is explained by the explanatory variables. If you&#8217;ve ever taken<\/p>\n<div class=\"more-link-wrapper\"><a class=\"more-link\" href=\"https:\/\/marcfbellemare.com\/wordpress\/9906\">Continue reading<span class=\"screen-reader-text\">Goodness of Fit in Binary Choice Models [Technical]<\/span><\/a><\/div>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-9906","post","type-post","status-publish","format-standard","hentry","category-uncategorized","entry"],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p1gPg8-2zM","_links":{"self":[{"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/posts\/9906","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/comments?post=9906"}],"version-history":[{"count":4,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/posts\/9906\/revisions"}],"predecessor-version":[{"id":9911,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/posts\/9906\/revisions\/9911"}],"wp:attachment":[{"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/media?parent=9906"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/categories?post=9906"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/marcfbellemare.com\/wordpress\/wp-json\/wp\/v2\/tags?post=9906"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}