{"id":40802236191,"date":"2010-06-18T16:12:00","date_gmt":"2010-06-18T20:12:00","guid":{"rendered":"https:\/\/www.monmouth.edu\/polling-institute\/2010\/06\/18\/are-nate-silvers-pollster-ratings-done-right\/"},"modified":"2023-07-05T10:34:18","modified_gmt":"2023-07-05T14:34:18","slug":"are-nate-silvers-pollster-ratings-done-right","status":"publish","type":"post","link":"https:\/\/www.monmouth.edu\/polling-institute\/2010\/06\/18\/are-nate-silvers-pollster-ratings-done-right\/","title":{"rendered":"Are Nate Silver\u2019s Pollster Ratings \u201cDone Right\u201d"},"content":{"rendered":"<p><em><strong>This originally appeared as a guest column on <a href=\"http:\/\/www.pollster.com\" target=\"_blank\" rel=\"noopener noreferrer\">Pollster.com<\/a>.<\/strong><\/em><\/p>\n<p>The motto of Nate Silver\u2019s website, <a href=\"http:\/\/www.fivethirtyeight.com\/\">www.fiverthirtyeight.com<\/a>, is \u201cPolitics Done Right.\u201d I\u2019m not sure that his latest round of pollster ratings lives up to that moniker.<\/p>\n<p>As most poll followers know, Nate shot to fame during the 2008 election, taking the statistical skills he developed to predict baseball outcomes and applying them to election forecasting. His approach was pretty accurate in that presidential race (although it\u2019s worth noting that <a href=\"http:\/\/election.princeton.edu\/2008\/11\/11\/post-election-evaluation-part-2\" target=\"_blank\" rel=\"noopener noreferrer\">other poll aggregators were similarly accurate<\/a>).<\/p>\n<p>Nate recently released a new set of pollster ratings that has raised some concerns among the polling community.<\/p>\n<p>First, there are some questions about the accuracy of the underlying data he uses. Nate claims to have culled his results from 10 different sources, but he seems to not to have cross-checked those sources or searched original sources for verification.<\/p>\n<p>I asked for Monmouth University\u2019s poll data and found errors in the 17 poll entries he attributes to us \u2013 including six polls that were actually conducted by another pollster before we partnered with the New Jersey Gannett newspapers, one omitted poll that should have been included, two incorrect election results, and one incorrect candidate margin. <i>[Nate emailed me that he will correct these errors in his update later this summer.]<\/i><\/p>\n<p>Mark Blumenthal also noted errors and omissions in the data used to arrive at Research2000\u2019s rating. I found evidence that suggest these errors may be fairly widespread.<\/p>\n<p>In the case of prolific pollsters, like Research2000, these errors may not have a major impact on the ratings. But just one or two database errors could significantly affect the vast majority of pollsters with relatively limited track records \u2013 such as the 157 pollsters out of 262 pollsters on his list who have fewer than 5 polls to their credit.<\/p>\n<p>Some observers have called on Nate to demonstrate transparency in his own methods by releasing that database. Nate has refused to do this (with a dubious rationale that the information may be proprietary) &#8211; but he does now have a process in place for pollsters to verify their own data. <i>[If you do, make sure to check the accuracy of the actual election results as well.]<\/i><\/p>\n<p>I\u2019d be interested to see how many other pollsters find errors in their data. But the issue that has really generated buzz in our field is Nate\u2019s claim that pollsters who either were members of the National Council on Public Polls or had committed to the <a href=\"https:\/\/aapor.org\/\" target=\"_blank\" rel=\"noopener noreferrer\">American Association for Public Opinion Research (AAPOR) Transparency Initiative<\/a> by June 1, 2010 exhibit superior polling performance. For these pollsters, he awards a very sizable \u201ctransparency bonus\u201d in his latest ratings.<\/p>\n<p>One of the obvious problems with his use of the bonus is that the June 1 cut-off is arbitrary. Those pollsters who signed onto the initiative by June 1, 2010 were either involved in the planning or happened to attend the AAPOR national conference in May. A general call to support the initiative did not go out until June 7 \u2013 the day after Nate\u2019s ratings were published.<\/p>\n<p>Thus, the theoretical claim regarding a transparency bonus is at least partially dependent on there also being a relationship between pollster accuracy <i>and<\/i> AAPOR conference attendance. Others have remarked on the apparent arbitrariness of this \u201ctransparency bonus\u201d cutoff date Nate claims that regardless of how a pollster made it onto the list, there is statistical evidence these pollsters are simply better at election forecasting. I don\u2019t quite see it.<\/p>\n<p>His methodology statement includes a regression analysis of pollster ratings that is presented as evidence for using the bonus.<\/p>\n<p>The problem is that even in this equation, the transparency score just misses most researcher\u2019s threshold for being significant (p&lt;.05). More to the point, his model \u2013 using dummy variables to identify \u201ctransparent\u201d pollsters, partisan pollsters, and internet pollsters \u2013 is incomplete. The adjusted R-square is .03. In other words, 3% of total variance in pollster raw scores (i.e. error) is predicted by the model.<\/p>\n<p>Interestingly of the three variables \u2013 transparency, partisan, and internet \u2013 only partisan polling shows a significant relationship. He decided to calculate different benchmarks that award transparent polls and penalize internet polls (even though the latter was based on only 4 cases and not statistically significant). And oddly, he does not treat partisan pollsters any differently than other pollsters, even though this was the only variable with a significant relationship to rawscore.<\/p>\n<p>I decided to look at this another way, using a simple means analysis. The average error among all pollsters is +.54 (positive error is bad, negative is good). Among \u201ctransparent\u201d pollsters it is -.63 (se=.23) and among other pollsters it is +.68 (se=.28).<\/p>\n<p>But let\u2019s isolate the more prolific pollsters, say the 63 organizations with at least 10 polls to their names who are included in Nate\u2019s first chart. Among these pollsters, the 19 \u201ctransparent\u201d ones have an average score of -.32 (se=.23) and the other 44 pollsters average +.03 (se=.17). The difference is not so stark now.<\/p>\n<p>Firms with fewer than 10 polls to their credit have an average error score of -1.38 (se=.73) if they are \u201ctransparent\u201d (all 8 of them) and a mean of +.83 (se=.28) if they are not. That\u2019s a much larger difference.<\/p>\n<p>I also ran some ANOVA tests for the effect of the transparency variable on pollster raw scores for various levels of polling output (e.g. pollsters with more than 10 polls, pollsters with only 1 or 2 polls, etc.). The F values for this test range from only 1.2 to 3.6, and none were significant at p&lt;.05. In other words, there is more error variance within the two separate groups of transparent versus non-transparent pollsters than there is between the two groups.<\/p>\n<p>I can only surmise that the barely significant relationship between the arbitrary transparency designation and polling accuracy is pointing to other more significant factors, including pollster output.<\/p>\n<p>Consider this &#8211; 70% of \u201ctransparent\u201d pollsters on Nate\u2019s list are have 10 or more polls to their credit, whereas only 19% of the \u201cnon-transparent\u201d ones do. In other words, Nate\u2019s \u201cbonus\u201d is actually a sizable penalty levied against more prolific pollsters in the latter group. \u201cNon-transparent pollsters happen to be affiliated with a large number of organizations with only a handful of polls to their name \u2013 i.e. pollsters who are prone to greater error.<\/p>\n<p>For comparison, re-ran Nate\u2019s PIE (Pollster Introduced Error) calculation using a level playing field for all 262 pollsters on the list. I set the error mean at +.50 (which is approximately the mean error among all pollsters).<\/p>\n<p>Comparing the relative pollster ranking between the two lists produced some intriguing results. The vast majority of pollster ranks (175) did not change by more than 10 spots on the table. Another 67 had rank changes between 11 to 40 spots on the two lists; 11 shifted by 41 to 100 spots, and 9 pollsters gained more than 100 spots in the rankings because of the transparency bonus. Of this latter group, only 2 of the 9 had more than 15 polls recorded in the database.<\/p>\n<p>Nate says that the main purpose of his project is not to rate pollsters\u2019 past performance but to determine probable accuracy going forward. But one wonders if he needs to go this particular route to get there. Other aggregators use less elaborate methods \u2013 including straightforward mean scores \u2013 and seem to be just as accurate.<\/p>\n<p>His methodology statement is about 4,800 words (with 18 footnotes). It reminds me of a lot of the techies I have worked with over the years \u2013 the kind of person who will make three left turns to go right.<\/p>\n<p>This time I think Nate may have taken one left turn to many. We\u2019ll know in November.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>This originally appeared as a guest column on Pollster.com. The motto of Nate Silver\u2019s website, www.fiverthirtyeight.com, is \u201cPolitics Done Right.\u201d I\u2019m not sure that his latest round of pollster ratings lives up to that moniker. As most poll followers know, Nate shot to fame during the 2008 election, taking the statistical skills he developed to [&hellip;]<\/p>\n","protected":false},"author":939,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-40802236191","post","type-post","status-publish","format-standard","hentry","category-uncategorized"],"_links":{"self":[{"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/posts\/40802236191","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/users\/939"}],"replies":[{"embeddable":true,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/comments?post=40802236191"}],"version-history":[{"count":6,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/posts\/40802236191\/revisions"}],"predecessor-version":[{"id":40802255466,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/posts\/40802236191\/revisions\/40802255466"}],"wp:attachment":[{"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/media?parent=40802236191"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/categories?post=40802236191"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.monmouth.edu\/polling-institute\/wp-json\/wp\/v2\/tags?post=40802236191"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}