One of the more consistent criticisms of the news media over the past decade or two has been that it is overly reliant on coverage of the “horse race� — that is, a focus on who is likely to win an election rather than stories about the actual candidates and their positions.
At times, that criticism is fair. New polling data provides new information about a campaign that triggers a response from the news industry that consistent rhetoric from candidates doesnâ€
But thereâ€
With that admittedly defensive context established, we can now turn our attention to the point of this article: differentiating between assessments of the likely outcome of the race that are useful and those that are garbage. Both exist! None is perfect! And some predictors that are garbage or garbage-approximate might end up close to the mark simply by virtue of the broken-clock truism. If you want to understand what might happen next month, though, it’s useful to know where to look.
Without further ado, here is an assessment of some of those predictive vehicles, arranged from least to most sophisticated.
Social media surveys
These fall into the category of “garbage.�
To be clear, weâ€
Itâ€
Because the next predictive vehicle is …
Betting markets
Betting markets, sites where people can invest in the likelihood of a particular electoral outcome, are relatively new and exist in a murky and evolving legal landscape. The theory, though, is uncomplicated: Let people put money on how they expect an election to unfold, and the wisdom of the market will produce predictive results.
These markets donâ€
And then thereâ€
Past predictors
Another way in which people attempt to predict the outcome of the election is to look at certain indicators that correlate to past results, like presidential approval ratings and shifts in the economy. The most famous purveyor of this approach is American University professor Allan Lichtman, who generates media attention every four years with his assessments of what the indicators he looks at say about the upcoming race.
So how has he done? Well, Lichtman predicted that Joe Biden would win in 2020, which he did, and that Trump would win four years before that … which he did, despite losing the popular vote. In 2000, Lichtman predicted that Al Gore would win, his sole “wrong� prediction since 1984 — except that Gore won the popular vote, too. In national races where the popular-vote margin was 3 percentage points or less, in other words, Lichtman is 2 for 3, depending on whether you want to say he got 2000 or 2016 wrong.
This year, he says Harris will emerge victorious.
Statistically weighted polls
We have, at last, arrived at attempts to actually measure support among American voters.
Before we get too far, though, letâ€
This isnâ€
The business model for betting markets is making money on betting. If they get the results right, great. The business model for pollsters is providing accurate assessments of opinion.
Most pollsters. There are polling firms that work for candidates or that seem to have found a niche in providing partisan media outlets with talking points. 538â€
Election polls also tend to jump around a lot, particularly in a close race. As weâ€
Math can get weird, so itâ€
Polling averages
One way to accommodate those mathematical fluctuations is with an average of polls. For this part, letâ€
Imagine a race between candidates from two parties, the Circles and the Squares. Over the last 100 days of the election, both parties have their conventions and both campaigns are rocked by scandals. The actual support each candidate has — that is, the support each candidate would see if the election were that day — goes up and down in a range from 45 percent to 50 percent, as below.
This data is fake, mind you, generated solely for illustrative purposes. For the same reason, we also generated polling of the race from four different pollsters, each with different margins of error (from 4 to 6 percent), different polling frequencies and different “house effects� — tendencies of different pollsters to advantage one party or the other.
Below we show how those four firms “polled� the race. (To generate these, we shifted the “real� value of support for a given day based on randomized consideration of house effects and margins of error.) Firm A had a low margin of error (MOE) and low house effect. B had a high MOE and low house effect. C had a higher MOE and modest house effect, while D had a low MOE and big house effect. We assumed each poll lasted three days; the release date of the poll (the day after it was completed) is shown.
All over the place! With 50 days until the election, for example, at a point when the “real� support had the Circle Party with a 1-point lead, the most recent polls from the four pollsters showed Square plus-2, Circle plus-6, Circle plus-3 and Circle plus-6. Hard to know what to think!
One issue is that those polls were taken at different times. Another is that the race changed in the days before the 50-day mark, as our “realâ€� data shows. Polls wouldnâ€
If we look at the average of the four polls (using a seven-day average of when polls were actually being conducted), the trends become clearer.
In fact, the average comports well with the “real� values. At the 50-day mark, Circle still has a 4-point lead in the average, but in less than a week it has the two candidates running even.
Notice that the end result, though, isnâ€
Overall, though, the average was a better predictor of “real� sentiment over the course of the last 100 days. It was, on average, about 0.1 points away from the “real� margin between the candidates on any given day. The pollsters ranged between 0.5 points (Pollster C) and 1.7 points (Pollster D) away from the “real� values on the dates their polls were released — in part because the release dates of polls themselves are later than support is actually measured.
Again, this is just an example, done with randomized values. But the point is the same: Averages end up giving a better sense of the course of an election, albeit an imperfect one. And the more polling, the better the average tends to do.
Weighted polling averages
One way in which poll watchers and the media try to ensure more accuracy is by eliminating or de-emphasizing dubious or historically inaccurate polls. The Washington Postâ€
Thereâ€
How effective are the results? Well, we donâ€
Election forecasts
Of course, those national averages are also hobbled by the same asterisk that tripped up Lichtman: The president isnâ€
Right now, 538 suggests that Harris would win 53 times if the election were run 100 times as polls stand at the moment. This doesnâ€
If I said that 53 percent of the judges in a baking competition thought you had the better pie — a measure of support equivalent to a polling average — youâ€
After the 2016 election, 538 (then under Silverâ€
The current forecasts are probably the most useful predictor of what will happen, precisely because they demonstrate so much uncertainty about the outcome. Unlike Lichtman or the anonymous investors in betting markets, forecasts based on polling averages suggest that the race is (and has long been) a toss-up.
Might as well add that to our list, in fact:
Tossing a coin
This is admittedly not the most sophisticated means of determining who will win. But it remains the approach that best captures the state of the race.