Follow Us

Follow us on Twitter  Follow us on LinkedIn
 

25 October 2022

NIESR's Kellner: The Value of Opinion Polls and Economic Forecasts


There is a danger that political polls, like economic forecasts, are taken as fact. In this blog, Peter Kellner, former president of YouGov explores their limitations and goes on to explain why these are precisely what makes them so valuable.

Adlai Stevenson got it right. The Democratic presidential candidate who lost twice to Dwight Eisenhower in the 1950s once said: “Polls should be taken but not inhaled”.

Pollsters, like economic forecasters, suffer from the habit of journalists, politicians and others to judge them as simply “right” or “wrong”, rather than the best efforts of (mainly) intelligent people to tell us what they think is going on. They are usually somewhere near the truth but seldom precisely accurate – and often mocked for errors that are actually quite modest. Imagine a guest arriving two minutes late at a restaurant and being scolded as fiercely as if they were two hours late.

That, though, is only the start of our problems. Here are five factors that intelligent pollwatchers should take into account.

First, pundits predict; polls don’t. Polls provide our best estimate of the state of public opinion at the time they were conducted. We can never be completely certain what voters will think in a month or a year’s time. The one partial exception to this is the final estimate the polls produce on the eve of a general election, when voters have little time to change their mind. But even then, polls could be caught out by differential turnout – the supporters of one party being more determined to vote than the supporters of another.

Second, polls outside an election campaign differ on what they are actually measuring. In the aftermath of Kwasi Kwarteng’s mini budget, two respected companies produced sharply different figures for Labour’s lead: 33%, said YouGov; 19% said Opinium. Much of the difference can be explained by the fact that the two companies were answering two questions that sound alike but are in fact different:

“What would be the figures for each party if we managed to speak to every elector in Great Britain and added up the numbers supporting each party?”

Or:

“If nobody switched parties between now and the next general election, what would the result be?”

YouGov’s polls seek to answer the first question, while Opinium seeks to answer the second. They are different because of the way they deal with people who say they currently don’t know how they would vote. YouGov’s data indicates that of the 14 million people who voted Conservative election, five million now don’t know how they would vote. They are omitted from the voting intention figures. Opinium assumes that many of them will in the end vote, and that most of them will return to the Tory fold. It also draws on past election data which shows consistently that slightly more Labour than Conservative supporters end up staying at home. If YouGov had adopted Opinium’s approach, it would have cut Labour’s lead by around ten points. Which method you prefer is a matter of judgement about how best to treat those pesky “don’t knows”.

That issue connects to a wider third point. Answering a mid-term poll is different from deciding a general election vote. A worried Conservative could tell a pollster that they would vote Labour or Liberal Democrat in order to give vent to their frustrations, knowing that they will not wake up next morning to a change of government. By-elections are often similar: big anti-government swings in mid term send a message without handing power to the opposition. This is why one of the most common patterns in polls for more than 60 years is for support for the governing party to slump in mid-term but recover as the election approaches. It was even true in the run-up to Labour’s landslide in 1997. Two years beforehand, the Tories were 40 points behind (sometimes more). They lost the popular vote in the election by 13 points – a lot, but far less than 40.

A pundit might well say now that the Conservatives will recover to some extent between now and the next election. I, for one, expect this. But, as those financial ads warn, past record is no guarantee of  future performance.

Fourth, polls, however intelligently designed, are prone to systematic errors. Long gone are the days when most people would happily give their views to pollsters in the street or on the phone. Response rates for both have collapsed, making them increasingly expensive. Most political polls these days are conducted online.  (YouGov pioneered this in Britain 22 years ago; other companies have now switched to online research.)

However, only a minority of electors belong to online panels, which generate the email addresses for pollsters to contact – and by no means all of them agree to take part in political polls. This means that pollsters must take the answers from the people they can reach and deduce the views of those they can’t reach. They are in the business of modelling public opinion, not simply measuring it. They use various methods to select and weight their samples in order to reflect the electorate as a whole – by age, gender, education, social class, region, past vote and so on. But it is always possible that, at any given time, some other factor influences the way people vote, and it turns out that the sampling and weighting system has not allowed for this.

Moreover, when one party is out of favour, its supporters might be under-recorded, either because they are more reluctant to take part in polls at all, or to admit their loyalty. This might be one reason why so many people who voted Conservative in 2019 now say “don’t know”.

Fifth, and lastly, even the best poll, getting its modelling right, is prone to straightforward sampling deviation. By convention, polls refer to the 95% confidence limit (for statisticians, the two-sigma deviation level). That is, the range which should be accurate 19 times out of 20. In a normal political poll, reporting the voting intentions of around 1500 people, its figures for each of the main parties should be within 2.5 percentage points of its true support (assuming there are no other sources of error). However, that figure should be qualified. Two-thirds of the time, the margin for error is half that – 1 ¼ points – but one poll in 20 will be outside that range. It will be a rogue poll not because the pollsters are no good, but because they are unlucky.

In September, the month that saw Liz Truss becoming Prime Minister, her first, and now former, Chancellor delivering his mini Budget, the financial markets going haywire and Tory support tanking, I counted 38 published voting intention polls. Statistically, two of them were probably rogue polls. Which two? Good question, but forgive me if I leave you to check them out and make up your own mind.

NIER



© NIESR


< Next Previous >
Key
 Hover over the blue highlighted text to view the acronym meaning
Hover over these icons for more information



Add new comment