Superforecasting: The Art and Science of Prediction
Rate it:
Open Preview
2%
Flag icon
Bill has answered roughly three hundred questions like “Will Russia officially annex additional Ukrainian territory in the next three months?” and “In the next year, will any country withdraw from the eurozone?”
2%
Flag icon
Some of the questions are downright obscure, at least for most of us.
2%
Flag icon
When Bill first sees one of these questions, he may have no clue how to answer
2%
Flag icon
He gathers facts, balances clashing arguments, and settles on an answer.
2%
Flag icon
We know that because each one of Bill’s predictions has been dated, recorded, and assessed for accuracy by independent scientific observers. His track record is excellent.
2%
Flag icon
I call them superforecasters because that is what they are. Reliable evidence proves it. Explaining why they’re so good, and how others can learn to do what they do, is my goal in this book.
3%
Flag icon
The one undeniable talent that talking heads have is their skill at telling a compelling story with conviction, and that is enough.
3%
Flag icon
It was easiest to beat chance on the shortest-range questions that only required looking one year out, and accuracy fell off the further out experts tried to forecast—approaching the dart-throwing-chimpanzee level three to five years out.
4%
Flag icon
Often the connections are harder to spot, but they are all around us, in things like the price we pay at the gas station or the layoffs down the street.
5%
Flag icon
How predictable something is depends on what we are trying to predict, how far into the future, and under what circumstances.
5%
Flag icon
Accuracy is seldom determined after the fact and is almost never done with sufficient regularity and rigor that conclusions can be drawn. The reason? Mostly it’s a demand-side problem: The consumers of forecasting—governments, business, and the public—don’t demand evidence of accuracy. So there is no measurement. Which means no revision.
5%
Flag icon
“I have been struck by how important measurement is to improving the human condition,” Bill Gates wrote. “You can achieve incredible progress if you set a clear goal and find a measure that will drive progress toward that goal….This may seem basic, but it is amazing how often it is not done and how hard it is to get right.”
6%
Flag icon
IARPA is an agency within the intelligence community that reports to the director of National Intelligence and its job is to support daring research that promises to make American intelligence better at what it does.
6%
Flag icon
but they do have a real, measurable skill at judging how high-stakes events are likely to unfold three months, six months, a year, or a year and a half in advance.
6%
Flag icon
Foresight isn’t a mysterious gift bestowed at birth. It is the product of particular ways of thinking, of gathering information, of updating beliefs.
7%
Flag icon
bet and a 40/60 bet. And yet, if it’s possible to improve foresight simply by measuring, and if the rewards of improved foresight are substantial, why isn’t measuring standard practice? A big part of the answer to that question lies in the psychology that convinces us we know things we really don’t—things like whether Tom Friedman is an accurate forecaster or not.
7%
Flag icon
When physicians finally accepted that their experience and perceptions were not reliable means of determining whether a treatment works, they turned to scientific testing—and medicine finally started to make rapid advances. The same revolution needs to happen in forecasting.
7%
Flag icon
As with the experts who had real foresight in my earlier research, what matters most is how the forecaster thinks.
7%
Flag icon
superforecasting demands thinking that is open-minded, careful, curious, and—above all—self-critical. It also demands focus. The kind of thinking that produces superior judgment does not come effortlessly. Only the determined can deliver it reasonably consistently, which is why our analyses have consistently found commitment to self-improvement to be the strongest predictor of performance.
7%
Flag icon
The point is now indisputable: when you have a well-validated statistical algorithm, use it.
8%
Flag icon
“mimicking human meaning,” and thereby better at predicting human behavior, but “there’s a difference between mimicking
8%
Flag icon
That’s a space human judgment will always occupy.
11%
Flag icon
As Daniel Kahneman puts it, “System 1 is designed to jump to conclusions from little evidence.”13
14%
Flag icon
Progress only really began when physicians accepted that the view from the tip of their nose was not enough to determine what works.
15%
Flag icon
The first step in learning what works in forecasting, and what doesn’t, is to judge forecasts, and to do that we can’t make assumptions about what the forecast means. We have to know. There can’t be any ambiguity about whether a forecast is accurate or not and Ballmer’s forecast is ambiguous. Sure, it looks wrong. It
15%
Flag icon
It is far from unusual that a forecast that at first looks as clear as a freshly washed window proves too opaque to be conclusively judged right or wrong.
15%
Flag icon
The letter says nothing about the time frame. It wouldn’t matter if Ritholtz waited until 2014 or 2015 or 2016. No matter how much time passed, someone could always say, “Just wait. It’s coming.”6 It also isn’t clear how
Niladri
Time is everything - what time and by how much
16%
Flag icon
Judging forecasts is much harder than often supposed, a lesson I learned the hard way—from extensive and exasperating experience.
17%
Flag icon
That’s why forecasts without timelines don’t appear absurd when they are made.
17%
Flag icon
This sort of vague verbiage is more the rule than the exception. And it too renders forecasts untestable.
17%
Flag icon
Hence forecasting is all about estimating the likelihood of something happening,
18%
Flag icon
But it was never adopted. People liked clarity and precision in principle but when it came time to make clear and precise forecasts they weren’t so keen on numbers.
18%
Flag icon
A more serious objection—then and now—is that expressing a probability estimate with a number may imply to the reader that it is an objective fact,
18%
Flag icon
It’s to inform readers that numbers, just like words, only express estimates—opinions—and nothing more.
18%
Flag icon
Also, bear in mind that words like “serious possibility” suggest the same thing numbers do, the only real difference being that numbers make it explicit, reducing the risk of confusion.
18%
Flag icon
metacognition.
18%
Flag icon
But a more fundamental obstacle to adopting numbers relates to accountability and what I call the wrong-side-of-maybe fallacy.
18%
Flag icon
But people do judge. And they always judge the same way: they look at which side of “maybe”—50%—the probability was on. If the forecast said there was a 70% chance of rain
19%
Flag icon
So what’s the safe thing to do? Stick with elastic language.
19%
Flag icon
With perverse incentives like these, it’s no wonder people prefer rubbery words over firm numbers.
19%
Flag icon
Every “may” is accompanied by an asterisk and the words “or may not” are buried in the fine print. But the interviewer didn’t notice the fine print in Ferguson’s forecast, so he did not ask him to clarify.16
19%
Flag icon
They must use numbers. And one more thing is essential: we must have lots of forecasts.
19%
Flag icon
If a meteorologist says there is a 70% chance of rain tomorrow, that forecast cannot be judged, but if she predicts the weather tomorrow, and the day after, and the day after that, for months, her forecasts can be tabulated and her track record determined. If her forecasting is perfect, rain happens 70% of the time when she says there is 70% chance of rain, 30% of the time when she says there is 30% chance of rain, and so on. This is called calibration. It can be plotted on a simple chart. Perfect calibration is captured by the diagonal line on this chart:
19%
Flag icon
We could focus on the state level in presidential elections, for example, which would give us fifty results per election,