Noise: A Flaw in Human Judgment
Rate it:
Read between July 31 - August 28, 2024
56%
Flag icon
To reduce the role of luck, the researchers examined how participants did, on average, across numerous forecasts.
56%
Flag icon
make their forecasts in terms of probabilities that an event would happen,
56%
Flag icon
However, given our objective ignorance of future events, it is much better to formulate probabilistic forecasts.
56%
Flag icon
well calibrated.
57%
Flag icon
They gave the participants the opportunity to revise their forecasts continuously in light of new information.
57%
Flag icon
allowed the forecasters to update their forecasts.
57%
Flag icon
each one of these updates is treated as a new forecast.
57%
Flag icon
“When the facts change, I change my mind. What do you do?”)
57%
Flag icon
Brier scores, as they are known, measure the distance between what people forecast and what actually happens.
57%
Flag icon
pervasive problem associated with probabilistic forecasts: the incentive for forecasters to hedge their bets by never taking a bold stance.
57%
Flag icon
Margaret’s forecasts are well calibrated but also practically useless.
57%
Flag icon
high resolution in addition to good calibration.
57%
Flag icon
Brier scores are based on the logic of mean squared errors, and lower scores are better: a score of 0 would be perfect.
57%
Flag icon
overwhelming majority of the volunteers did poorly, but about 2% stood out.
57%
Flag icon
calls these well-performing people superforecasters.
57%
Flag icon
This comparison is worth pausing over.
57%
Flag icon
superforecasters
57%
Flag icon
unusually intelligent.
57%
Flag icon
superforecasters are unusually good with numbers.
57%
Flag icon
ease in thinking analytically and probabilistically.
57%
Flag icon
structure and disaggregat...
This highlight has been truncated due to consecutive passage length restrictions.
57%
Flag icon
break it up into its component parts.
57%
Flag icon
they ask and try to answer an assortment of subsidiary questions.
57%
Flag icon
taking the outside view, and they care a lot about base rates.
57%
Flag icon
isn’t their sheer intelligence; it’s how they apply it.
57%
Flag icon
high level of “active open-mindedness.”
57%
Flag icon
not shy about updating their judgments (without overreacting) when new information becomes available.
57%
Flag icon
“the strongest predictor of rising into the ranks of superforecasters is perpetual beta, the degree to which one is committed to belief updating and self-improvement.”
57%
Flag icon
They like a particular cycle of thinking: “try, fail, analyze, adjust, try again.”
57%
Flag icon
tested the effect of different interventions on the quality of subsequent judgments.
57%
Flag icon
three of the strategies we have described to improve judgments:
57%
Flag icon
probabilistic reasoning.
57%
Flag icon
learned about various biases
57%
Flag icon
importance of averaging multiple predictions from...
This highlight has been truncated due to consecutive passage length restrictions.
57%
Flag icon
Teaming could increase accuracy by encouraging forecasters to deal with opposing arguments and to be actively open-minded.
58%
Flag icon
three interventions worked,
58%
Flag icon
three major reasons why some forecasters can perform better or worse than others:
58%
Flag icon
more skilled at finding and analyzing data in the environment that are relevant to the prediction they have to make.
58%
Flag icon
forecasters may have a general tendency to err on a particular side of the true value
58%
Flag icon
noisy in their use of the probability scale.
58%
Flag icon
BIN
58%
Flag icon
model for forecasting.
58%
Flag icon
all three interventions worked primarily by reducing noise.
58%
Flag icon
boosted accuracy, it worked mainly by suppressing random errors in judgment.
58%
Flag icon
Tetlock’s training is designed to fight psychological biases.
58%
Flag icon
training forecasters to fight their psychological biases works—by reducing noise.
58%
Flag icon
When working in groups, the superforecasters seem capable of avoiding the dangers of group polarization and information cascades.
58%
Flag icon
Selection had the largest total effect.
58%
Flag icon
But the main effect of selection is, again, to reduce noise.
58%
Flag icon
may owe their success more to superior discipline in tamping down measurement error,
1 17 26