Noise: A Flaw in Human Judgment
Rate it:
Open Preview
Read between October 22 - December 9, 2021
1%
Flag icon
Bias and noise—systematic deviation and random scatter—are different components of error. The targets illustrate the difference.
1%
Flag icon
A general property of noise is that you can recognize and measure it while knowing nothing about the target or bias.
2%
Flag icon
as we will see, noise is the more important problem. But in public conversations about human error and in organizations all over the world, noise is rarely recognized. Bias is the star of the show. Noise is a bit player, usually offstage. The topic of bias has been discussed in thousands of scientific articles and dozens of popular books, few of which even mention the issue of noise. This book is our attempt to redress the balance.
6%
Flag icon
Acceptable lotteries are used to allocate “goods,” like courses in some universities, or “bads,” like the draft in the military. They serve a purpose. But the judgment lotteries we talk about allocate nothing.
6%
Flag icon
By our measure, the median difference in underwriting was 55%, about five times as large as was expected by most people, including the company’s executives.
6%
Flag icon
A defining feature of system noise is that it is unwanted, and we should stress right here that variability in judgments is not always unwanted.
6%
Flag icon
But diversity of tastes can help account for errors if a personal taste is mistaken for a professional judgment.
7%
Flag icon
The noise audits suggested that respected professionals—and the organizations that employ them—maintained an illusion of agreement while in fact disagreeing in their daily professional judgments.
7%
Flag icon
We hold a single interpretation of the world around us at any one time, and we normally invest little effort in generating plausible alternatives to it. One interpretation is enough, and we experience it as true. We do not go through life imagining alternative ways of seeing what we see.
7%
Flag icon
Bad judgment is much easier to identify than good judgment. The
7%
Flag icon
Our conclusion is simple: wherever there is judgment, there is noise, and more of it than you think.
8%
Flag icon
Arguably, there is a continuum, not a category difference, between singular and recurrent decisions.
8%
Flag icon
The nature of singular decisions raises an important question for the study of noise. We have defined noise as undesirable variability in judgments of the same problem. Since singular problems are never exactly repeated, this definition does not apply to them. After all, history is only run once.
8%
Flag icon
Our observations here suggest the opposite advice. From the perspective of noise reduction, a singular decision is a recurrent decision that happens only once.
9%
Flag icon
Judgment can therefore be described as measurement in which the instrument is a human mind. Implicit in the notion of measurement is the goal of accuracy
9%
Flag icon
Judgment, like measurement, refers both to the mental activity of making a judgment and to its product.
9%
Flag icon
We will rely extensively on the analogy between judgment and measurement because it helps explain the role of noise in error.
11%
Flag icon
We have contrasted two ways of evaluating a judgment: by comparing it to an outcome and by assessing the quality of the process that led to it.
12%
Flag icon
System noise is inconsistency, and inconsistency damages the credibility of the system.
12%
Flag icon
All we need to measure noise is multiple judgments of the same problem.
12%
Flag icon
Its basic message is straightforward: in professional judgments of all kinds, whenever accuracy is the goal, bias and noise play the same role in the calculation of overall error.
13%
Flag icon
wherever achieving accuracy is the goal. The weighting of errors by their square is central to statistics.
13%
Flag icon
bias and noise play identical roles in the error equation. They are independent of each other and equally weighted in the determination of overall error.
14%
Flag icon
In terms of overall error, noise and bias are independent: the benefit of reducing noise is the same, regardless of the amount of bias.
14%
Flag icon
It provides the rationale for the goal of reducing system noise in predictive judgments, a goal that is in principle as important as the reduction of statistical bias.
15%
Flag icon
Keep your values and your facts separate.”
15%
Flag icon
The variance of biases across cases—some positive, some negative—is an important source of error and unfairness.
15%
Flag icon
Now, add noise by going down each column and changing some numbers here and there—sometimes by adding prison time to the mean sentence, sometimes by subtracting from it. Because the changes you make are not all the same, they create variability within the column. This variability is noise.
15%
Flag icon
The essential result of this study is the large amount of noise observed within the judgments of each case.
16%
Flag icon
For these reasons, we suspect that the amount of noise defendants face in actual courtrooms is even larger than what we see here.
16%
Flag icon
Not surprisingly, conservative ideology was also related to severity of sentences.
16%
Flag icon
noted, “Patterned differences between judges in the influence of offense/offender characteristics” are “an additional form of sentence disparity.”
17%
Flag icon
System noise is undesirable variability in the judgments of the same case by multiple individuals.
17%
Flag icon
Variability is expected, not just between players but within players. The
17%
Flag icon
we do not always produce identical judgments when faced with the same facts on two occasions.
17%
Flag icon
We have described the process that picks an underwriter, a judge, or a doctor as a lottery that creates system noise. Occasion noise is the product of a second lottery. This lottery picks the moment when the professional makes a judgment, the professional’s mood, the sequence of cases that are fresh in mind, and countless other features of the occasion.
17%
Flag icon
When people form a carefully considered professional opinion, they associate it with the reasons that justify their point of view.
18%
Flag icon
Vul and Pashler drew inspiration from the well-known phenomenon known as the wisdom-of-crowds effect: averaging the independent judgments of different people generally improves accuracy.
18%
Flag icon
if you can get independent opinions from others, do it—this real wisdom of crowds is highly likely to improve your judgment. If you cannot, make the same judgment yourself a second time to create an “inner crowd.”
19%
Flag icon
The propensity to find meaning in such statements is a trait known as bullshit receptivity. (Bullshit has become something of a technical term since Harry Frankfurt, a philosopher at Princeton University, published an insightful book, On Bullshit, in which he distinguished bullshit from other types of misrepresentation.)
19%
Flag icon
you are not the same person at all times.
19%
Flag icon
This behavior reflects a cognitive bias known as the gambler’s fallacy: we tend to underestimate the likelihood that streaks will occur by chance.
20%
Flag icon
way. If our mind is a measuring instrument, it will never be a perfect one.
20%
Flag icon
And because of the dynamics among group members—our emphasis here—the level of noise can be high.
21%
Flag icon
If a song benefited from early popularity, it could do really well. If it did not get that benefit, the outcome could be very different.
22%
Flag icon
Very similar groups can end up in divergent places because of social pressures.
23%
Flag icon
deliberating juries were far noisier than statistical juries—a clear reflection of social influence noise. Deliberation had the effect of increasing noise.
23%
Flag icon
compares the accuracy of predictions made by professionals, by machines, and by simple rules.
23%
Flag icon
A measure that captures this intuition is the percent concordant (PC), which answers a more specific question: Suppose you take a pair of employees at random. What is the probability that the one who scored higher on an evaluation of potential also performs better on the
23%
Flag icon
standard measure that social scientists use. The standard measure is the correlation coefficient (r), which varies between 0 and 1 when two variables are positively related. In the preceding example, the correlation between height and foot size is about .60.
« Prev 1 3