Thinking, Fast and Slow by Noble Prize winner Daniel Kahneman provides a new
model
of human behavior and how the mind works which aims to transform the way we
take decisions and experience the world. In writing the book, Kahneman intended
to “improve the ability to identify and understand errors of judgment and
choice, in others and eventually in ourselves, by providing a richer and more
precise language to discuss them.”[1]
Kahneman
starts with the notion that our minds contain two interactive modes of
thinking:
System
1 operates automatically and quickly, with little or no effort and no sense
of
voluntary control.
System
2 allocates attention to the effortful mental activities that demand it,
including
complex computations. The operations of System 2 are often associated
with
the subjective experience of agency, choice, and concentration.[2]
In
other words, System 1 is unconscious, intuitive thought (automatic pilot), while
slower System 2 is conscious, rational thinking (effortful system).
When
we are awake, most of our actions are controlled automatically by System 1. The
mind
cannot consciously perform the thousands of complex tasks per day that human
functioning requires. System 2 is normally in a low-effort mode. System 2 activates
when System 1 cannot deal with a task–when more detailed processing is needed;
only System 2 can construct thoughts in a step-by-step fashion. In addition, it
continuously monitors human behavior. The interactions of Systems 1 and 2 are
usually highly efficient. However, System 1 is prone to biases and errors, and
System 2 is often lazy.
System
2 requires effort and acts of self-control in which the intuitions and impulses
of System 1 are overcome. Attention and Effort requires the lazy system 2 to act.[3]
System
2 is by its nature lazy.[4]
System
1 works in a process called associative
activation: ideas that have been evoked trigger connected coherent ideas.[5]
System
2 is required when cognitive strain
comes up due to unmet demands of the situation which require the system 2 to
focus.[6]
Our
system 1 develops our image of what is normal
and associative ideas are formed which represent the structure of events in our
life and represent the structure of events in our life and interpretation of
the present as well as expectation of the future.[7]
System
1 is a kind of machine which jumps to conclusions,
which can lead to intuitive errors, which may be prevented by a deliberate
intervention of System 2.[8]
System
1 forms basic assesments by continuously monitoring what is going on inside and
outside the mind, and continuously generating assessments of various aspects of
the situation without specific intention and with little or no effort. These
basic assessments are easily substituted for more difficult questions.[9]
Kahneman
next discusses heuristics–“a simple procedure that helps find adequate, though
humans
to act fast, but they can also lead to wrong conclusions (biases) because they
sometimes substitute an easier question for the one asked. A type of heuristic
is the halo effect–“the tendency to like (or dislike) everything about a
person–including things you have not observed. ..”[11] A simple example is rating a baseball
player as good at pitching because he is handsome and athletic.
Kahneman
then discusses the law of small numbers which basically states that researchers
who pick too small a sample leave themselves at the mercy of sampling luck.[12]
The
main lesson to be learned in this chapter is about the difficult relationship
between our mind and statistics. System 1 is highly adept in one form of
thinking-it automatically and effortlessly identifies causal connections
between events, sometimes even when the connection is spurious. System 1 is
inept when faced with ‘merely statistical’ facts, which change the probability
of outcomes but do not cause them to happen.
A
random event, by definition, does not lend itself to explanation, but
collections of random events do behave in a highly regular fashion. Also
predicting the results is based on the following facts:
Of
course results of large samples deserve more trust than smaller samples, we
know this as the law of large numbers.
But
also, more significantly, the following two statements mean exactly the same
thing:
· large samples are more precise than small samples.
· Small samples yield extreme results more often than large
samples do.
Let
us repeat the following result: “researchers who pick too small a sample leave
themselves at the mercy of sampling luck”, traditionally psychologists do not
use calculations to decide on sample
size. They use their judgement, which is commonly flawed.
People
are not adequately sensitive to sample-size. System 1 is not prone to doubt. It
suppresses ambiguity and spontaneously constructs stories that are as coherent
as possible. System 2 is capable of doubt, because it can maintain incompatible
possibilities at the same time.
The
strong bias toward believing that small samples closely resemble the population
from which they are drawn is also part of a larger story: we are prone to
exaggerate the consistency and coherence of what we see.
Our
prelidiction for causal thinking exposes us to serious mistakes in evaluating
the randomness of truly random events. We are pattern seekers, believers in a
coherent world, in which regularities appear not by accident but as a result of
mechanical causality or of someone’s intention. We do not expect to see
regularity produced by a random process, and when we detect what appears to be
a rule, we quickly reject the idea that the process is truly random. Random
processes produce many sequences that the convince people that the process is
not random after all.
The
law of small number is part of two larger stories about the workings of the
mind.
· The exaggerated faith in small samples is only one example
of a more general illusion-we pay more attention to the content of messages
than to information about their reliability.
· Statistics produce many observations that appear to beg for
causal explanations but do not lend themselves to such explanations. Many facts
of the world are due to chance, including accidents of sampling. Causal
explanations of chance events are inevitably wrong.
An
example of a heuristic bias is when judgments are influenced by an
uninformative
number
(an anchor), which results from an associative activation in System 1. People
are
influenced
when they consider a particular value for an unknown number before estimating
that number. The estimate for a number then stays close to the anchor. For
example, two groups estimated Gandhi’s age when he died. The first group were
initially asked whether he was more than 114; a second group was asked whether
he was 35 or older. The first group then estimated a higher number for when he
died than the second one.[13]
Two
different mechanisms produce anchoring effects-one for each system. There is a
form of anchoring that occurs in a deliberate process of adjustment, an
operation of System 2. And there is anchoring that occurs by a priming effect,
an automatic manifestation of System 1.
Insufficient
adjustment neatly explains why you are likely to drive too fast when you come
off the highway into city streets-especially if you are talking with someone as
you drive.
Adjustment
is a deliberate attempt to find reasons to move away from the anchor: people
who are instructed to shake their head when they hear the anchor, as if they
rejected it, move farther from the anchor, and people who nod their head show
enhanced anchoring.
Adjustment
is an effortful operation. People adjust less (stay closer to the anchor) when
their mental resources are depleted, either because their memory is loaded with
digits or because they are slightly drunk. Insufficient adjustment is a failure
of a weak or lazy System 2.
Suggestion
is a priming effect, which selectively evokes compatible evidence. System 1
understands sentences by trying to make them true, and the selective activation
of compatible thoughts produces a family of systematic errors that make us
gullible and prone to believe too strongly whatever we believe.
A
process that resembles suggestion is indeed at work in many situations: System
1 tries its best toconstruct a world in which the anchor is the true number.
Suggestion
and anchoring are both explained by the same automatic operation of System 1.
A
key finding of anchoring research is that anchors that are obviously random can
be just as effective as potentially informative anchors. Anchors clearly do not
have their effects because people believe they are informative.
Anchoring
effects-sometimes due to priming, sometimes to insufficient adjustment-are
everywhere. The psychological mechanisms that produce anchoring make us far
more suggestible than most of us would want to be. And of course there are
quite a few people who are willing and able to exploit our gullibility.
A
strategy of deliberately “thinking the opposite” may be a good defense against
anchoring effects, because it negates the biased recruitment of thoughts that
produces these effects.
System
2 is susceptible to the biasing influence of anchors that make some information
easier to retrieve.
A
message, unless it is immediately rejected as a lie, will have the same effect
on the associative system regardless of its reliability. The gist of the
message is the story, which is based on whatever information is available, even
if the quantity of of the information is slight and its quality is poor.
Anchoring
results from associative activation. Whether the story is true, or believable,
matters little, if at all. The powerful effect of random anchors is an extreme
case of this phenomenon, because a random anchor obviously provides no
information at all.
The
main moral of priming research is that our thoughts and our behaviour are
influenced, much more than we know or want, by the environment of the moment.
The
concept of availability is the
process of judging frequency by “ the ease with which instances come to mind.”,
this heuristic is known to be both a deliberate problem solving strategy and an
automatic operation.[14]
A
question considered early was how many instances must be retrieved to get an
impression of the ease with which they come to mind. We now know the answer:
none.
The
availability heuristic, like other heuristics of judgement, substitutes one
question for another: you wish to estimate the size of a category or the
frequency of an event, but you report an impression of the ease with which
instances come to mind. Substitution of questions inevitably produces
systematic errors.:
· A salient event that attracts your attention will be easily
retrieved from memory.
· A dramatic event temporarily increases the availability of
its category.
· Personal experiences, pictures, and vivid examples are more
available than incidents that happened to others, or mere words, or statistics.
Resisting this large collection of
potential availability biases is possible, but tiresome.
One of the best-known studies of
availability suggests that awareness of your own biases can contribute to peace
in marriages, and probably in other joint projects.
The ease with which instances comes to
mind is a System 1 heuristic, which is replaced by a focus on content when
System 2 is more engaged.
People who let themselves be guided by
System 1 are more strongly susceptible to availability biases than others who
are in a higher state of vigilance. The following are some conditions in which
people “go with the flow” and are affected more strongly by ease of retrieval
than by the content they retrieved:
· When they are engaged in another effortful task.
· When they are in a good mood.
· If they are depressed.
· If they are knowledgeable novices.
· Faith in intuition.
· Are or made to feel powerful.
The
concept of an affect heuristic is one in which people make judgements and
decisions by consulting their emotions, a particularly important concept is:
the availability cascade, the importance of an idea is often judged by the
fluency (and emotional charge ) with which that idea comes to mind, this has
impacts on public policy, particularly with reference to the effect of the
media.[15]
Availability
effects help explain the pattern of insurance purchases and protective action
after disasters. Victims and near victims are very concerned after a disaster. However,
the memories of the disaster dim over time, and so do worry and diligence.
Protective
actions, whether by individuals or governments, are usually designed to be
adequate to the worst disaster actually experienced.
Esimates
of causes of death are warped by media coverage. The coverage is itself biased
towards novelty and poignancy. The media do not just shape what the public is
interested in, but are also shaped by it.
Notion
of an affect heuristic was developed in which people make judgements and
decisions by consulting their emotions: Do I like it? Do I hate it? How
strongly do I feel about it?
“The
emotional tail wags the rational dog.” The affect heuristic simplifies our
lives by creating a world that is much tidier than reality. In the real world,
of course, we often face painful trade-offs between benefits and costs.
Availability
cascades are real and they undoubtedly distort priorities in the allocation of
public resources. One perspective is offered by Cass Sunstein who would seek
mechanisms that insulate decision makers from public pressures, letting the
allocation of resourcesa be determined by impartial experts who have a broad
view of all risks and of the resources available to reduce them. Paul Slovic on
the other hand trusts the experts much less and the public somewhat more than
Sunstein does, and he points out that insulating the experts from the emotions
of the public produces policies that the public will reject-an impossible
situation in a democracy.
People
who are asked to assess probability are not stumped, because they do not try to
judge probability as statisticians and philosophers use the word. A question
about probability or likelihood activates a mental shotgun, evoking answers to
easier questions. Judging probability by representativeness has important
virtues: the intuitive impressions that it produces are often-indeed,
usually-more accurate than chance guesses would be.[16]
In
the absence of specific information about a subject, you will go by the base
rates.
Activation
of association with a stereotype, is an automatic activity of System1.
Representativeness
involves ignoring both the base rates and the doubts about the veracity of the
description. This is a serious mistake, because judgements of similarity and
probability are not constrained by the same logical rules. It is entirely
acceptable for judgements of similarity to be unaffected by the base rates and
also by the possibility that the description was inaccurate, but anyone who
ignores base rates and the quality of evidence in probability assessments will
certainly make mistakes.
Logicians
and statisticians have developed competing definitions of probability, all very
precise.
In
contrast people who are asked to assess probability are not stumped, because
they do not try to judge probability as statisticians and philosophers use the
word. A question about probability or likelihood activates a mental shotgun,
evoking answers to easier questions.
Although
it is common, prediction by representativeness is not statistically optimal.
Judging
probability by representativeness has important virtues: the intuitive
impressions that it produces are often-indeed, usually-more accurate than
chance guesses would be.
In
other situations, the stereotypes are false and the representativeness
heuristic will mislead, especially if it causes people to neglect base-rate
information that points in another direction.
One
sin of representativeness is an excessive willingness to predict the occurrence
of unlikely (low base-rate) events.
People
without training in statistics are quite capable of using base rates in
predictions under some conditions.
Instructing
people to “think like a statistician” enhanced the use of base rate
information, while the instruction to “think like a clinician” had the opposite
effect.
Some
people ignore base rates because they believe them to be irrelevant in the
presence of individual information. Others make the same mistake because they
are not focussed on the task.
The
second sin of representativeness is insensitivity to the quality of evidence.
To
be useful your beliefs should be constrained by the logic of probability.
The
relevant “rules” for such cases are provided by Bayesian Statistics: the logic
of how people should change their mind in the light of evidence.
There
are two ideas to keep in mind about Bayesian reasoning and how we tend to mess
it up. The first is that base rates matter, even in the presence of evidence
about the case at hand. This is often not intuitively obvious. The second is
that intuitive impressions of the diagnosity of evidence are often exaggerated.
A
conjunction fallacy is one which people commit when they judge a conjunction of
two events to be more probable than one of the events in a direct comparision.[17]
When
you specify a possible event in greater detail you can only lower its
probability. So, there is a conflict between the intuition of
representativeness and the logic of probability.
The
word fallacy is used, in general, when people fail to apply a logical rule that
is obviously relevant. Amos and I introduced the idea of a conjunction fallacy,
which people commit when they judge a conjunction of two events to be more
probable than one of the events in a direct comparision.
The
fallacy remains attractive even when you recognise it for what it is.
The
uncritical substitution of plausibility for probability has pernicious effects
on judgements when scenarios are used as tools of forecasting.
Adding
detail to scenarios makes them more persuasive, but less likely to come true.
Less
is more:sometimes even in joint evaluation: the scenario that is judged more
probable is unquestionably more plausible, a more coherent fit with all that is
known.
A
reference to a number of individuals brings a spatial representation to mind.
The
frequency representation, as it is known, makes it easy to appreciate that one
group is wholly included in the other. The solution to the puzzle appears to be
that a question phrased as “how many?’
makes you think of individuals, but the same question phrased as “ what
percentage?” does not.
The
laziness of System 2 is an important fact of life, and the observation that
representativeness can block the application of an obvious logical rule is also
of some interest.
Intuition
governs judgments in the between-subjects condition:logic rules in joint
evaluation. In the Linda problem, in contrast, intuition often overcame logic
even in joint evaluation, although we identified some conditions in which logic
prevails.
The
blatant violations of the logic of probability that we had observed in
transparent problems were interesting.
Causes
trump statistics, in the sense that statistical base rates are generally
underweighted and causal base rates are considered as information about the
individual.[18]
This
chapter considers a standard problem of Bayesian inference. There are two items
of information: a base rate and the imperfectly reliable testimony of a
witness.
You
can probably guess what people do when faced wth this problem: they ignore the
base rate and go with the witness.
Now
consider a variation of the same story, in which only the presentation of the
base rate has been altered.
The
two versions of the problem are mathematically indistinguishable, but they are
psychologically quite different. People who read the first version do not know
how to use the base rate and often ignore it. In contrast, people who see the
second version give considerable weight to the base rate, and their average
judgment is not too far from the Bayesian solution. Why?
In
the first version, the base rate is a statistical fact. A mind that is hungry
for causal stories finds nothing to chew on.
In
the second version, in contrast, you formed a stereotype, which you apply to
unknown individual observations. The stereotype is easily fitted into a causal
story. In this version, there are two causal stories that need to be combined
or reconciled. The inferences from the two stories are contradictory and
approximately cancel each other. The Bayesian estimate is 41%, reflecting the
fact that the base rate is a little more extreme than the reliability of the
witness.
The
example illustrates two types of base rates. Statistical base rates are facts
about a population to which a case belongs, but they are not relevant to the
individual case. Causal base rates change your view of how the individual case
came to be. The two types of base-rate information are treated differently:
Statistical
base rates are generally underweighted, and sometimes neglected altogether, when
specific information about the case at hand is available.
Causal
base rates are treated as information about the individual case and are easily
combined with other case-specific information.
The
causal version of the cab problem had the form of a stereotype: Stereotypes are
statements about the group that are (at least tentatively) accepted as facts
about every member.
These
statements are readily interpreted as setting up a propensity in individual
members of the group, and they fit in a causal story.
Stereotyping
is a bad word in our culture, but in the authors usage it is neutral. One of
the bsic characteristics of System 1 is that it represents categories as norms
and prototypical examplars; we hold in memory a representation of one or more “normal”
members of each of these categories. When the categories are social, these
representations are called stereotypes. Some stereotypes are perniciously
wrong, and hostile stereotyping can have dreadful consequences, but the
psychological facts cannot be avoided: stereotypes, both correct and false, are
how we think of categories.
You
may note the irony. In the context of the
problem, the neglect of base-rate information is a cognitive flaw, a
failure of Bayesian reasoning, and the reliance on causal base rates is
desireable. Stereotyping improves the accuracy of judgement. In other contexts,
however, there is a strong social norm against stereotyping, which is also
embedded in the law.
The
social norm against stereotyping, including the opposition to profiling, has
been highly been highly beneficial in creating a more civilised and more equal
society. It is useful to remember, however, that neglecting valid stereotypes
inevitably results in suboptimal judgements.
The
explicitly stated base rates had some effects on judgment, but they had much
less impact than the statistically equivalent causal base rates. System 1 can
deal with stories in which the elements are causally linked, but it is weak in
statistical reasoning. For a Bayesian thinker, of course, the versions are
equivalent. It is tempting to conclude that we have reached a satisfactory
conclusion:causal base rates are used; merely statistical facts are more or
less neglected. The next study, however, shows that the situation is rather
more complex.
Individuals
feel relieved of responsibility when they know that others can take
responsibility.
Even
normal, decent people do not rush to help when they expect others to take on
the unpleasantness of dealing with a seizure.
Respondents
“quietly exempt themselves” (and their friends and acquaintances) from the
conclusions of experiments that surprise them.
To
teach students any psychology they did not know before, you must surprise them.
But which surprise will do? When respondents were presented with a surprising
statistical fact they managed to learn nothing at all. But when the students
were surprised by individual cases-two nice people who had not helped-they
immediately made the generalisation and inferred that helping is more difficult
than they had thought.
This
is a profoundly important conclusion. People who are taught surprising
statistical facts about human behaviour may be impressed to the point of
telling their friends about what they have heard, but this does not mean that
their understanding of the world has really changed. The test of learning
psychology is whether your understanding of situations you encounter has
changed, not whether you have learned a new fact. There is a deep gap between
our thinking about statistics and our thinking about individual cases.
Statistical results with a causal interpretation have a stronger effect on our
thinking than noncausal information. But even compelling causal statistics will
not change long-held beliefs or beliefs rooted in personal experience. On the
other hand, surprising individual cases have a powerful impact and are a more
effective tool for teaching psychology because the incongruity must be resolved
and embedded in a causal story.
Regression
to the mean involves moving closer to the average than the earlier value of the
variable observed. Also regression to the mean has an explanation, but does not
have a cause.[19]
An
important principle of skill training:rewards for improved performance work
better than punishment of mistakes. This proposition is supported by much
evidence from research.
Regression
to the mean, involves that poor performance is typically followed by
improvement and good performance by deterioration, without any help from either
praise or punishment.
The
feedback to which life exposes us is perverse. Because we tend to be nice to
other people when they please us and nasty when they do not, we are
statistically punished for being nice and rewarded for being nasty.
Regression
does not have a causal explanation. Regression effects are ubiquitous, and so
are misguided casual stories to explain them. The point to remember is that the
change from the first to the second occurrence does not need a causal
explanation. It is a mathematically inevitable consequence of the fact that
luck played a role in the outcome of the first occurence.
Regression
inevitably occurs when the correlation between two measures is less than
perfect.
The
correlation coefficient between two measures, which varies between 0 and 1, is
a measure of the relative weight of the factors they share.
Correlation
and regression are not two concepts-they are different perspectives on the same
concept. The general rule is straightforward but has surprising consequences:
whenever the correlation between two scores is imperfect, there will be
regression to the mean.
Our
mind is strongly biased toward causal explanations and does not deal well with
“mere statistics.” When our attention is called to an event, associative memory
will look for its cause-more precisely, activation will automatically spread to
any cause that is already stored in memory. Causal explanations will be evoked
when regression is detected, but they will be wrong because the truth is that regression
to the mean has an explanation but does not have a cause.
System
2 finds it difficult to understand and learn. This is due in part to the
insistent demand for causal interpritations, which is a feature of System 1.
Regression
effects are a common source of trouble in research, and experienced scientists
develop a healthy fear of the trap of unwarranted causal inference.
Intuitive
predictions need to be corrected because they are not based on regression to
the mean and are therefore biased. Correcting intuitive predictions are a task
for system 2.[20]
Life
presents us with many occasions to forecast. Some predictive judgments, rely
largely on precise calculations. Others involve intuition and System 1 in two
main varieties. Some intuitions draw primarily on skill and expertise acquired
by repeated experience.
Other
intuitions, which are sometimes subjectively indistinguishable from the first,
arise from the operation of heuristics that often substitute an easy question
for the harder one that was asked. Of course, many judgements, especially in
the professional domain, are influenced by a combination of analysis and
intuition.
We
are capable of rejecting information as irrelevant or false, but adjusting for
smaller weaknesses in the evidence is not something that system 1 can do. As a
result intuitive predictions are almost completely insensitive to the actual
predictive quality of the evidence. When a link is found, what you see is all
there is applies:your associative memory quickly and automatically constructs
the best possible story from the information available.
Next
the evidence is evaluated in relation to a relevant norm.
The
next step involves substitution and intensity matching.
The
final step is a translation from an impression of the relative position of the candidates performation to the result.
Intensity
matching yields predictions that are as extreme as the evidence on which they
are based. By now you should realise that all these operations are features of
system 1. You should imagine a process of spreading activation that is
initially prompted by the evidence and the question, feeds back upon itself,
and eventually settles on the most coherent solution possible.
The
prediction of the future is not distinguished from an evaluation of current
evidence-prediction matches evaluation.
This
is perhaps the best evidence we have for the role of substitution. People are
asked for a prediction but they substitute an evaluation of the evidence,
without noticing that the question they answer is not the one they were asked.
This process is guaranteed to generate predictions that are systematically
biased; they completely ignore regression to the mean.
Intuitive
predictions need to be corrected because they are not regressive and are
therefore are biased.
The
corrected intuitive predictions eliminate these biases, so that predictions
(both high and low) are about equally likely to overestimate and to
underestimate the true value. You will still make errors when your predictions
are unbiased, but the errors are smaller and do not favour either high or low
outcomes.
Correcting
your intuitive predictions is a task for System 2. Significant effort is
required to find the relevant reference category, estimate the baseline
prediction, and evaluate the quality of the evidence. The effort is justified
only when the stakes are high and when you are particularly keen not to make
mistakes.
The
objections to the principle of moderating intuitive predictions must be taken
seriously, because absence of bias is not always what matters most. A
preference for unbiased predictions is justified if all errors of prediction
are treated alike, regardless of their direction. But there are situations in
which one type of error is much worse than another.
For
a rational person, predictions that are unbiased and moderate should not
present a problem.
Extreme
predictions and a willingness to predict rare events from weak evidence are
both manifestations of System1. It is natural for the associative machinery to
match the extremeness of predictions to the perceived extremeness of evidence
on which it is based-this is how substitution works. And it is natural for
System1 to generate overconfident judgements, because confidence, as we have
seen, is determined by the coherence of the best story you can tell from the
evidence at hand. Be warned: your intuitions will deliver predictions that are
too extreme and you will be inclined to put far too much faith in them.
Regression
is also a problem for System 2. The very idea of regression is alien and
difficult to communicate and comprehend.
Matching
predictions to the evidence is not only something we do intuitively; it also
seems a reasonable thing to do. We will not learn to understand regression from
experience. Even when a regression is identified, it will be given a causal
interpretation that is lamost always wrong.
[2]
Ibid-page 21
[3]
Ibid-chapter 2
[4]
Ibid-chapter 3
[5]
Ibid-chapter 4
[6]
Ibid-chapter 5
[7]
Ibid-chapter 6
[8]
Ibid-chapter 7
[9]
Ibid-chapter 8
[10]
Ibid-page 98
[11]
Ibid-page 82
[12]
Ibid-chapter 10
[13]
Ibid-chapter 11.
[14] Ibid-chapter
12.
[15]
Ibid-chapter 13.
[16] Ibid-chapter
14.
[17]
Ibid-chapter 15
[18]
Ibid-chapter 16
[19]
Ibid-chapter 17
[20]
Ibid-chapter 18.