Thinking, Fast and Slow
Book Author | |
---|---|
Published | October 25, 2011 |
Pages | 499 |
Greek Publisher | Κάτοπτρο |
Intuition or deliberation? Where you can (and can’t) trust your brain
What’s it about?
Thinking, Fast and Slow (2011) – a recapitulation of the decades of research that led to Kahneman’s winning the Nobel Prize – explains his contributions to our current understanding of psychology and behavioral economics. Over the years, the research of Kahneman and his colleagues has helped us better understand how decisions are made, why certain judgment errors are so common, and how we can improve ourselves.
A note to readers: this Blink was redone especially for audio. This is the reason why the text version might differ from the audio version. If you’re trying to decide whether to listen or to read, we highly recommend listening!
About the author
Daniel Kahneman, PhD, won the Nobel Prize in Economics in 2002. He is the Senior Scholar at the Woodrow Wilson School of Public and International Affairs, Professor of Psychology and Public Affairs Emeritus at the Woodrow Wilson School, Eugene Higgins Professor of Psychology Emeritus at Princeton University, and a fellow of the Center for Rationality at the Hebrew University in Jerusalem.
Basic Key Ideas
There is a compelling drama going on in our minds, a filmlike plot between two main characters with twists, dramas and tensions. These two characters are the impulsive, automatic, intuitive System 1, and the thoughtful, deliberate, calculating System 2. As they play off against each other, their interactions determine how we think, make judgments and decisions, and act.
System 1 is the part of our brain that operates intuitively and suddenly, often without our conscious control. You can experience this system at work when you hear a very loud and unexpected sound. What do you do? You probably immediately and automatically shift your attention toward the sound. That’s System 1.
This system is a legacy of our evolutionary past: there are inherent survival advantages in being able to make such rapid actions and judgments.
System 2 is what we think of when we visualize the part of the brain responsible for our individual decision-making, reasoning and beliefs. It deals with conscious activities of the mind such as self-control, choices and more deliberate focus of attention.
For instance, imagine you’re looking for a woman in a crowd. Your mind deliberately focuses on the task: it recalls characteristics of the person and anything that would help locate her. This focus helps eliminate potential distractions, and you barely notice other people in the crowd. If you maintain this focused attention, you might spot her within a matter of minutes, whereas if you’re distracted and lose focus, you’ll have trouble finding her.
As we’ll see in the following blinks, the relationship between these two systems determines how we behave.
To see how the two systems work, try solving this famous bat-and-ball problem:
A bat and ball cost $1.10. The bat costs one dollar more than the ball. How much does the ball cost?
The price that most likely came to your mind, $0.10, is a result of the intuitive and automatic System 1, and it’s wrong! Take a second and do the math now.
Do you see your mistake? The correct answer is $0.05.
What happened was that your impulsive System 1 took control and automatically answered by relying on intuition. But it answered too fast.
Usually, when faced with a situation it can’t comprehend, System 1 calls on System 2 to work out the problem, but in the bat-and-ball problem, System 1 is tricked. It perceives the problem as simpler than it is, and incorrectly assumes it can handle it on its own.
The issue the bat-and-ball problem exposes is our innate mental laziness. When we use our brain, we tend to use the minimum amount of energy possible for each task. This is known as the law of least effort. Because checking the answer with System 2 would use more energy, our mind won’t do it when it thinks it can just get by with System 1.
This laziness is unfortunate, because using System 2 is an important aspect of our intelligence. Research shows that practicing System-2 tasks, like focus and self-control, lead to higher intelligence scores. The bat-and-ball problem illustrates this, as our minds could have checked the answer by using System 2 and thereby avoided making this common error.
By being lazy and avoiding using System 2, our mind is limiting the strength of our intelligence.
What do you think when you see the word fragment “SO_P”? Probably nothing. What if you first consider the word “EAT”? Now, when you look again at the word “SO_P,” you would probably complete it as “SOUP.” This process is known as priming.
We’re primed when exposure to a word, concept or event causes us to summon related words and concepts. If you had seen the word “SHOWER” instead of “EAT” above, you probably would’ve completed the letters as “SOAP.”
Such priming not only affects the way we think but also the way we act. Just as the mind is affected by hearing certain words and concepts, the body can be affected as well. A great example of this can be found in a study in which participants primed with words associated with being elderly, such as “Florida” and “wrinkle,” responded by walking at a slower pace than usual.
Incredibly, the priming of actions and thoughts is completely unconscious; we do it without realizing.
What priming therefore shows is that despite what many argue, we are not always in conscious control of our actions, judgments and choices. We are instead being constantly primed by certain social and cultural conditions.
For example, research done by Kathleen Vohs proves that the concept of money primes individualistic actions. People primed with the idea of money – for example, through being exposed to images of money – act more independently and are less willing to be involved with, depend on or accept demands from others. One implication of Vohs’s research is that living in a society filled with triggers that prime money could nudge our behavior away from altruism.
Priming, just like other societal elements, can influence an individual’s thoughts and therefore choices, judgment and behavior – and these reflect back into the culture and heavily affect the kind of society we all live in.
Imagine you meet someone named Ben at a party, and you find him easy to talk to. Later, someone asks if you know anybody who might want to contribute to their charity. You think of Ben, even though the only thing you know about him is that he is easy to talk to.
In other words, you liked one aspect of Ben’s character, and so you assumed you would like everything else about him. We often approve or disapprove of a person even when we know little about them.
Our mind’s tendency to oversimplify things without sufficient information often leads to judgment errors. This is called exaggerated emotional coherence, also known as the halo effect: positive feelings about Ben’s approachability cause you to place a halo on Ben, even though you know very little about him.
But this is not the only way our minds take shortcuts when making judgments.
There is also confirmation bias, which is the tendency for people to agree with information that supports their previously held beliefs, as well as to accept whatever information is suggested to them.
This can be shown if we ask the question, “Is James friendly?” Studies have shown that, faced with this question but no other information, we’re very likely to consider James friendly – because the mind automatically confirms the suggested idea.
The halo effect and confirmation bias both occur because our minds are eager to make quick judgments. But this often leads to mistakes, because we don’t always have enough data to make an accurate call. Our minds rely on false suggestions and oversimplifications to fill in the gaps in the data, leading us to potentially wrong conclusions.
Like priming, these cognitive phenomena happen without our conscious awareness and affect our choices, judgments and actions.
Often we find ourselves in situations where we need to make a quick judgment. To help us do this, our minds have developed little shortcuts to help us immediately understand our surroundings. These are called heuristics.
Most of the time, these processes are very helpful, but the trouble is that our minds tend to overuse them. Applying them in situations for which they aren’t suited can lead us to make mistakes. To get a better understanding of what heuristics are and what mistakes they can lead to, we can examine two of their many types: the substitution heuristic and the availability heuristic.
The substitution heuristic is where we answer an easier question than the one that was actually posed.
Take this question, for example: “That woman is a candidate for sheriff. How successful will she be in office?” We automatically substitute the question we’re supposed to answer with an easier one, like, “Does this woman look like someone who will make a good sheriff?”
This heuristic means that instead of researching the candidate’s background and policies, we merely ask ourselves the far easier question of whether this woman matches our mental image of a good sheriff. Unfortunately, if the woman does not fit our image of a sheriff, we could reject her – even if she has years of crime-fighting experience that make her the ideal candidate.
Next, there is the availability heuristic, which is where you overestimate the probability of something you hear often or find easy to remember.
For example, strokes cause many more deaths than accidents do, but one study found that 80 percent of respondents considered an accidental death a more likely fate. This is because we hear of accidental deaths more in the media, and because they make a stronger impression on us; we remember horrific accidental deaths more readily than deaths from strokes, and so we may react inappropriately to these dangers.
How can you make predictions on whether certain things will happen?
One effective way is to keep the base rate in mind. This refers to a statistical base, which other statistics rely on. For example, imagine a large taxi company has 20 percent yellow cabs and 80 percent red cabs. That means the base rate for yellow taxi cabs is 20 percent and the base rate for red cabs is 80 percent. If you order a cab and want to guess its color, remember the base rates and you will make a fairly accurate prediction.
We should therefore always remember the base rate when predicting an event, but unfortunately this doesn’t happen. In fact, base-rate neglect is extremely common.
One of the reasons we find ourselves ignoring the base rate is that we focus on what we expect rather than what is most likely. For example, imagine those cabs again: If you were to see five red cabs pass by, you’d probably start to feel it’s quite likely that the next one will be yellow for a change. But no matter how many cabs of either color go by, the probability that the next cab will be red will still be around 80 percent – and if we remember the base rate we should realize this. But instead we tend to focus on what we expect to see, a yellow cab, and so we will likely be wrong.
Base-rate neglect is a common mistake connected with the wider problem of working with statistics. We also struggle to remember that everything regresses to the mean. This is the acknowledgment that all situations have their average status, and variations from that average will eventually tilt back toward the average.
For example, if a football striker who averages five goals per month scores ten goals in September, her coach will be ecstatic; but if she then goes on to score around five goals per month for the rest of the year, her coach will probably criticize her for not continuing her “hot streak.” The striker wouldn’t deserve this criticism, though, because she is only regressing to the mean!
Our minds don’t remember experiences in a straightforward way. We have two different apparatuses, called memory selves, both of which remember situations differently.
First, there is the experiencing self, which records how we feel in the present moment. It asks the question: “How does it feel now?”
Then there is the remembering self, which records how the entire event unfolded after the fact. It asks, “How was it on the whole?”
The experiencing self gives a more accurate account of what occurred, because our feelings during an experience are always the most accurate. But the remembering self, which is less accurate because it registers memories after the situation is finished, dominates our memory.
There are two reasons why the remembering self dominates the experiencing self. The first of these is called duration neglect, where we ignore the total duration of the event in favor of a particular memory from it. Second is the peak-end rule, where we overemphasize what occurs at the end of an event.
For an example of this dominance of the remembering self, take this experiment, which measured people’s memories of a painful colonoscopy. Before the colonoscopy, the people were put into two groups: the patients in one group were given long, rather drawn-out colonoscopies, while those in the other group were given much shorter procedures, but where the level of pain increased towards the end.
You’d think the most unhappy patients would be those who endured the longer process, as their pain was endured for longer. This was certainly what they felt at the time. During the process, when each patient was asked about the pain, their experiencing self gave an accurate answer: those who had the longer procedures felt worse. However, after the experience, when the remembering self took over, those who went through the shorter process with the more painful ending felt the worst. This survey offers us a clear example of duration neglect, the peak-end rule, and our faulty memories.
Our minds use different amounts of energy depending on the task. When there’s no need to mobilize attention and little energy is needed, we are in a state of cognitive ease. Yet, when our minds must mobilize attention, they use more energy and enter a state of cognitive strain.
These changes in the brain’s energy levels have dramatic effects on how we behave.
In a state of cognitive ease, the intuitive System 1 is in charge of our minds, and the logical and more energy-demanding System 2 is weakened. This means we are more intuitive, creative and happier, yet we’re also more likely to make mistakes.
In a state of cognitive strain, our awareness is more heightened, and so System 2 is put in charge. System 2 is more ready to double-check our judgments than System 1, so although we are far less creative, we will make fewer mistakes.
You can consciously influence the amount of energy the mind uses to get in the right frame of mind for certain tasks. If you want a message to be persuasive, for example, try promoting cognitive ease.
One way to do this is to expose ourselves to repetitive information. If information is repeated to us, or made more memorable, it becomes more persuasive. This is because our minds have evolved to react positively when repeatedly exposed to the same clear messages. When we see something familiar, we enter a state of cognitive ease.
Cognitive strain, on the other hand, helps us succeed at things like statistical problems.
We can get into this state by exposing ourselves to information that is presented to us in a confusing way, for example, via hard-to-read type. Our minds perk up and increase their energy levels in an effort to comprehend the problem, and therefore we are less likely to simply give up.
The way we judge ideas and approach problems is heavily determined by the way they are expressed to us. Slight changes to the details or focus of a statement or question can dramatically alter the way we address it.
A great example of this can be found in how we assess risk.
You may think that once we can determine the probability of a risk occurring, everyone will approach it in the same way. Yet, this isn’t the case. Even for carefully calculated probabilities, just changing the way the figure is expressed can change how we approach it.
For example, people will consider a rare event as more likely to occur if it’s expressed in terms of relative frequency rather than as a statistical probability.
In what’s known as the Mr. Jones experiment, two groups of psychiatric professionals were asked if it was safe to discharge Mr. Jones from the psychiatric hospital. The first group were told that patients like Mr. Jones had a “10 percent probability of committing an act of violence,” and the second group were told that “of every 100 patients similar to Mr. Jones, 10 are estimated to commit an act of violence.” Of the two groups, almost twice as many respondents in the second group denied his discharge.
Another way our attention is distracted from what is statistically relevant is called denominator neglect. This occurs when we ignore plain statistics in favor of vivid mental images that influence our decisions.
Take these two statements: “This drug protects children from disease X but has a 0.001 percent chance of permanent disfigurement” versus “One of 100,000 children who take this drug will be permanently disfigured.” Even though both statements are equal, the latter statement brings to mind a disfigured child and is much more influential, which is why it would make us less likely to administer the drug.
How do we as individuals make choices?
For a long time, a powerful and influential group of economists suggested that we made decisions based purely on rational argument. They argued that we all make choices according to utility theory, which states that when individuals make decisions, they look only at the rational facts and choose the option with the best overall outcome for them, meaning the most utility.
For example, utility theory would posit this kind of statement: if you like oranges more than you like kiwis, then you’re also going to take a 10 percent chance of winning an orange over a 10 percent chance of winning a kiwi.
Seems obvious, right?
The most influential group of economists in this field centered on the Chicago School of Economics and their most famous scholar Milton Friedman. Using utility theory, the Chicago School argued that individuals in the marketplace are ultra-rational decision-makers, whom economist Richard Thaler and lawyer Cass Sunstein later named Econs. As Econs, each individual acts in the same way, valuing goods and services based on their rational needs. What’s more, Econs also value their wealth rationally, weighing only how much utility it provides them.
So imagine two people, John and Jenny, who both have fortunes of $5 million. According to utility theory, they have the same wealth, meaning they should both be equally happy with their finances.
But what if we complicate things a little? Let’s say that their $5 million fortunes are the end-result of a day at the casino, and the two had vastly different starting points: John walked in with a mere $1 million and quintupled his money, whereas Jenny came in with $9 million that dwindled down to $5 million. Do you still think John and Jenny are equally happy with their $5 million?
Unlikely. Clearly then, there is something more to the way we value things than pure utility.
As we’ll see in the next blink, since we don’t all see utility as rationally as utility theory thinks, we can make strange and seemingly irrational decisions.
If utility theory doesn’t work, then what does?
One alternative is prospect theory, developed by the author.
Kahneman’s prospect theory challenges utility theory by showing that when we make choices, we don’t always act in the most rational way.
Imagine these two scenarios for example: In the first scenario, you’re given $1,000 and then must choose between receiving a definite $500 or taking a 50 percent chance to win another $1,000. In the second scenario, you’re given $2,000 and must then choose between a sure loss of $500 or taking a 50 percent chance on losing $1,000.
If we made purely rational choices, then we would make the same choice in both cases. But this isn’t the case. In the first instance, most people choose to take the sure bet, while in the second case, most people take a gamble.
Prospect theory helps to explain why this is the case. It highlights at least two reasons why we don’t always act rationally. Both of them feature our loss aversion — the fact that we fear losses more than we value gains.
The first reason is that we value things based on reference points. Starting with $1,000 or $2,000 in the two scenarios changes whether we’re willing to gamble, because the starting point affects how we value our position. The reference point in the first scenario is $1,000 and $2,000 in the second, which means ending up at $1,500 feels like a win in the first, but a distasteful loss in the second. Even though our reasoning here is clearly irrational, we understand value as much by our starting point as by the actual objective value at the time.
Second, we’re influenced by the diminishing sensitivity principle: the value we perceive may be different from its actual worth. For instance, going from $1,000 to $900 doesn’t feel as bad as going from $200 to $100, despite the monetary value of both losses being equal. Similarly in our example, the perceived value lost when going from $1,500 to $1,000 is greater than when going from $2,000 to $1,500.
In order to understand situations, our minds naturally use cognitive coherence; we construct complete mental pictures to explain ideas and concepts. For example, we have many images in our brain for the weather. We have an image for, say, summer weather, which might be a picture of a bright, hot sun bathing us in heat.
As well as helping us to understand things, we also rely on these images when making a decision.
When we make decisions, we refer to these pictures and build our assumptions and conclusions based on them. For example, if we want to know what clothes to wear in summer, we base our decisions on our image of that season’s weather.
The problem is that we place too much confidence in these images. Even when available statistics and data disagree with our mental pictures, we still let the images guide us. In summer, the weather forecaster might predict relatively cool weather, yet you might still go out in shorts and a T-shirt, as that’s what your mental image of summer tells you to wear. You may then end up shivering outside!
We are, in short, massively overconfident of our often faulty mental images. But there are ways to overcome this overconfidence and start making better predictions.
One way to avoid mistakes is to utilize reference class forecasting. Instead of making judgments based on your rather general mental images, use specific historical examples to make a more accurate forecast. For example, think of the previous occasion you went out when it was a cold summer day. What did you wear then?
In addition, you can devise a long-term risk policy that plans specific measures in the case of both success and failure in forecasting. Through preparation and protection, you can rely on evidence instead of general mental pictures and make more accurate forecasts. In the case of our weather example, this could mean bringing along a sweater just to be safe.
The key message of this book is:
Thinking, Fast and Slow shows us that our minds contain two systems. The first acts instinctively and requires little effort; the second is more deliberate and requires much more of our attention. Our thoughts and actions vary depending on which of the two systems is in control of our brain at the time.
Actionable advice
Repeat the message!
Messages are more persuasive when we’re repeatedly exposed to them. This is probably because we evolved in a way that made repeated exposure to things that had no bad consequences seem inherently good.
Don’t be influenced by rare statistical events that are over-reported in newspapers.
Disasters and other events are an important part of our history, but we often overestimate their statistical probability because of the vivid images we associate with them from the media.
You’re more creative and intuitive when you’re in a better mood.
When you’re in a better mood, the part of the mind that is alert and analytical tends to relax. That cedes control of your mind to the more intuitive and quicker thinking system, which also makes you more creative.
SECOND REVIEW FROM SHORTFORM
About Book
We’re so self-confident in our rationality that we think all our decisions are well-considered. When we choose a job, decide how to spend our time, or buy something, we think we’ve considered all the relevant factors and are making the optimal choice. In reality, our minds are riddled with biases leading to poor decision making. We ignore data that we don’t see, and we weigh evidence inappropriately.
Thinking, Fast and Slow is a masterful book on psychology and behavioral economics by Nobel laureate Daniel Kahneman. Learn your two systems of thinking, how you make decisions, and your greatest vulnerabilities to bad decisions.
Thinking, Fast and Slow concerns a few major questions: how do we make decisions? And in what ways do we make decisions poorly?
The book covers three areas of Daniel Kahneman’s research: cognitive biases, prospect theory, and happiness.
System 1 and 2
Kahneman defines two systems of the mind.
System 1: operates automatically and quickly, with little or no effort, and no sense of voluntary control
- Examples: Detect that one object is farther than another; detect sadness in a voice; read words on billboards; understand simple sentences; drive a car on an empty road.
System 2: allocates attention to the effortful mental activities that demand it, including complex computations. Often associated with the subjective experience of agency, choice and concentration
- Examples: Focus attention on a particular person in a crowd; exercise faster than is normal for you; monitor your behavior in a social situation; park in a narrow space; multiply 17 x 24.
System 1 automatically generates suggestions, feelings, and intuitions for System 2. If endorsed by System 2, intuitions turn into beliefs, and impulses turn into voluntary actions.
System 1 can be completely involuntary. You can’t stop your brain from completing 2 + 2 = ?, or from considering a cheesecake as delicious. You can’t unsee optical illusions, even if you rationally know what’s going on.
A lazy System 2 accepts what the faulty System 1 gives it, without questioning. This leads to cognitive biases. Even worse, cognitive strain taxes System 2, making it more willing to accept System 1. Therefore, we’re more vulnerable to cognitive biases when we’re stressed.
Because System 1 operates automatically and can’t be turned off, biases are difficult to prevent. Yet it’s also not wise (or energetically possible) to constantly question System 1, and System 2 is too slow to substitute in routine decisions. We should aim for a compromise: recognize situations when we’re vulnerable to mistakes, and avoid large mistakes when the stakes are high.
Cognitive Biases and Heuristics
Despite all the complexities of life, notice that you’re rarely stumped. You rarely face situations as mentally taxing as having to solve 9382 x 7491 in your head.
Isn’t it profound how we can make decisions without realizing it? You like or dislike people before you know much about them; you feel a company will succeed or fail without really analyzing it.
When faced with a difficult question, System 1 substitutes an easier question, or the heuristic question. The answer is often adequate, though imperfect.
Consider the following examples of heuristics:
- Target question: Is this company’s stock worth buying? Will the price increase or decrease?
- Heuristic question: How much do I like this company?
- Target question: How happy are you with your life?
- Heuristic question: What’s my current mood?
- Target question: How far will this political candidate get in her party?
- Heuristic question: Does this person look like a political winner?
These are related, but imperfect questions. When System 1 produces an imperfect answer, System 2 has the opportunity to reject this answer, but a lazy System 2 often endorses the heuristic without much scrutiny.
Important Biases and Heuristics
Confirmation bias: We tend to find and interpret information in a way that confirms our prior beliefs. We selectively pay attention to data that fit our prior beliefs and discard data that don’t.
“What you see is all there is”: We don’t consider the global set of alternatives or data. We don’t realize the data that are missing. Related:
- Planning fallacy: we habitually underestimate the amount of time a project will take. This is because we ignore the many ways things could go wrong and visualize an ideal world where nothing goes wrong.
- Sunk cost fallacy: we separate life into separate accounts, instead of considering the global account. For example, if you narrowly focus on a single failed project, you feel reluctant to cut your losses, but a broader view would show that you should cut your losses and put your resources elsewhere.
Ignoring reversion to the mean: If randomness is a major factor in outcomes, high performers today will suffer and low performers will improve, for no meaningful reason. Yet pundits will create superficial causal relationships to explain these random fluctuations in success and failure, observing that high performers buckled under the spotlight, or that low performers lit a fire of motivation.
Anchoring: When shown an initial piece of information, you bias toward that information, even if it’s irrelevant to the decision at hand. For instance, in one study, when a nonprofit requested $400, the average donation was $143; when it requested $5, the average donation was $20. The first piece of information (in this case, the suggested donation) influences our decision (in this case, how much to donate), even though the suggested amount shouldn’t be relevant to deciding how much to give.
Representativeness: You tend to use your stereotypes to make decisions, even when they contradict common sense statistics. For example, if you’re told about someone who is meek and keeps to himself, you’d guess the person is more likely to be a librarian than a construction worker, even though there are far more of the latter than the former in the country.
Availability bias: Vivid images and stronger emotions make items easier to recall and are overweighted. Meanwhile, important issues that do not evoke strong emotions and are not easily recalled are diminished in importance.
Narrative fallacy: We seek to explain events with coherent stories, even though the event may have occurred due to randomness. Because the stories sound plausible to us, it gives us unjustified confidence about predicting the future.
Prospect Theory
Traditional Expected Utility Theory
Traditional “expected utility theory” asserts that people are rational agents that calculate the utility of each situation and make the optimum choice each time.
If you preferred apples to bananas, would you rather have a 10% chance of winning an apple, or 10% chance of winning a banana? Clearly you’d prefer the former.
The expected utility theory explained cases like these, but failed to explain the phenomenon of risk aversion, where in some situations a lower-expected-value choice was preferred.
Consider: Would you rather have an 80% chance of gaining $100 and a 20% chance to win $10, or a certain gain of $80?
The expected value of the former is greater (at $82) but most people choose the latter. This makes no sense in classic utility theory—you should be willing to take a positive expected value gamble every time.
Furthermore, it ignores how differently we feel in the case of gains and losses. Say Anthony has $1 million and Beth has $4 million. Anthony gains $1 million and Beth loses $2 million, so they each now have $2 million. Are Anthony and Beth equally happy?
Obviously not – Beth lost, while Anthony gained. Puzzling with this concept led Kahneman to develop prospect theory.
Prospect Theory
The key insight from the above example is that evaluations of utility are not purely dependent on the current state. Utility depends on changes from one’s reference point. Utility is attached to changes of wealth, not states of wealth. And losses hurt more than gains.
Prospect theory can be summarized in 3 points:
1. When you evaluate a situation, you compare it to a neutral reference point.
- Usually this refers to the status quo you currently experience. But it can also refer to an outcome you expect or feel entitled to, like an annual raise. When you don’t get something you expect, you feel crushed, even though your status quo hasn’t changed.
2. Diminishing marginal utility applies to changes in wealth (and to sensory inputs).
- Going from $100 to $200 feels much better than going from $900 to $1,000. The more you have, the less significant the change feels.
3. Losses of a certain amount trigger stronger emotions than a gain of the same amount.
- Evolutionarily, the organisms that treated threats more urgently than opportunities tended to survive and reproduce better. We have evolved to react extremely quickly to bad news.
There are a few practical implications of prospect theory.
Possibility Effect
Consider which is more meaningful to you:
- Going from a 0% chance of winning $1 million to 5% chance
- Going from a 5% chance of winning $1 million to 10% chance
Most likely you felt better about the first than the second. The mere possibility of winning something (that may still be highly unlikely) is overweighted in its importance. We fantasize about small chances of big gains. We obsess about tiny chances of very bad outcomes.
Certainty Effect
Now consider how you feel about these options on the opposite end of probability:
- In a surgical procedure, going from a 90% success rate to 95% success rate.
- In a surgical procedure, going from a 95% success rate to 100% success rate
Most likely, you felt better about the second than the first. Outcomes that are almost certain are given less weight than their probability justifies. 95% success rate is actually fantastic! But it doesn’t feel this way, because it’s not 100%.
Status Quo Bias
You like what you have and don’t want to lose it, even if your past self would have been indifferent about having it. For example, if your boss announces a raise, then ten minutes later said she made a mistake and takes it back, this is experienced as a dramatic loss. However, if you heard about this happening to someone else, you likely would see the change as negligible.
Framing Effects
The context in which a decision is made makes a big difference in the emotions that are invoked and the ultimate decision. Even though a gain can be logically equivalently defined as a loss, because losses are so much more painful, different framings may feel very different.
For example, a medical procedure with a 90% chance of survival sounds more appealing than one with a 10% chance of mortality, even though they’re identical.
Happiness and the Two Selves
The new focus of Kahneman’s recent research is happiness. Happiness is a tricky concept. There is in-the-moment happiness, and there is overall well being. There is happiness we experience, and happiness we remember.
Kahneman presents two selves:
- The experiencing self: the person who feels pleasure and pain, moment to moment. This experienced utility would best be assessed by measuring happiness over time, then summing the total happiness felt over time. (In calculus terms, this is integrating the area under the curve.)
- The remembering self: the person who reflects on past experiences and evaluates it overall.
The remembering self factors heavily in our thinking. After a moment has passed, only the remembering self exists when thinking about our past lives. The remembering self is often the one making future decisions.
But the remembering self evaluates differently from the experiencing self in two critical ways:
- Peak-end rule: The overall rating is determined by the peak intensity of the experience and the end of the experience. It does not care much about the averages throughout the experience.
- Duration neglect: The duration of the experience has little effect on the memory of the event.
We tend to prioritize the remembering self (such as when we choose where we book vacations, or in our willingness to endure pain we will forget later) and don’t give enough to the experiencing self.
For example, would you take a vacation that was very enjoyable, but required that at the end you take a pill that gives you total amnesia of the event? Most would decline, suggesting that memories are a key, perhaps dominant, part of the value of vacations. The remembering self, not the experiencing self, chooses vacations!
Kahneman’s push is to weight the experiencing self more. Spend more time on things that give you moment-to-moment pleasures, and diminish moment-to-moment pain. Try to reduce your commute, which is a common source of experienced misery. Spend more time in active pleasure activities, such as socializing and exercise.
Focusing Illusion
Considering overall life satisfaction is a difficult System 2 question. When considering life satisfaction, it’s difficult to consider all the factors in your life, weigh those factors accurately, then score your factors.
As is typical, System 1 substitutes the answer to an easier question, such as “what is my mood right now?”, focusing on significant events (both achievements and failures), or recurrent concerns (like illness).
The key point: Nothing in life is as important as you think it is when you are thinking about it. Your mood is largely determined by what you attend to. You get pleasure/displeasure from something when you think about it.
For example, even though Northerners despise their weather and Californians enjoy theirs, in research studies, climate makes no difference in life satisfaction. Why is this? When people are asked about life satisfaction, climate is just a small factor in the overall question—they’re much more worried about their career, their love life, and the bills they need to pay.
When you forecast your own future happiness, you overestimate the effect a change will have on you (like getting a promotion), because you overestimate how salient the thought will be in future you’s mind. In reality, future you has gotten used to the new environment and now has other problems to worry about.