Intuition: What’s the deal with intuition these days? It seems to be on everyone’s mind, a
brainworm on the loose. People are
claiming that tablets like the iPad “tap into intuition,” that Steve Jobs had
an “intuitive designer’s sense,” George Bush “trusted his gut feelings” about
the presence of WMD in Iraq, Kim Kardashian decided that “intuition led me to
divorce” after 72 days of wedded bliss, and the Huffington Post writes “Science says
to trust your gut.” Intuition seems to
be more valued than ever, although there seems to have been no recent upgrade in
our collective intuitive skills. Is intuition really just the flowering of some inner, secret power?
By contrast, the current movie “Moneyball” is about the
success of the Oakland A’s team of 2002, a team that was put together with
guidance of some clever statistical
analysis by their general manager, a baseball quant-jock if there ever was
one. Even Roger Ebert says the film is
about “the war between intuition and statistics.” Is there really a war going on?
In his book “Blink,” Malcolm Gladwell talks about intuition
as “thin-slicing” experience based on training up your experience base after 10,000
hours of practice in a field. If you have that much time-on-task, practicing and learning how to operate efficiently in a field, then you’re
capable of rapidly assessing a situation based on few, rapidly scanned clues,
and somehow coming up with a quick recognition of what’s going on.
Here's the thing to know: How much time have you spent doing search? My best estimate is that I've done about 5,000 hours of search since I first started using Google in late 1998. That is, I started practicing my search skills around 13 years (or 4748 days) ago.
If I've done a bit over 1 hour of search / day since then (which seems reasonable), that means I've invested ~5,000 hours of practice. While that's a lot of time, it's only half of the 10,000 hours that are usually needed for real expertise.
System 1, System 2: Nobel Prize winner Daniel Kahneman points out that people have two
different and parallel systems of thought when they confront problems. The cleverly named “System 1” is a fast recognizer
of situations in context—it identifies and labels objects, picks up on
relationships, and does so by recognition, rapidly, rather than by
deliberation.
Then there’s “System 2,”
the slower, more deliberate, symbol-pushing and rational part of our
minds.
As an example, System 1 recognizes that the number sequence
2, 4, 8, 16, 32… is just doubling from one to the next. It would be System 2 that lets you realize
that this is also the powers of 2. Of course,
if you’re a computer scientist, the powers of 2 has become, over many repeated
exposures, something that’s a System 1 effect.
For non-CS majors the problem 2 * 256 is a System 2 task. For CS-majors, it’s a System 1 task—you
recognize the pattern and say, “it’s 512…” without thinking much. In this sense, intuition is what you’ve been
trained to expect to perceive. It is the
power of repeated exposure and the accumulation of inarticulate recognition
skills.
“We are prone to think that
the world is more regular and predictable than it really is, because our memory
automatically and continuously maintains a story about what is going on, and
because the rules of memory tend to make that story as coherent as possible and
to suppress alternatives. Fast thinking is not prone to doubt.”
Nor is System 1 particularly good at noticing
contradictions. We swim in a sea of
counterintuitions--things that seem to be intuitively correct, but are not. We have difficulty seeing the sea we swim in simply
because we swim in it all the time.
Some examples:
- How
is it that clouds made of water vapor yet can float mid-air? (Water is awfully heavy.)
- The world is visibly and obviously flat—yet
we now believe that it’s intuitively obvious that the world is round. (Trust me, historically speaking, that wasn’t
obvious at all!)
- When we’re on a
merry-go-round, our intuition tells us that the force is outward—that
centrifugal force is really trying to thrust us radially away from the center,
not on a tangent along the direction of travel.
- Dense things are typically opaque, except for glass and
water, which are “intuitively obvious” exceptions to the rule.
Counterintuitive: Except that they’re not obvious, it’s just the pattern you’ve seen so often that System 1 doesn’t even pick up on the contradiction.
Intuitive thinking is primarily what you’ve experienced and
on patterns you pick up. While it’s
fast, it’s also errorful. Dan Ariely has asked hundreds of Princeton undergraduates the following question: “A bat and ball together cost $1.10.
The bat costs $1 more than the ball.
How much is the ball?” It’s a
simple question, but around 50% of Princeton seniors get it wrong and say that
the bat costs 0.10. (That can’t make
sense. If the ball is 0.10, then “$1
more than the ball” means that the bat is $1.10 – adding the bat and ball
prices together means the bat + ball are $1.20.)
So, why do they get it wrong so often?
Because they don’t check their work. Why?
Because it’s a bit of a hassle, and the answer is so obvious and
apparent (and the stakes are so low) that it’s not worth the effort. This is characteristic of many intuitive
answers—it’s so *obvious* that it’s not worth checking.
Seems to me that there’s not really a contradiction between
System1 (intuition) and System 2 (rational) modes of knowing. Instead, one’s just faster than the
other. Both are useful, both are
necessary. But the great achievement of
science has been to tell us that intuition always has to be checked. In essence, science is the overturning of
intuition by slow, painful, careful System 2 reasoning. Clouds float because cloud-borne water
droplets are really, really tiny and are kept aloft by random air movements.
A note from Dan’s System 2:
Cloud droplets average between 1 and 100 microns. A typical droplet 20 microns in diameter is
4.2 picoliters in volume with a weight of 4.2 nanograms, falling at a terminal
velocity of 0.02 mph. Thus, an updraft
at just over 0.02 mph will keep the cloud in the sky. That’s not much updraft. Or to look at it another way, if the cloud
formed at 10,000 feet, at 0.02 mph it will take nearly 4 days to fall to the
ground.
The trick is to know when to use one mode versus the
other. “Trust the force, Luke” Obi-Wan
whispers into our ears. But the same
time, you can’t build a pan-galactic empire with massive infrastructures based
on your intuitions about mechanical engineering, power grid design, or waste
management systems.
We know there are a number of rather rapid inferences that
are System 1. A while ago Bob Zajonc
showed that determining preference is much faster than cognition… that is, it
anticipates cognition. You see this when
making choices about which kind of thing you prefer over something else—be it
mates, brands of peanut butter or random Chinese characters. As he said, “Preferences need no inferences.” It’s fast and gets there
before you’ve figured out what you really should want. That’s
what System 2 reasoning is for.
It strikes me that we have a lot of brainware dedicated to
specific fast System 1 type tasks, and even those mechanisms get it wrong much
of the time. We’ve all seen various
visual illusions that we KNOW can’t be true—a vertical table that looks much
longer than a horizontal table; two patches of color that look very different
in a scene, but can be easily shown to be the same color when placed
side-by-side.
The strange thing is that visual illusions persist, even
when you know they’re wrong; even when you’ve measure the differences; even
when you put the color patches side by side.
Even though you do the measurements yourself, you can’t convince your
vision system to accept your measurement as truth. It’s as though you haven’t learned anything
in the last minute of experience.
(In the example below, squares A and B are the same color. Use an eyedropper tool to measure the colors and find they're the same. This illusion image is from Wikipedia, created by Edward H. Adelson, Professor of Vision Science at MIT in 1995.)
Even though we DO more vision than anything else we do with
our brains, we still have predictable,
repeatable mistakes in vision. How can
we be better in domains where we don’t have so much hardwired brain?
This is the realm of the cognitive illusion, and is even
more problematic. We know that if you
set up a program with an opt-out vs. an opt-in choice, the differences in
selection rate are huge. In one European
study, the rate of people selecting to be in a organ donation program was
massively different: in Sweden the participation rate is 86%, while in Denmark
the participation rate is only 4%. The
difference is that one is opt-in while the other is opt-out.
Dan Ariely points out that when we’re filling out the form
we
“feel as though we’re making decisions… but the person who designs the form that
is actually making the decision for you.
You like to think that the options don’t influence us, but that’s not
true.”
The illusion is that you have
control over your destiny. That is, you
think your System 2 rational decision-making system will actually express your
deeply-held, fundamental beliefs. “I
believe that organ donation is a social good, so I’ll participate.” That’s the story you might tell yourself. But statistically speaking, the choice is
made for you when the form is designed as opt-out.
We don’t recognize the illusions we all live with
day-to-day, we’ve come to accept them as intuitive and obvious. Normal is what you’ve grown up with, the
non-intuitive, confusing, difficult high-technology world is whatever was invented
after you turned 8, and stopped seeing everything as pre-destined.
Isn't that intuitive?