The Folly of Fools: The Logic of Deceit and Self-Deception in Human Life

Home > Other > The Folly of Fools: The Logic of Deceit and Self-Deception in Human Life > Page 17
The Folly of Fools: The Logic of Deceit and Self-Deception in Human Life Page 17

by Robert Trivers


  The psychological and immune systems are deeply intertwined, cause and effect go in either direction, and it is hardly possible for one system to react without affecting the other. For reasons that are not always obvious, self-deception appears to have strong immune effects, usually according to the rule more self-deception, lower immune strength, but occasionally, more self-deception, better immune function.

  This field is still in its infancy. Some interesting things are known, but much more remains to be found out. Which levels of information suppression are associated with what immune effects? And what chemicals are common to the brain and the immune system, leading to important trade-offs between the two? And what questions do we not even know enough to ask?

  CHAPTER 7

  The Psychology of Self-Deception

  How do we achieve our various self-deceptions? If not in precise mechanistic terms, then in psychological ones, what are the psychological processes that help us achieve self-deception? We both seek out information and act to destroy it, but when do we do which and how do we do it? To give an answer to this, we need to trace the flow of information from the moment it arrives until the moment it leaves, that is, is represented to others. From the “rooter to the tooter,” as we say for pigs. At every single stage—from its biased arrival, to its biased encoding, to organizing it around false logic, to misremembering and then misrepresenting it to others, the mind continually acts to distort information flow in favor of the usual good goal of appearing better than one really is—beneffective to others, for example. Misrepresentation of self to others is believed to be the primary force behind misrepresentation of self to self. This is way beyond simple computational error, the problems of subsampling from larger samples, or valid systems of logic that occasionally go awry. This is self-deception, a series of biasing procedures that affect every aspect of information acquisition and analysis. It is systematic deformation of the truth at each stage of the psychological process. This is why psychology is both the study of information acquisition and analysis and also the study of its continual degradation and destruction.

  One important fact is worth stressing at the outset. Self-deception does not require that the truth and falsehood regarding something be simultaneously stored—as in our example of voice recognition (Chapter 3). Falsehood alone may be stored. As we saw for the old-age positivity bias (Chapter 6), the earlier the information is shunted aside—or indeed entirely avoided—the less storage of truth occurs and the less need there will be for (potentially costly) suppression later on. At the same time, since less information is stored, there are greater potential costs associated with complete ignorance. As time after acquisition increases, the choice between suppressing and retaining the truth should be more subtle and complex. The study of exactly how these conflicting forces have played out over time is a completely open field whose exploration will be most revealing.

  In what follows, I begin with a review of some of the biasing that takes place during information processing. This is by no means an exhaustive look but more an impressionistic one of the ways in which various psychological processes support a deceptive function. This may include biases in predicting future feelings. Especially important are the roles of denial, projection, and cognitive dissonance in molding deceit and self-deception.

  AVOIDING SOME INFORMATION AND SEEKING OUT OTHER

  However much we champion freedom of thought, we actually spend much of our time censoring input. We seek out publications that mirror or support our prior views and largely avoid those that don’t. If I see yet another article suggesting the medical benefits of marijuana, you can trust me to give it a careful read; an article on its health hazards is worth at best a quick glance. Regarding tobacco, I couldn’t care less. The scientific facts were established decades ago and it has been years since my last cigarette. So this bias in my attention span is both directly adaptive—I smoke marijuana, so I am interested in its effects—and serves self-deception, because I hype the positive and neglect the negative, the better to defend the behavior from my own inspection and that of others.

  A lab experiment measured this kind of bias precisely by confronting people with the chance that they might have a tendency toward a serious medical condition and telling them a simple test would suggest whether they were vulnerable. If they applied their saliva to a strip of material and it changed color, this indicated either vulnerability or not (depending on experimental group). People led to believe that a color change was good looked at the strip 60 percent longer than did those who thought it would be bad (actually the strip never changed color). In another experiment, people listened to a tape describing the dangers of smoking, while being asked to pay attention to content. Meanwhile, there was some background static and the subjects had the option of decreasing its volume. Smokers chose not to decrease the static, while nonsmokers lowered the level, the better to hear what was being said.

  Some people avoid taking HIV and other diagnostic tests, the better not to hear bad news. “What I don’t know can’t hurt me.” As expected, this is especially likely when little or nothing can be done either way. It is also not surprising that those who feel more secure about themselves are more willing to consider negative information. In short, we actively avoid learning negative information about ourselves, especially when it can’t lead to any useful counteraction and when we feel otherwise insecure about ourselves. Self-deception is here acting in service of maintaining and projecting a positive self-view.

  In many situations, we can choose what to concentrate on. At a cocktail party, we could overhear two conversations. Depending on which views we wish to hear, we may attend to one conversation instead of the other. We are likely to be aware of the general tenor of the information we are avoiding but none of its details, so here again biased processes of information-gathering may work early enough to leave no information at all that may later need to be hidden. In one experiment, people were convinced that they were likely—or highly unlikely—to be chosen for a prospective date. If yes, they spent slightly more time studying the positive rather than negative attributes of the prospective date, but if no, they spent more time looking at the negative, as if already rationalizing their pending disappointment.

  BIASED ENCODING AND INTERPRETATION OF INFORMATION

  Assuming we do attend to incoming information, we can still do so in a biased way. One experiment invited people to look at a figure that could be either a capital B or the number 13 (or a horse or a seal) and were told the stimulus could be either a letter or a number (or a farm animal or ocean animal). Having been provided differential food reward for the general categories ahead of time, people quickly developed a sharp perceptual bias in the appropriate direction on items presented for only four hundred milliseconds, that is, ones just reaching consciousness. Eye tracking showed that the first look was usually toward the preferred category (about 60 percent). These studies suggest that the impact of motivation on processing information extends to preconscious processing of visual stimuli and thus guides what the visual system presents to conscious awareness. Similar work has now been done using colors.

  The point is that our perceptual systems are set up to orient very quickly toward preferred information—in this case, shapes associated with food rewards. This itself has nothing to do with deceit and self-deception—it will often give direct benefits. But the same quick-biasing procedure is available to us when the information is preferred because it boosts our self-esteem, or our ability to fool others. There are few more powerful forces in the service of self-deception than personal fantasies, so when these are aroused, selective attention is expected to be especially intense.

  The related effect was shown sixty years ago: hungrier children, when asked to draw a coin, draw coins larger. Instruments to gain satisfaction (money buys food) are more attractive and are perceived as being larger. Recent confirmation shows that a glass appears larger to you when you are thirstier, especially if attention is called to your thirst
, and even garden implements appear larger if gardening has subliminally been linked to suggestions that it is fun.

  Our initial biases may have surprisingly strong effects. In one experiment, people were preselected for strong attitudes for and against capital punishment. They were then presented with a mixed bag of facts supporting both positions. Instead of leading to group cohesion, this action split the group more sharply. Those who were already against capital punishment now had a new set of arguments at hand, and vice versa. Biased interpretation ran the process. Those in favor of capital punishment accepted pro arguments as sound and rejected anti arguments as unsound. As before, self-affirming thoughts were negatively associated with this behavior—think better of yourself, you practice less self-deception. One important implication is that self-deception is a force that often drives people apart—certainly friends, lovers, neighbors—although under common group aims, such as war, shared self-deceptions are also uniquely powerful in binding people together.

  BIASED MEMORY

  There are also many processes of memory that can be biased to produce welcome results. We more easily remember positive information about ourselves and either forget the negative or, with time, transmute it to be neutral or even positive. Differential rehearsal, as in telling others, can itself produce the effect, an example of self-deception at the end of the process (the “tooter”) affecting earlier processes. Complementary memory biases may actively work in the same direction. When given a “skills class,” people remember their skills prior to the class as being worse than they rated them at the time, probably to create an illusion of progress. They then later misremember their actual performance after the class as being better than it was, presumably in service of the same delusion. What we are doing here is producing a consistent set of biases in our own favor by a series of biased memories.

  Memories are continually distorted in self-serving ways. Men and women both remember having fewer sexual partners, and more sex with each partner, than was actually true. People likewise remember voting in elections they did not and giving to charity when they did not. If they did vote, they remember supporting the winning candidate rather than the one they actually voted for. They remember their children as being more precocious and talented than they were. And so on.

  Although people often think of memory as a photo whose sharpness gradually degrades with time, we know that memory is both reconstructive and easily manipulated. That is, people continually re-create their own memories, and it is relatively easy to affect this process in another person. If a police officer asks a witness about a nonexistent red sports car right near an accident, the officer will often learn about the red sports car in subsequent questioning—it can sometimes end up as one of the most vividly remembered details of the accident itself. As mentioned, differential rehearsal of material after the fact can produce reliable biases in memory.

  Take another example. Health information can easily be distorted in memory even when it is presented in a clear and memorable fashion. People were given a cholesterol screening and then one, three, and six months later tested for the memory of the result. Respondents usually (89 percent) recalled their risk category accurately and their memory did not decay with time, but more than twice as many people remembered their cholesterol level as lower rather than higher than it actually was. This same kind of memory bias is true of daily experiences in which people recall their good behavior more easily than bad but show no such bias in recalling the behavior of others.

  Or we can invent completely fictitious memories. As has been said, “My memory is so good I can remember things that never happened.” One case is memorable in my own life. For many years I told the story of how in 1968 I went deep into the bowels of Harvard’s Widener Library to find a book coauthored by my father in 1948, published by the State Department, which laid out the de-Nazification procedures for all Nazis too unimportant to be hung at Nuremberg. It was a complex system of graded steps. If you were a member of SA, two slaps on the wrist; if SS, you lost your job for five years—that kind of thing. Yet almost none of this is true. No such book exists. Yes, the trip to the bowels took place and a book on Nazis was located with my father as a coauthor and it was published by the State Department. Only it was published in 1943 and is a minor piece on the structure of Nazi organizations in Nazi-occupied territories. Hardly the basis for the reinvention of Germany, but is this not the point of false memory—to improve things, especially appearances? I added nice little touches along the way. I liked to say that I trusted no one, including myself, and thus went to the bowels to see whether this family story was true. But this added to the falsehood, since there really was no “family story” about this minor 1943 work, and is a general feature of false-memory construction—new details are added that support the general argument and then become part of memory.

  One can even reverse exactly who is saying what to whom. Gore Vidal remembers an interview with Tom Brokaw on NBC’s morning Today Show in which Brokaw began by asking about Vidal’s writings on bisexuality, to which Vidal replied that he was there to talk politics. Brokaw persisted with bisexuality; Vidal stood firm until they concentrated on politics. Yet years later, when Brokaw was asked what his most difficult interview had been, he cited his interview with Vidal. Reason: Vidal kept insisting they talk about bisexuality when all he wanted to discuss was politics. Positions exactly reversed—and, as expected, in the service of self-improvement: Brokaw looks better being interested in politics than in bisexuality.

  In arguments with other people, lab work shows that we naturally tend to remember the good arguments on our side and the poor ones on the other, and to forget those that turn out badly for us and good for the other. This bolsters our own side and image, of course, which presumably is its function. Memory distortions are more powerful the more they are motivated to maintain our self-esteem, to excuse failures or bad decisions, and to push into the deeper past causes of current problems. Thus, most people maintain the illusion of improvement, where such mistakes as must be acknowledged can at least be attributed to the failings of an earlier version of oneself.

  RATIONALIZATION AND BIASED REPORTING

  We reconstruct internal motives and narratives to rationalize otherwise bad or questionable behavior. We can attribute behavior to external contingencies rather than internal, thereby helping defend ourselves. So a general belief that cheating is not bad—or is unintentional or occurs in a world without free will—will all serve to rationalize our cheating, as indeed they do.

  Biases show up in unexpected places, even when there are no clear benefits or costs at issue. The classic experiment in this domain was beautifully designed to put people in an awkward situation with one of two escapes. People were offered the chance to sit next to a crippled person or one who was not. Each was watching a television set in front of him or her. Sometimes the two TVs had the same show, sometimes different ones. When it was the same show, people preferentially chose to sit next to the handicapped person, as if demonstrating their lack of bias, but if the two TVs had different shows, people chose to sit away from the crippled person, as if now having a justification (more interesting show) for an otherwise arbitrary choice. Similarly, a meta-analysis of many studies shows that white Americans choose to help black Americans more or less equally (compared to helping whites) but not when they can rationalize less helping on grounds such as distance or risk. Here people are not denying or misremembering their behavior—rather, they are denying the underlying intention and rationalizing it as the product of external forces. This has the advantage of reducing their responsibility for behavior performed.

  A belief in determinism can provide a ready excuse for misbehavior, just as can unconsciousness: the “I had no choice” defense. Relatively deterministic views of human behavior may provide some cover for socially malevolent behavior. Experimentally inducing a deterministic view (reading an essay on how genes and environment together determine human behavior) increases cheating on
a computer-based task that permits cryptic cheating. What this work shows is that by manipulating a variable that reduces personal responsibility, we easily induce immoral behavior in ourselves (at least as viewed by others).

  PREDICTING FUTURE FEELINGS

  It is an interesting fact that we show systematic biases in our ability to predict our own future feelings. We make systematic errors in the process, under the general rule that what we are feeling at the present will extend into the future. When imagining a good outcome, we overestimate our future happiness, and vice versa for a bad one. It is as if we assay our current feelings and then project them into the future. We do not imagine that we will “regress to the mean,” that is, return naturally to the average value of happiness. We do not assume we will be less happy in the future than our current state of happiness or happier in the future if we are currently down. Thus, one week after the 2004 US elections, Kerry supporters were less dejected than they thought they would be and Bush supporters, less ecstatic.

  There is evidence that we make similar mistakes when trying to predict the feelings of others, whether friends or strangers. We overestimate the effect of an emotional event on their future feelings, much as we do for ourselves. Indeed, our forecasting of them is positively correlated with their own, but neither is very predictive of the future.

 

‹ Prev