EFTA01157466
EFTA01157468 DataSet-9
EFTA01157487

EFTA01157468.pdf

DataSet-9 19 pages 11,856 words document
P17 P22 V14 P24 V9
Open PDF directly ↗ View extracted text
👁 1 💬 0
📄 Extracted Text (11,856 words)
Al memo No. 603, November, 1980, published in Minsky, Marvin, "Jokes and their Relation to the Cognitive Unconscious," Cognitive Constraints on Communication, Vaina and Hintikka (eds.) Reidel, 1981. This version of the paper may have misprints because it is a prepublication draft. JOKES and the Logic of the Cognitive Unconscious Marvin Minsky, MIT Abstract: Freud's theory of jokes explains how they overcome the mental "censors" that make it hard for us to think "forbidden" thoughts. But his theory did not work so well for humorous nonsense as for other comical subjects. In this essay I argue that the different forms of humor can be seen as much more similar, once we recognize the importance of knowledge about knowledge and, particularly, aspects of thinking concerned with recognizing and suppressing bugs—ineffective or destructive thought processes. When seen in this light, much humor that at first seems pointless, or mysterious, becomes more understandable. Introduction A gentleman entered a pastry-cook's shop and ordered a cake; but he soon brought it back and askedfor a glass ofliqueur instead. He drank it and began to leave without havingpaid. The proprietor detained him. "You've not paidfor the liqueur." "But 1 gave you the cake in exchange for it." "You didn't payfor that either" "But 1 hadn't eaten it". —Freud (1905). In trying to classify humorous phenomena, Sigmund Freud asks whether this should be called a joke, 'for thefact is we do not yet know in what the characteristic ofbeing a joke resides." Let us agree that some of the cake-joke's humor is related to a logical absurdity—leaving aside whether it is in the logic itself, or in keeping track of it. Later Freud goes on to ask what is the status of a "knife without a blade which has no handle?" This absurdity has a different quality; some representation is being misused—like a frame without a picture. Freud, who never returned to the subject after writing his 1905 book on the theory of jokes {0}, suggested that "censors" in the mind form powerful, unconscious barriers that make it difficult to think "forbidden" thoughts. But jokes can elude these censors—to create the pleasure of unearned release of psychic energy, which is discharged in the form of laughter. He explains why jokes tend to be compact and condensed, with double meanings: this is to fool the childishly simple-minded censors, who see only innocent surface meanings and fail to penetrate the disguise of the forbidden wishes. But Freud's theories do not work as well for humorous nonsense as for humorous aggression and sexuality. {0) In this essay I try to show how these different forms of humor can be seen as much more similar, once we make certain observations about the nature of commonsense reasoning. Here is our thesis: I. Common sense logic is too unreliable for practical use. It cannot be repaired, so we must learn to avoid its most common malfunctions. Humor plays a special role in learning and communicating about such matters. 2. It is not enough to detect errors in reasoning; one must anticipate and prevent them. We embody much of our knowledge about how to do this in the form of "censors" that suppress unproductive mental states. This is why humor is so concerned with the prohibited. 3. Productive thinking depends on knowing how to use Analogy and Metaphor. But analogies are often false, and metaphors misleading. So the "cognitive unconscious" must suppress inappropriate comparisons. This is why humor is so concerned with the nonsensical. EFTA01157468 4. The consequences of intellectual failure are manifest in one's own head, while social failures involve other people. Intellect and Affect seem less different once we theorize that the "cognitive unconscious" considers faulty reasoning to be just as "naughty" as the usual "Freudian" wishes. 5. Humor evolved in a social context. Its forms include graciously disarming ways to instruct others about inappropriate behavior and faulty reasoning. This deviousness makes the subject more confusing. Our theory emphasizes the importance of knowledge about knowledge and, particularly, aspects of thinking concerned with recognizing and suppressing bugs—ineffective or destructive thought processes. When seen in this light, much humor that at first seems pointless, or mysterious, becomes more understandable. {1} I. PROBLEMS OF COMMON SENSE REASONING When you tell a young child "Iam telling a lie" then, if he is old enough to reason so, he will think: "Ifthat isfalse, then he's not telling a lie. But, then it must be true. But then, it must be a lie,for it says so. But then...". And so on, back and forth. A child might find this situation disagreeable for several reasons. It challenges the belief that propositions are always either true or false. It threatens to propagate through his knowledge- structure, creating other inconsistencies. And he can make no progress when his mind returns again and again to the same state. {2} Common sense can go awry in endless ways. Beliefs can be wrong from the start, one can make mistakes from each step to the next, and one can wander aimlessly, getting nowhere. But before we discuss these, we should observe what actually happens when you say things like "this statement is false." Often the listener first seems puzzled, then troubled, and finally laughs. "That's funny," he says. 'Tell me another liar joke". THE PROBLEM OF TRUTH: How do we know where to begin? The conclusions of even die best reasoning can be no better than its premises. In mathematics, this is of little concern (because one cares more where premises lead than where they come from). But in real life, few propositions are perfectly trustworthy. What does one do when an accepted "fact" turns out false? One of my children was once entranced by an ornamental fish-shaped bowl with four short legs. After a while she announced, somewhat uncertainly: "Some fish don't have legs." We never know anything for certain. What should one do upon reaching a conclusion that appears false erase all the assumptions? When a long held belief turns false, should one erase all that has been deduced from it? When an acquaintance once tells a lie, should one reject everything else he ever said? There is no simple, single rule: each person must find his own ways to maintain his knowledge about his knowledge. (Ref. Truth Maintenance Systems) WHENCE THE RULES OF INFERENCE? How do we know how to infer? Most people believe that if most A's are B's, and if most B's are C's, then most A's are C's. Though false, this has great heuristic value, especially for children who encounter few exceptions. Psychologically, I see no great difference between heuristic and logical reasoning; deduction is "just another" kind of evidence. We firmly believe a deduced conclusion only when it seems plausible on other grounds as well. At one time, many philosophers held that faultless "laws of thought" were somehow inherent, a priori, in the very nature of mind. This belief was twice shaken in the past century; first when Russell and his successors showed how the logic men employ can be defective, and later when Freud and Piaget started to reveal the tortuous ways in which our minds actually develop. After Russell observed that the seemingly frivolous "Who shaves the Barber, who shaves everyone who does not shave himself?" was a truly serious obstacle to formalizing common sense logic, he and others tried to develop new formalisms that avoided such problems by EFTA01157469 preventing the fatal self-references. But the proposed substitutes were much too complicated to serve for everyday use. I am inclined to doubt that anything very resembling formal logic could be a good model for human reasoning. (The paper by Hewitt and Komfeld might suggest a possible avenue of compromise (Ref. BAK).) In particular, I doubt that any logic that prohibits self-reference can be adequate for psychology: no mind can have enough power without the power to think about Thinking itself. Without Self-Reference it would seem immeasurably harder to achieve Self- Consciousness which, so far as I can see, requires at least some capacity to reflect on what it does. {3} If Russell shattered our hopes for making a completely reliable version of commonsense reasoning, still we can try to find the islands of "local consistency," in which naive reasoning remains correct. It seems that only certain kinds of expressions lead to paradoxes and inconsistencies, and it seems worth taking some risks, gambling for greater power provided we can learn, over time, to avoid the most common disasters. We all know the legend of the great mathematician who, warned that his proof would lead to a paradox if he took one more step. He replied "Ah, but I shall not take that step." (4} One would miss the point to treat this as a "mere" joke. What it means, really, is that we build into our minds two complementary functions: We work to discover "islands of power" within which commonsense reasoning seems safe. We work also to find and mark the unsafe boundaries of those islands. In civilized communities, guardians display warnings to tell drivers about sharp turns, skaters about thin ice. Similarly, our philosophers and mathematicians display paradigms like the Barber, the Tortoise, and the Liar to tell us where to stop and laugh. I suggest that when such paradigms are incorporated into the mind, they form intellectual counterparts to Freud's emotional censors. This would help explain why purely logical nonsense so often has the same humorous quality as do jokes about injury and discomfort the problem that bothered Freud. The cake-joke reminds us, somewhat obscurely, to avoid a certain kind of logical absurdity lest we do ourselves some vaguely understood cognitive harm. Hence our thesis: since we have no systematic way to avoid all the inconsistencies of commonsense logic, each person must find his own way by building a private collection of "cognitive censors" to suppress the kinds of mistakes he has discovered in the past. HEURISTIC CONTROL OF LOGIC: How do we know what next to do? I once tutored a student having trouble with middle-school geometry. I began to explain the axioms, and how proofs were structured. "I understand all that," he said, "only I was sick the day the teacher explained how to find the proofs." It is not enough just to know the principles of reasoning; one must know also when to apply them. We each know millions of facts, and perhaps millions of rules of inference. Which to apply to what, and when? There is a basic problem of direction, of not meandering, lest one aimlessly derive billions of inferences, all perfectly logical but none relevant to any goal. First, some "plan" is required. Next, one must avoid circling returning to the same place again and again. Finally, to avoid confusion, one needs an administrative structure to keep track of what one is doing, and why. The new science called Artificial Intelligence is concerned with just such issues of the efficiency and effectiveness of Reason matters rarely discussed in Logic or Philosophy, which focus on verifying that proofs are valid, or that arguments are sound, rather than on how proofs are discovered. Much has been learned, in "Al", about how to avoid excessive meandering and confusion, by using goal-structures and plans techniques for insuring progress. Using such methods, some modern computer programs can thread their ways through some quite complicated situations. EFTA01157470 Nevertheless, the problem of meandering is certain to re-emerge once we learn how to make machines that examine themselves to formulate their own new problems. Questioning one's own "top-level" goals always reveals the paradox-oscillation of ultimate purpose. How could one decide that a goal is worthwhile unless one already knew what it is that is worthwhile? How could one decide when a question is properly answered unless one knows how to answer that question itself? Parents dread such problems and enjoin kids to not take them seriously. We learn to suppress those lines of thoughts, to "not even think about them" and to dismiss the most important of all as nonsensical, viz. the joke "Life is like a bridge." "In what way?" "How should I know?" Such questions lie beyond the shores of sense and in the end it is Evolution, not Reason, that decides who remains to ask them. II. CENSORSHIP Just like taboos in human societies, certain things must not be thought inside the Mind. The best way for a child to learn not to do a certain bad thing would be to learn not to even to think of it. But isn't that like trying "not to think of a monkey"? Contrast two ways: (i) suppress an idea already in the mind or (ii) prevent it from being thought in the first place: (i) Stop thinking that! (ii) Don't even (or ever) think that! It is easy to begin to make a type (i) censor: wait for the specified "bad" event to happen, and then suppress it. But how does one prevent it from recurring? It seems harder to begin to make a type (ii) censor, because it must be able to recognize the repressed thought's Precursors— but it is easier to see what to do next, since each Precursor usually leads to several options, and suppressing one still leaves the others. So we shall discuss only the second censor-type. So, our censor-builder has to learn to recognize Precursors mind-brain states that precede a recognizable (and, here, to be prohibited) activity. To do this, it will need a short term memory (to remember what just happened) and a long term memory (to store the result of learning). The latter may eventually become quite large, because a prohibited event may have many different precursors. In any case, a experienced type (ii) censor can recognize its joke by the situation and need not wait for the punch line. A type (i) censor makes you wait till the comedian finishes: only then can you complain "Oh, I've heard that one before!" To place these in a larger framework, Let's consider a simple "two-brain" theory. An "A-brain" has sensory inputs from the world, and motor outputs to the world. The B-brain's inputs come from the interior of A so B can perceive "A-states"— and its outputs go into A, so B can affect activities in A. Thus, B can "see" what is happening inside A, and act to influence it, just as A can "see" and affect what happens in the world. B need not and probably can not—know what A-events "mean," vis-a-vis the world, but B is in a position to recognize such metapsychological conditions such as A being "meandering, circling, or confused." (51 When a B-censor acts, it must disturb the A-brain so as to suppress the undesired activity. (It would be even better for B to remember, from past events, which is a good way to go, but that is outside this essay's concern.) In any case, the point is that precursor-sensitive censors can do their work before the problems that they evade actually arise. Probably, also, they can do this so quickly and gently as to produce no noticeable mental phenomenology. This would explain why censors are (in the) unconscious. The censorship theory explains why a joke is not so funny if you've heard it before; this is because a new censor has been constructed, or an old one extended. Freud touches on "novelty" as a component of humor, but never dwells on why old jokes get few laughs. I presume that he simply considered it too obvious to mention, that censors are learners. How big must the censor memory be, to protect us from naive reasoning mistakes? Probably not so large for formal logic, considering how rarely we discover a new paradox. But for avoiding nonsense in general, we might accumulate millions of censors. For all we know, this "negative EFTA01157471 meta-knowledge"—about patterns of thought and inference that have been found defective or harmful—may be a large portion of all we know. Consider the activities called play and practice. The insensitive learning theories of behavioristic psychology regard play not at all, and see practice as reinforcing something repetitive. But practice (I conjecture) is often far from mere repetition and refinement of the same thing; often it is exploratory, testing out a skill's minute variations and perturbations—and learning which of them to enhance or suppress. Similarly play, (commonly seen as "mere") is often also an exploration of variations on a larger scale. Many other everyday activities can so be seen as ways to learn to avoid bugs and mistakes. I know a young child whose sense of humor tends toward jokes like "what if the spoon were rubber," apparently observing that a flexible spoon would be absurd because the food would fall out of it. Is he enforcing a "must-be-rigid" property of some spoon-frame, or is he censoring the use of some spoon-frame that lacks the property? The humor-behavior of children also needs more study. III. MEANING AND METAPHOR 7ivo villagers decided to go bird-hunting. They packed their guns and set out, with their dog, into thefields. Near evening, with no success at all, one said to the othet; "We must be doing something wrong". "Yes", agreed hisfriend. "Perhaps we're not throwing the dog high enough." When you want to fasten a screw and therefore reach for a certain screwdriver, your mind has chosen to see that screwdriver as a screw-driver; you could have seen it as a kind of dull knife, or as a hammer without a head. When we see something only in its "intended" aspect, we are "confusing the thing with itself." As Korzybski (Ref. AK) intoned, "whatever a thing is, it is not". (6), (7) FRAMES: I have suggested in (Ref. MM-74) that perceptions are ordinarily interpreted by the mind in terms of previously acquired description-structures called Frames. A frame is a way to represent a stereotyped situation, like being in a certain kind of room, or going to a certain kind of party. Attached to each frame are several kinds of information; some about how to use the frame, some about what one might expect to happen next, some about what to do if those expectations are not confirmed, and so forth. This theory was proposed to explain the speed and virtual absence of noticeable phenomenology in perceiving and thinking, and here I propose to sketch just enough of it to explain some features, and some "bugs," of reasoning. Then we can return to unconscious censorship and error-correction. Each frame includes, among other things, a variety of terminals to which other frames are attached. Thus a chair-frame specifies that a (certain kind of) chair has a seat, a back, and four legs. The details of these would be described, not in the chair-frame itself, but in other frames attached to its terminals. Each frame includes also a set of features which, if enough of them are present, may activate the Frame. So, when you see enough parts of a chair, these will activate one of your chair-frames which, in turn, will activate the sub-frames attached to its terminals. These, then, will "look for" other chair-parts that were not recognized at first—because of being unusual, or partially hidden from view, or whatever. Finally, if some elements required by the frame are not seen at all—one rarely sees all the legs of a chair, and never all the sides of a box —the missing elements are supplied by default. This is easy because most terminals of most frames have certain sub-frames already attached as default assignments. When one reads in a story about some shoe or some chair, these cause one, "by default" to assume a certain kind of shoe or chair. The concept of default goes much further. When one sees a person in a sitting posture then, even if every part of his chair is hidden from view, one will pseudo-see a chair under him. Unless your attention is drawn to the fact, you never notice that no chair was actually seen. This is EFTA01157472 because the "sitting" frame includes a "must-be-supported-by" sub-frame terminal, and this is attached to some "typical" chair-frame, for its default assignment. (8) The chair-frame that is selected can depend, however, on the context. for default assignments are weak and easy to "displace". If there are other chairs around, the invisible chair will be assumed to be like one of them. If the scene is set in a park, then a park-bench frame might be activated to serve as default. If one then noticed an arm-chair arm, the system would replace the weakly- attached bench-frame by one that better suits what was seen—and one now "sees" an armchair. According to the theory in (Ref. MM-74), this is done very swiftly because the "corresponding" terminals of related frames are already pre-connected to one another. This makes it easy to change a faltering interpretation or a frustrated expectation. Shifting from one related frame to another should be so fast and efficient as to be imperceptible to introspection. This is why one so easily recognizes any chair as "a chair", even though particular chairs are so different from one another. I do not suggest that all this happens by magic; the interconnecting network is constructed over a lifetime of experience. In (Ref. MM-77) I discuss how new frames arise usually as revised versions of older ones, bringing those "common terminals" along with them. In (Ref. MM-79) are more details, but one must read between the lines of that paper because it does not use the terminology of frames. Frames and frame-systems are used at conceptual as well as perceptual levels, and there we find other kinds of frame-systems—families of interconnected frames—that are not transformed so easily, hence more effort is noticed. Consider, for example, Wittgenstein's paradigmatic question about defining "game." (Ref. Wittgenstein) The problem is that there is no property common to all games, so that the most usual kinds of definition fail. Not every game has a ball, nor two competing teams; even, sometimes, there is no notion of "winning." In my view, the explanation is that a word like "game" points to a somewhat diffuse "system" of prototype frames, among which some frame-shifts are easy, but others involve more strain. The analogy between football and chess is strained only a little, more with solitaire, and soon. Shifting from one familiar kind of kitchen-chair to another is imperceptible, but changing a park-bench to an arm-chair would be strain enough to "surprise". Now I propose that much of commonsense logic itself is based on learning to make shifts between frames that have terminals in common. For example, if a situation fits a frame like A implies B, and B implies C, a simple frame-shift re-represents it as A implies C. Seen in this light, Freud's cake story appears to display some sort of incorrect-logic script in which each consecutive pair of sentences matches some such tried-and-true kind of reasoning step. I presume that when we "understand" this sort of story, we represent it in our minds as a series of pairs of overlapping assignments of things to terminals of such frames. And somewhere along the way, in the cake story, there is an improper assignment-change. Is it the payment moving from the cake to the drink? Is it a pivot between "owns" and "possesses?" Each listener must make his own theory of what is wrong—and devise his own way to avoid this confusion in the future. Some people will do better at this than others. METAPHOR: All intelligent persons also possess some larger-scale frame-systems whose members seemed at first impossibly different—like water with electricity, or poetry with music. Yet many such analogies—along with the knowledge of how to apply them—are among our most powerful tools of thought. They explain our ability sometimes to see one thing—or idea— as though it were another, and thus to apply knowledge and experience gathered in one domain to solve problems in another. It is thus that we transfer knowledge via the paradigms of Science. We learn to see gases and fluids as particles, particles as waves, and waves as envelopes of growing spheres. How are these powerful connections discovered? For the simple ones, there is no great problem: some frames are easily recognized as similar because their terminals accept the same sorts of entities; these could be located and classified by simple algorithms, e.g., searches for best EFTA01157473 match. As for those most subtle, once-in-a-lifetime insights, whose analogical powers are hidden deep in the procedural structures that operate on them, we hardly need a general theory to account for them since—like favorable evolutionary mutations—few are ever discovered by any single individual, and those can be thereafter transmitted through the culture. In any case, putting aside the origins of those rare, greatest insights, each individual must have his own ways to build new connections amongst his frames. I will contrast particular methods against general methods, arguing that the two have problems that seem somewhat opposite in character—that the errors of "particular" methods can be managed additively„ while bugs in the "general" methods must be repaired subtractively. This last point will eventually bring us back to censors. PARTICULAR ANALOGIES: In the course of thinking, we use different frames from one moment to the next. But frequently one of the active frames will fail—"that's not a door, only a big window." Winston (PHW-70) suggests that whenever such an error is (somehow) detected, described, and corrected, we can attach to the failing frame a "pointer" to some other frame that has been found to work in this circumstance. The pointer must contain, of course, a description of the failure circumstance. A family of frames connected in such a way is called a difference network. (9} We can explain (Ref. MM-74) the definition difficulty (e.g., that of defining "game") by supposing that such words point not to any single frame, but into such a network. Any such link between two frames implies a generalization of the form "A's are like B's, except for D". Thus, such a link is a fragment of an analogy. Of course, like any generalization, it will likely soon fail again and need refinement. Winston's thesis (Ref. PHW-70) suggests ways to do this. The important point here is that, particular analogies discovered in the course of experience, can be remembered by adding positive, active links between frames. GENERAL ANALOGY METHODS: What if one is confronted with a novel situation that does not arouse any particular frame? Then, it makes sense to try some "general" method., e.g., to compare the situation to some large class of frames, and, select the one that "best fits." Such a method can do much better than chance, but what if it yields a result that does more harm than good? I will argue that one now must build a censor, and that there is a general principle here that learning theorists have not appreciated: positive general principles need always to be supplemented by negative, anecdotal censors. For, it hardly ever pays to alter a general mechanism to correct a particular bug. Almost every instance-specific modification of a best-match mechanism would reduce its general usefulness. So when the general mechanism yields a bad result, one can only remember to suppress its subsequent appearances—that is, to build a censor. If a child wants his sibling's toy, he will first try seizing it—the most general way to get things one wants. Once the parents see to it that this is censored, then he can find other ways. But he must not learn in general not to take things he wants, lest he become helpless. Some forms of humor, notably puns, turn on changing the meaning-sense of a word. (Besides the easily distinguished dictionary senses, most words have also many others that would never be separated in a dictionary. "Lift," for example, has different implications when an object weighs one gram, or a thousand, {7} and really to understand lifting requires a network for making appropriate shifts among such different micro-senses.) While verbal sense-shifting can be funny, and even useful, it is dangerous, and especially hazardous to be subject to the fortuitous, meaningless sense-shifts that depend on superficial word-sound similarities. In a fragment of a schizophrenic's transcript, the patient sees a penny in the street, says "copper, that's a conductor," then must run to a street car to speak to the conductor. Perhaps this disorder is one of frame-shift control, either disabling the bad-analogy suppressors or irresponsibly enhancing the general analogy-finder. The element that seems to me most common to all the different kinds of humor is that of unexpected frame-substitution, in which a scene is first described from one viewpoint and then EFTA01157474 suddenly—typically by a single word—one is made to view all the scene-elements in another, quite different way. Some such shifts are insightful, of course, while others are mere meaningless accidents. Next we turn to a kind that could be turning points in each individual's personal evolution. IV. TRAUMATIC COGNITIVE EXPERIENCES "Yields truth when appended to its own quotation" yields truth when appended to its own quotation.—W V Quine In the popular psychology of our day, Intellect is seen as straightforward, deliberate, conscious, and emotionally neutral; but in regard to emotional matters, the public has come generally to accept the psychoanalytic view that Affect is dominated by unknown terrors and traumas, lurking in the unconscious since childhood. Now I want to challenge this affect-intellect "dumbbell". I certainly do not mean to suggest that there are no differences of kinds in mental affairs—only that this particularly popular distinction, however useful in everyday life, does more harm than good in psychology. (Ref. MM-79) In any case, the Affect-Intellect distinction would lose much of its force if Reason, too, employed a powerful "cognitive unconscious"—and that is exactly what I shall argue: that Intellect, too, has its own buried secrets. In Freud's scenario, the ego develops largely in contexts of fear of deprivation, punishment or mutilation; of anxiety about uncertainty and insecurity; terror of losing the esteem or person of parent or attachment-figure. New families of censors—new domains of repression—are created when wishes conflict enough with reality to justify attempting to keep them from becoming conscious. Does anything like that happen in the intellectual sphere? One might suppose not, because a necessary element is missing—that of a beloved—or punitive—authority figure. However, while Freudian taboos originate from outside, the child needs no external authority-figure to point out his gross cognitive failures; he needs no parent to scold him, when an encounter with paradox throws his mind into a frightening cyclone. The momentary loss of mental control should provoke anxiety in its own right. But, one might ask, if we bear the scars of frightening cognitive experiences, why do they not reveal themselves (like the affect-laden ones) in nightmares, compulsions, phobias, and the like? Perhaps they do; it would not show in the interpretations of present-day psychiatry. But every teacher knows (and fears) the rigid inhibition of children's cognitive phobias: "I don't want to learn this; I couldn't possibly do that". Let us speculate on how such fears might originate. Consider the paradox of Nearness: Every child must once have said to himself: Hmm. Ten is almost Eleven. And Eleven is nearly Twelve. And so on; Ninety-Nine is nearly a Hundred. But then Ten must be almost a Hundred! To an adult, this is not even a stupid joke. But we each must once have thought something like: "there is obviously something wrong here. Is it in my premises or in my logic? Well, what's my premise? Obviously, that "if A is near B, and if B is near C, then A must be near C." Nothing wrong with that. So there must be something wrong with my logic. But I'm only using something like: "If A implies B, and if B implies C, then A implies C." "How could that be wrong? No way!" To be sure, not everyone remembers such experiences as frightening. In fact, ad hominem, readers of essays like this one would more likely complain that they like such problems and cannot see them as "traumatic." No matter, such readers are just the ones who have found ways to transform—what did Freud call it?—to sublimate such problems into constructive thinking about thinking. In any case, in one private manner or another, everyone somehow comes to deal with such problems, and I see only one practical way: we must each grow for ourselves some structure, of large complexity and little elegance, to tell us when —and when not—to trust each EFTA01157475 such pattern of inference. For example, we each learn never to repeat a near deduction more than a few times in any one argument. Furthermore, the accumulation of such experiences leads us eventually to realize that this is not peculiar to "near" alone: perhaps one shouldn't use any inference method too many times. What is "too many?" There is, I fear, no elegant answer. We each have to learn and master large of bodies knowledge about the limitations of each species of reasoning. It might be useful to try to catalog the kinds of cognitive incidents that must have so baffled each of us in early life. Each reader must recall the distress of being made to discern the arbitrary boundaries between one ocean and another, or at trying to answer "which came first, the chicken or the egg?" Every child must have wondered about his origin and from whence came the first person. Only the dullest child never found for himself some sort of Zeno paradox, or Achilles and Tortoise problem. I remember being especially disturbed to discover that there were questions that adults had no way to answer. MYSTICAL EXPERIENCE: Another family of disturbances arise when we question our own purposes. "Why should I do this,"—whatever it is—one asks, and proposes some answer. But then one is impelled to continue, "why should I want that," and so forth. There is a not uncommon phenomenon—sometimes called mystical experience—from which a person emerges with the conviction that some unsolvable problem (like the purpose of existence) has been completely explained; one can't remember quite how, only that it was answered so well as to leave no doubt at all. This, I venture, reflects some mental mechanism (perhaps one of last resort) that, in a state of particularly severe turmoil or stress, can short-circuit the entire intellectual process—by creating the illusion that the problem has been settled. Powerful but dangerous, such a mechanism short-cuts the canons of normal confirmation. One kind of confusion-cycle is thereby broken, but this may damage other ways in which sane minds confront beliefs with evidence. Then, anything can happen. {10} V. HUMOR AND EVOLUTION Ifyou wish to study a granfalloon just remove the skin ofa toy balloon.—Kurt Vonnegut, in Cat's Cradle In the 1912 edition Freud, still perplexed about the purpose of nonsense, recounts a joke of this form: {11 "A man at the dinner table dipped his hands in the mayonnaise and then ran them through his hair. When his neighbor looked astonished, the man apologized: "I'm so sorry. I thought it was spinach." We have argued that learning about bugs is central to the growth of reason. But reason itself grows in no vacuum; most of our ideas—and our ideas about ideas—come via our families and cultures, and this poses some special communication problems. For one, it is risky to point out the mistakes of a person one wants to please. So this must be done in some "conciliatory" manner—and humor seems involved in this. For another, if learning about bugs involves a special kind of memory, then this communication must somehow engage that memory. In this section we propose that humor—and more specifically, laughter—is innately enabled to do this, too. But first we digress to discuss an important methodological problem: Why is it so hard to explain why jokes are funny? Why, for that matter, is it so hard to say precisely what is a joke? We have already mentioned Wittgenstein's problem of defining "game": one can find no single quality common to all the different kinds of examples—and one finds a similar problem in attempting to define "humor." But we did not stop to ask why this is so. One might suppose it a mere surface difficulty, and hope that we may yet find a single underlying structure from which all funny things spring—some basic "grammar of humor," or "comical deep structure." Not so, I EFTA01157476 fear; when we look deeper for that underlying structure of humor we shall still find a vexing lack of unity. I argue that this is a consequence of the way things usually evolve in biology. In mechanisms designed by plan, it is reasonable to ask about purpose or cause. What is the purpose of that beam in this house? Simple: to hold up this roof—and, perhaps, to hold those walls apart. But, when we ask questions about structures created by evolution, we find that only rarely does one evolutionary increment serve a single purpose—and rarely is one alone in serving any particular purpose. Behavior emerges from a network of interdependent mechanisms, and one cannot expect any compactly circumscribed theory (or mechanism) completely to "explain" any single surface component of behavior. What a theory can do, though, is to describe some fragment of that larger network of interacting subsystems. Humor, like games, serves and exploits many different needs and mechanisms. It lacks sharp, natural boundaries because those underlying things themselves overlap and exploit one another. When we employ a word like "humor," one has the illusion of designating something sharper than this kind of complex web of relations among laughter, faulty reasoning, taboos and prohibitions, and unconscious suppressor mechanisms. But, I think the very clarity of words is itself a related illusion; as noted in {7}, language itself works only because oversimplification is more useful than realistic confusion—that is, in real life, if not for thinking about psychology. ROLES OF LAUGHTER: Consider what happens when a thought-situation comes to be perceived as funny or absurd: further reasoning is drowned in a flood of activity—furious motions of thorax, abdomen, head, limbs and face, accompanied by loud barking, wheezing, and choking noises. To a Martian, an epileptic seizure would be less alarming. Adults, of course, can train themselves to suppress this, but that is another matter. Laughter disrupts reasoning: The laughter reaction is so distracting as to keep the mind from proceeding further along the prohibited or ridiculous path it has started. Whatever that line of thought, the disruption prevents you from "taking it seriously," from acting upon it or considering its further logical consequences. At the same time, laughter exercises another, complementary function. Laughter focuses attention: While disrupting further reasoning, laughter takes a firm grip on the thought itself, holding up the absurdity in sharp focus. Perhaps the joke-thought is given full attention, holding the incongruity in some "short term memory"—so that "censor-learning" agency can absorb it. Thus "humor" might serve to mediate the process in which the censors learn, much as "pleasure" is often supposed to mediate ordinary learning. {12} EVOLUTION OF HUMOR: How might all this have evolved? We conjecture that it happened while our evolving minds passed through stages of increasing ability to reflect—to think not merely about the physical problems at hand, but about how we should apply our mental resources to them; in a word, when we were learning to plan. In order to make realistic plans, we had to learn to take account of what we could make our minds do. This ability could not have emerged all at once. There must have been intermediate steps—such as the appearance of multi-level schemes like the one suggested above, in which an A-mind is monitored by a B-mind. Eventually we became able to symbolize and manipulate representations of plans, and this allowed the first direct references to our own mental activities. Now, suddenly, we could do such tricks as to relate statements to their own quotations, and make propositions that (for better or for worse) could discuss their own truth—as in Quine's "yields truth" tour de force. {13} In any case, once able to accomplish intricate chains of reasoning, we became vulnerable to new kinds of bugs: faulty variable bindings, subtle changes of sense, and more obscurely circular EFTA01157477 logic. This same epoch probably saw also the emergence of Language for social communication, which also converged toward using more concise and manipulable symbolic representations. And, because we could not weaken the expressiveness of symbol-manipulation without losing its power, we had to evolve those censor memories. Of course, what actually happened was surely not this simple, but we had better return again to our speculations about laughter. Laughter's facial component suggests that it evolved in connection with social communication. It appears to be derived (ethologically) in part from a "conciliatory" expression, but it includes also a baring of teeth that suggests a defensive-aggressive mixture. (14} Laughter's bizarre vocal component also suggests social functions that combine ancestral "releasers" for both conciliation and aggression. Perhaps it came somehow to serve as a signal to induce another person to stop whatever he was doing: whether because dangerous, pointless, objectionable, ridiculous, or otherwise forbidden. Later, this function became internalized. If a person could feel and hear himself laugh, grimace, and shake, why not exploit these side-effects also to make one's own self to stop doing something ridiculous or prohibited? Perhaps, literally, men first teamed to laugh at their own mistakes, and later learned to censure themselves in silence. I make no plea for this particular scenario, only that something of this general sort must have happened, in which several pre-existing complexes grew together. They each brought along a variety of older interactions with other systems and purposes—that were exploited to produce the puzzling combinations of conciliation, aggression, sexuality, and nonsense that are now mixed together in humor. If other mental structures also share this kind of tangled ethological ancestry, then a mind grown this way must now resemble a great spider-web, in which many threads of different biological purpose intersect in many nodes of different multi-purpose mechanisms. If so, the goals of psychological theories must be to describe different fragments of that web—each to make a map of some few of those threads and nodes. By a curious coincidence, our theories of how minds work must probably have themselves this same peculiar, web-like character—albeit for a different reason. For (I think) the only way a person can understand anything very complicated is to understand it at each moment only locally —like the spider itself, seeing but a few threads and crossings from each viewpoint. Strand by strand, we build within our minds these webs of theory, from hard-earned locally intelligible fragments. The mind-spider's theory is correct to the extent that the model in his head corresponds to the mechanism in his head. (15} Finally it is probably futile to ask precisely what humor is. Korzybski's injunction applies here especially: the word "humor" points to no real thing. Instead, in each different person's mind it points to a slightly different web-model. We each use—and mean—the word a little differently —just as we each laugh at different jokes. Now I do not mean to hint that the problem is unreal, or even that it is especially incomprehensible. I only want to suggest that "humor" may be less a Thing-Part of the mind, and more a Thing-Theory in the mind. This makes it no less worthy of study, but one must be clear about what one is doing. If we get confused between making theories about theories and making theories about things, we may spin forever. (16} TIME CONSTANTS: According to our thesis, familiar types ofjokes should seem less funny, because the censors have learned more about them. Why, then, do some kinds ofjokes remain so persistently funny? People tire of old nonsense jokes, but not ofjokes about forbidden aspects of sex. Does this falsify our theory? Not necessarily; it may mean only that the censors for this particular subject are much slower to learn and to change. Is that plausible? Most psychological theories of our day—both popular and professional—seem to suppose that all memories are made of the same stuff, stored in the same huge container. (I argue otherwise in (Ref. MM-79).) But, on reflection, we see that some memories ought to be less changeable EFTA01157478 than others. Contemplate, for example, the plight of a mother with a new infant, that will demand her time and attention for several years. Why doesn't she wonder "Why am I doing this," or "what could this baby do for me to justify such a sacrifice." One might argue "To preserve the race," or "because you will love it," or "because it will repay you some day," but these would hardly convince any rational person; raising a child is not, let's face it, a notably sensible enterprise. Conventional wisdom recognizes that love is far from rational, and holds that an "instinctive" attachment is somehow created and somehow protected from casual alteration. Clearly so, but we must know more about those "somehow"s. This maternal self-questioning doesn't usually go too far, perhaps because we are protected by the web of personal pleasure and social compulsion surrounding child-rearing. But the problem is real, and there are occasional (and invariably concealed) tragedies in which a young mother's frustration overwhelms her attachment. The simplest way might be to build attachment into a special kind of memory that, once established, tends to persist for several years. After all, long time-constants characterize other aspects of attachment. Some persons always choose different partners of similar appearance; as though unable to alter a fixed stereotype. Others find themselves in the grip of undesired infatuations, that reason declares inappropriate. And most familiar is the seemingly inexorable time-span of mourning—the year or two it takes to adjust to separation or loss. All these could be by-products of adaptations of older mechanisms whose slowness was/is of value in our sociobiological evolution. {17} Perhaps one can also interpret in this light the prolonged, mourning-like depression associated with sexual assault, presuming that the momentary association of violence with sexuality somehow impairs the entire attachment machinery. The large time-constants make recovery slow, from a profound disturbance of normal social attachments. No matter if the victim manages to view the incident "rationally;" this does not automatically restore those sluggish mechanisms to their normal state, and one must suffer the prolonged functional deprivation of an important mind-part. All this suggests that the curious robustness of sexual humor may reflect only that the associated censors are among the "slow learners" of the mind, like retarded children. Perhaps they indeed are retarded children— the nearly static remnants of our own early selves. They change only slowly, and our tireless enjoyment of certain censured subjects may be a side-effect of that circumstance. So, we finally conclude: jokes are not really funny at all, but reflect the most serious of concerns; the pursuit of sobriety through the suppression of the absurd. Cambridge, Mass. May-July, 1980 NOTES [NOTE 0] Freud seemed somewhat puzzled by "nonsense jokes" and suggested, to explain the worst of them, that they "give the teller the pleasure of misleading and annoying, by rousing the expectation of a joke and then frustrating the listener"—who in turn—"damps down his annoyance by determining to tell them himself later on." The enjoyment of nonsense, he goes on, might also reflect a wish to return to a childhood of relaxed, careless th
ℹ️ Document Details
SHA-256
a04d91970b56fef4bb579bacd7f4e24503b4f9e86a67f06c04942b27ba82c2dc
Bates Number
EFTA01157468
Dataset
DataSet-9
Document Type
document
Pages
19

Comments 0

Loading comments…
Link copied!