Marvin Lee Minsky (born August 9, 1927) is an American scientist in the field of artificial intelligence (AI), co-founder of MIT's AI laboratory, author of several texts on AI and philosophy, and winner of the 1969 Turing Award.
- In today's computer science curricula … almost all their time is devoted to formal classification of syntactic language types, defeatist unsolvability theories, folklore about systems programming, and generally trivial fragments of "optimization of logic design" — the latter often in situations where the art of heuristic programming has far outreached the special-case "theories" so grimly taught and tested — and invocations about programming style almost sure to be outmoded before the student graduates.
- Computer languages of the future will be more concerned with goals and less with procedures specified by the programmer.
- Turing Award Lecture "Form and Content in Computer Science" (1969), in Journal of the Association for Computing Machinery 17 (2) (April 1970)
- Once the computers got control, we might never get it back. We would survive at their sufferance. If we're lucky, they might decide to keep us as pets.
- Life Magazine (20 November 1970), p. 68
- Speed is what distinguishes intelligence. No bird discovers how to fly: evolution used a trillion bird-years to 'discover' that – where merely hundreds of person-years sufficed.
- "Communication with Alien Intelligence", in Extraterrestrials: Science and Alien Intelligence (1985) edited by Edward Regis also published in Byte Magazine (April 1985)
- Will robots inherit the earth? Yes, but they will be our children.
- When David Marr at MIT moved into computer vision, he generated a lot of excitement, but he hit up against the problem of knowledge representation; he had no good representations for knowledge in his vision systems.
- Marvin Minsky in: David G. Stork (1998). HAL's Legacy: 2001's Computer As Dream and Reality. p. 16
- An ethicist is someone who sees something wrong with whatever you have in mind.
- You don't understand anything until you learn it more than one way.
- In Managing an Information Security and Privacy Awareness and Training Program (2005) by Rebecca Herold, p. 101
Jokes and their Relation to the Cognitive Unconscious (1980)Edit
- [http://web.media.mit.edu/~minsky/papers/jokes.cognitive.txt "Jokes and their Relation to the Cognitive Unconscious" AI memo No. 603, (November 1980), also published in Cognitive Constraints on Communication (1981) edited by Vaina and Hintikka
- I am inclined to doubt that anything very resembling formal logic could be a good model for human reasoning. In particular, I doubt that any logic that prohibits self-reference can be adequate for psychology: no mind can have enough power — without the power to think about Thinking itself. Without Self-Reference it would seem immeasurably harder to achieve Self-Consciousness — which, so far as I can see, requires at least some capacity to reflect on what it does. If Russell shattered our hopes for making a completely reliable version of commonsense reasoning, still we can try to find the islands of "local consistency," in which naive reasoning remains correct.
- Since we have no systematic way to avoid all the inconsistencies of commonsense logic, each person must find his own way by building a private collection of "cognitive censors" to suppress the kinds of mistakes he has discovered in the past.
- Questioning one's own "top-level" goals always reveals the paradox-oscillation of ultimate purpose. How could one decide that a goal is worthwhile — unless one already knew what it is that is worthwhile?
- For avoiding nonsense in general, we might accumulate millions of censors. For all we know, this "negative meta-knowledge" — about patterns of thought and inference that have been found defective or harmful — may be a large portion of all we know.
- All intelligent persons also possess some larger-scale frame-systems whose members seemed at first impossibly different — like water with electricity, or poetry with music. Yet many such analogies — along with the knowledge of how to apply them — are among our most powerful tools of thought. They explain our ability sometimes to see one thing — or idea — as though it were another, and thus to apply knowledge and experience gathered in one domain to solve problems in another. It is thus that we transfer knowledge via the paradigms of Science. We learn to see gases and fluids as particles, particles as waves, and waves as envelopes of growing spheres.
- Positive general principles need always to be supplemented by negative, anecdotal censors. For, it hardly ever pays to alter a general mechanism to correct a particular bug.
K-Linesː A Theory of Memory (1980)Edit
- [http://csjarchive.cogsci.rpi.edu/1980v04/i02/p0117p0133/MAIN.PDF "K-Linesː A Theory of Memory" in Cognitive Science 4 (1980), pp.117-133
- When you "get an idea," or "solve a problem," or have a "memorable experience," you create what we shall call a K-line. This K-line gets connected to those "mental agencies" that were actively involved in the memorable event. When that K-line is later "activated," it reactivates some of those mental agencies, creating a "partial mental state" resembling the original.
- We usually say that one must first understand simpler things. But what if feelings and viewpoints are the simpler things?
- We shall envision the mind (or brain) as composed of many partially autonomous "agents"—as a "Society" of smaller minds. ...It is easiest to think about partial states that constrain only agents within a single Division. ...(we suggest) the local mechanisms for resolving conflicts could be the precursors of what we know later as reasoning — useful ways to combine different fragments of knowledge.
- Concrete concepts are not necessarily the simplest ones. A novice best remembers "being at" a concert. The amateur remembers more of what it "sounded like." Only the professional remembers the music itself, timbres, tones and textures.
- Old answers never perfectly suit new questions, except in the most formal, logical circumstances.
- Get the mind into the (partial) state that solved the old problem; then it might handle the new problem in the "same way."
- Changing the states of many agents grossly alters behavior, while changing only a few just perturbs the overall disposition a little.
- A memory should induce a state through which we see current reality as an instance of the remembered event — or equivalently, see the past as an instance of the present. ...the system can perform a computation analogous to one from the memorable past, but sensitive to present goals and circumstances.
- It would seem that making unusual connections is unusually difficult and, often, rather "indirect"—be it via words, images, or whatever. The bizarre structures used by mnemonist (and, presumably unknowingly, by each of us) suggests that arbitrary connections require devious pathways.
- Most theories of learning have been based on ideas of "reinforcement" of success. But all these theories postulate a single, centralized reward mechanism. I doubt this could suffice for human learning because the recognition of which events should be considered memorable cannot be a single, uniform process. It requires too much "intelligence." Instead I think such recognitions must be made, for each division of the mind, by some other agency that has engaged the present one for a purpose.
- Each subsociety of mind must have its own internal epistemology and phenomenology, with most details private, not only from the central processes, but from one another.
- Each part of the mind sees only a little of what happens in some others, and that little is swiftly refined, reformulated and "represented." We like to believe that these fragments have meanings in themselves — apart from the great webs of structure from which they emerge — and indeed this illusion is valuable to us qua thinkers — but not to us as psychologists — because it leads us to think that expressible knowledge is the first thing to study.
Music, Mind, and Meaning (1981)Edit
- "Music, Mind, and Meaning" (1981), a revised version of AI Memo No. 616, MIT; also published in the Computer Music Journal, Vol. 5, Number 3 (Fall 1981)
- Only the surface of reason is rational. I don't mean that understanding emotion is easy, only that understanding reason is probably harder.
- Our culture has a universal myth in which we see emotion as more complex and obscure than intellect. Indeed, emotion might be "deeper" in some sense of prior evolution, but this need not make it harder to understand; in fact, I think today we actually know much more about emotion than about reason.
- If explaining minds seems harder than explaining songs, we should remember that sometimes enlarging problems makes them simpler! The theory of the roots of equations seemed hard for centuries within its little world of real numbers, but it suddenly seemed simple once Gauss exposed the larger world of so-called complex numbers. Similarly, music should make more sense once seen through listeners' minds.
- We find things that do not fit into familiar frameworks hard to understand – such things seem meaningless.
- What is the difference between merely knowing (or remembering, or memorizing) and understanding? ...A thing or idea seems meaningful only when we have several different ways to represent it — different perspectives and different associations. ...Then we can turn it around in our minds, so to speak: however it seems at the moment, we can see it another way and we never come to a full stop. In other words, we can 'think' about it. If there were only one way to represent this thing or idea, we would not call this representation thinking.
- Of what use is musical knowledge? Here is one idea. Each child spends endless days in curious ways; we call this play. A child stacks and packs all kinds of blocks and boxes, lines them up, and knocks them down. … Clearly, the child is learning about space! ...how on earth does one learn about time? Can one time fit inside another? Can two of them go side by side? In music, we find out!
- The way the mathematics game is played, most variations lie outside the rules, while music can insist on perfect canon or tolerate a casual accompaniment.
- Most adults have some childlike fascination for making and arranging larger structures out of smaller ones.
- Perhaps the music that some call 'background' music can tranquilize by turning under-thoughts from bad to neutral, leaving the surface thoughts free of affect by diverting the unconscious.
- Theorems often tell us complex truths about the simple things, but only rarely tell us simple truths about the complex ones. To believe otherwise is wishful thinking or "mathematics envy."
- Music... immerses us in seemingly stable worlds! How can this be, when there is so little of it present at each moment?
- Hearing music is like viewing scenery and... when we hear good music our minds react in very much the same way they do when we see things.
- Our eyes are always flashing sudden flicks of different pictures to our brains, yet none of that saccadic action leads to any sense of change or motion in the world; each thing reposes calmly in its "place"! ...What makes us such innate Copernicans?
- How do both music and vision build things in our minds? Eye motions show us real objects; phrases show us musical objects. We "learn" a room with bodily motions; large musical sections show us musical "places." Walks and climbs move us from room to room; so do transitions between musical sections. Looking back in vision is like recapitulation in music; both give us time, at certain points, to reconfirm or change our conceptions of the whole.
- Innate sentic detectors could help by teaching children about their own affective states. For if distinct signals arouse specific states, the child can associate those signals with those states. Just knowing that such states exist, that is, having symbols for them, is half the battle.
- When no idea seems right, the right one must seem wrong.
The Society of Mind (1987)Edit
- We'll show you that you can build a mind from many little parts, each mindless by itself.
- This book... too, is a society — of many small ideas. Each by itself is only common sense, yet when we join enough of them we explain the strangest mysteries of mind.
- Unless we can explain the mind in terms of things that have no thoughts or feelings of their own, we'll only have gone around in a circle.
- How many processes are going on, to keep that teacup level in your grasp? There must be a hundred of them.
- The "laws of thought" depend not only on the property of brain cells, but also on how they are connected. And these connections are established not by the basic, "general" laws of physics... To be sure, "general" laws apply to everything. But, for that very reason, they can rarely explain anything in particular. ...Each higher level of description must add to our knowledge about lower levels.
- Questions about arts, traits, and styles of life are actually quite technical. They ask us to explain what happens among the agents of our minds. But this is a subject about which we have never learned very much... Such questions will be answered in time. But it will just prolong the wait if we keep using pseudo-explanation words like "holistic" and "gestalt." …It's harmful, when naming leads the mind to think that names alone bring meaning close.
- One's present personality cannot share all the thoughts of one's older personalities — and yet it has some sense that they exist. This is one reason why we feel that we possess an inner Self — a sort of ever-present person-friend, inside the mind, whom we can always ask for help.
- We rarely recognize how wonderful it is that a person can traverse an entire lifetime without making a single really serious mistake — like putting a fork in one's eye or using a window instead of a door.
- For generations, scientists and philosophers have tried to explain ordinary reasoning in terms of logical principles — with virtually no success. I suspect this enterprise failed because it was looking in the wrong direction: common sense works so well not because it is an approximation of logic; logic is only a small part of our great accumulation of different, useful ways to chain things together.
- p. 187
- What magical trick makes us intelligent? The trick is that there is no trick. The power of intelligence stems from our vast diversity, not from any single, perfect principle.
- p. 308
The Emotion Machine (2006)Edit
- Perhaps it is no accident that one meaning of the word express is "to squeeze"—for when you try to "express yourself," your language resources will have to pick and choose among the descriptions your other resources construct—and then attempt to squeeze a few of these through your tiny channels of phrases and gestures.
- I suspect our human "thinking processes" often "break down," but you rarely notice anything's wrong, because your systems so quickly switch you to think in different ways, while the systems that failed are repaired or replaced.
- Most of our future attempts to build large, growing Articial Intelligences will be subject to all sorts of mental disorders.
- We still remain prone to doctrines, philosophies, faiths, and beliefs that spread through the populations of entire civilizations. It is hard to imagine any foolproof ways to protect ourselves from such infections. ...the best we can do is to try to educate our children to learn more skills of critical thinking and methods of scientific verification.
- Every system that we build will surprise us with new kinds of flaws until those machines become clever enough to conceal their faults from us.
- I cannot articulate enough to express my dislike to people who think that understanding spoils your experience… How would they know?
- Mat Buckland, AI Techniques for Game Programming (2002), Cincinnati, OH: Premier Press, 36 (ISBN 1-931841-08-X).
Quotes about Marvin MinskyEdit
- Frank Rosenblatt... invented a very simple single-layer device called a Perceptron. ...Unfortunately, its influence was damped by Marvin Minsky and Seymour Papert, who proved [in Perceptrons: An Introduction to Computational Geometry (1969)] that the Perceptron architecture and learning rule could not execute the "exclusive OR" and therefore could not learn. This killed interest in Perceptrons for a number of years... It is possible to construct multilayer networks of simple units that could easily execute the exclusive OR... Minsky and Papert would have contributed more if they had produced a solution to this problem rather than beating the Perceptron to death.
- Francis Crick, The Astonishing Hypothesis: The Scientific Search for the Soul (1994)
- Although my own previous enthusiasm has been for syntactically rich languages like the Algol family, I now see clearly and concretely the force of Minsky's 1970 Turing lecture, in which he argued that Lisp's uniformity of structure and power of self reference gave the programmer capabilities whose content was well worth the sacrifice of visual form.