Computers

general-purpose device for performing arithmetic or logical operations
(Redirected from Compute)

A computer is a machine for manipulating data or storing data according to a list of instructions. Computers take many forms, from early room-sized complexes to modern personal computers (PCs) and personal digital assistants (PDAs) to tiny embedded systems that add sophisticated capabilities to other devices like toys and electronic appliances.

Trust The Computer. The Computer is your friend. ~ "Computer's Credo"

Quotes

edit
 
I have bought this wonderful machine- a computer. Now I am rather an authority on gods, so I identified the machine- it seems to me to be an Old Testament god with a lot of rules and no mercy. ~ Joseph Campbell
 
If you don't know anything about computers, just remember that they are machines that do exactly what you tell them but often surprise you in the result. ~ Richard Dawkins
 
Where a calculator like the ENIAC today is equipped with 18,000 vacuum tubes and weighs 30 tons, computers in the future may have only 1,000 vacuum tubes and perhaps weigh only 1½ tons. ~ Andrew Hamilton
 
What do such machines really do? They increase the number of things we can do without thinking. Things we do without thinking — there's the real danger. ~ Frank Herbert
 
The Joker: Do you know how many times we've come close to world war three over a flock of geese on a computer screen? ~ Alan Moore
 
Never underestimate the bandwidth of a station wagon full of tapes hurtling down the highway. ~ Andrew S. Tanenbaum
 
A computer would deserve to be called intelligent if it could deceive a human into believing that it was human. ~ Alan Turing
  • On two occasions, I have been asked [by members of Parliament], "Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?"...I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.
  • It makes sense to examine Plato and pottery together in order to understand the Greek world, Descartes and the mechanical clock together in order to understand Europe in the seventeenth and eighteenth centuries. In the same way, it makes sense to regard the computer as a technological paradigm for the science, the philosophy, even the art of the coming generation.
  • The clock has been the center of Western technology since its invention in the Middle Ages. Computer technology too finds it indispensable, although it has changed the clock from a mechanical device to a wholly electronic one.
  • In Hollywood, they think drawn animation doesn't work anymore, computers are the way. They forget that the reason computers are the way is that Pixar makes good movies. So everybody tries to copy Pixar. They're relying too much on the technology and not enough on the artists.
  • I have bought this wonderful machine- a computer. Now I am rather an authority on gods, so I identified the machine- it seems to me to be an Old Testament god with a lot of rules and no mercy.
  • Trust The Computer. The Computer is your friend.
  • To an outsider, the most significant innovation in the global warming controversy is the overt reliance that is being placed on models. Back in the days of nuclear winter, computer models were invoked to add weight to a conclusion: "These results are derived with the help of a computer model." But now, large-scale computer models are seen as generating data in themselves. No longer are models judged by how well they reproduce data from the real world—increasingly, models provide the data. As if they were themselves a reality. And indeed they are, when we are projecting forward.
  • This fascination with computer models is something I understand very well. Richard Feynman called it a disease. I fear he is right.
  • If you don't know anything about computers, just remember that they are machines that do exactly what you tell them but often surprise you in the result.
  • The simple fact is that without supporting directives or a mechanism for feedback, security is defined differently by each person and verified by no one. There is no metric for compliance with a "culture", and a "culture of security" is overridden by a culture of "get the job done" every time. If there are rules, write them down. If technology is put in place to implement or monitor the rules, write that down too. If people break the rules, follow up. If the rules prevent legitimate business from getting done, change them. It's that simple.
  • It always bothers me that, according to the laws as we understand them today, it takes a computing machine an infinite number of logical operations to figure out what goes on in no matter how tiny a region of space, and no matter how tiny a region of time. How can all that be going on in that tiny space? Why should it take an infinite amount of logic to figure out what one tiny piece of space/time is going to do? So I have often made the hypotheses that ultimately physics will not require a mathematical statement, that in the end the machinery will be revealed, and the laws will turn out to be simple, like the chequer board with all its apparent complexities.
  • Spock: Computers make excellent and efficient servants, but I have no wish to serve under them. Captain, a starship also runs on loyalty to one man, and nothing can replace it or him.
  • "So computers are tools of the devil?" thought Newt. He had no problem believing it. Computers had to be the tools of somebody, and all he knew for certain was that it definitely wasn't him.
  • Where a calculator like the ENIAC today is equipped with 18,000 vacuum tubes and weighs 30 tons, computers in the future may have only 1,000 vacuum tubes and perhaps weigh only 1½ tons.
    • Andrew Hamilton, "Brains that Click", Popular Mechanics 91 (3), March 1949, (pp. 162 et seq.) at p. 258.
  • What do such machines really do? They increase the number of things we can do without thinking. Things we do without thinking — there's the real danger.
  • Computers are good at following instructions, but not at reading your mind.
    • Donald Knuth (1984) cited in: Jorge Angeles (2011) Dynamic Response of Linear Mechanical Systems. p. 419
  • These machines have no common sense; they have not yet learned to "think," and they do exactly as they are told, no more and no less. This fact is the hardest concept to grasp when one first tries to use a computer.
    • Donald Knuth Knuth, Donald (1968). "Preface". The Art of Computer Programming, Volume 1: Fundamental Algorithms. Addison-Wesley. 
  • The Analytical Engine has no pretentions whatever to originate anything. It can do whatever we know how to order it to perform.
    • Ada Lovelace, An Account of the Analytical Engine, 1842, reprinted in Faster than Thought, ed. B.V. Bowden (Pitman, 1953), cited in Andrew Hodges Alan Turing: The Enigma of Intelligence (Unwin 1985) (probably the first known instance of "Computers can only do what they are programmed to do")
  • Dare to be gorgeous and unique. But don't ever be cryptic or otherwise unfathomable. Make it unforgettably great.
    • Robert J. Mical, Amiga Intuition Reference Manual, 1986, ISBN 0201110768, p. 231
  • Today's computers are not even close to a 4-year-old human in their ability to see, talk, move, or use common sense. One reason, of course, is sheer computing power. It has been estimated that the information processing capacity of even the most powerful supercomputer is equal to the nervous system of a snail—a tiny fraction of the power available to the supercomputer inside [our] skull.
    • Steven Pinker, How Unique You Are!; Is There a Creator Who Cares About You? (1998), published by Jehovah's Witnesses.
  • But if these machines were ingenious, what shall we think of the calculating machine of Mr. Babbage? What shall we think of an engine of wood and metal which can not only compute astronomical and navigation tables to any given extent, but render the exactitude of its operations mathematically certain through its power of correcting its possible errors? What shall we think of a machine which can not only accomplish all this, but actually print off its elaborate results, when obtained, without the slightest intervention of the intellect of man?
    • Edgar Allan Poe, "Maelzel's Chess-Player" (Southern Literary Messenger, April 1836)
  • Never underestimate the bandwidth of a station wagon full of tapes hurtling down the highway.
  • An adversary capable of implanting the right virus or accessing the right terminal can cause massive damage.
    • George Tenet, director of the U.S. Central Intelligence Agency. Cited in Awake! magazine 2001, 5/22.
  • A computer would deserve to be called intelligent if it could deceive a human into believing that it was human.
  • Computers with their binary on–off logic seem to appeal to the military mind. This is because the military, in order to counter the inherent confusion and danger of war, is forever seeking ways to make communications as terse and unambiguous as humanly possible. Computers by their very nature do just that. Had they only been able to stand at attention and salute, in many ways they would have made ideal soldiers.
  • With computers acting as the stimulus, the theory of war was assimilated into that of microeconomics. . . . Instead of evaluating military operations by their product –that is, victory – calculations were cast in terms of input–output and cost effectiveness. Since intuition was replaced by calculation, and since the latter wasto be carried out with the aid of computers, it was necessary that all the phenomena of war be reduced to quantitative form. Consequently everything that could be quantified was, while everything that could not be tended to be thrown onto the garbage heap.
  • It used to be said of a man who had suffered a catastrophic setback in his line of work that he had been handed his head on a platter. We are being handed our heads with tweezers now.
  • The danger of computers becoming like humans is not as great as the danger of humans becoming like computers.
    • Konrad Zuse (2005) in: Hersfelder Zeitung. Nr. 212, 12. September 2005.

Software

edit
Main article: Software
  • Anyone who slaps a 'this page is best viewed with Browser X' label on a Web page appears to be yearning for the bad old days, before the Web, when you had very little chance of reading a document written on another computer, another word processor, or another network.
  • Interviewer: Is studying computer science the best way to prepare to be a programmer? Bill Gates: No. the best way to prepare is to write programs, and to study great programs that other people have written. In my case, I went to the garbage cans at the Computer Science Center and I fished out listings of their operating system. You got to be willing to read other people's code, then write your own, then have other people review your code. You've got to want to be in this incredible feedback loop where you get the world-class people to tell you what you're doing wrong.
    • Bill Gates cited in: "Programmers at Work: Interviews With 19 Programmers Who Shaped the Computer Industry", Tempus, by Susan Lammers (Editor)
  • Around computers it is difficult to find the correct unit of time to measure progress. Some cathedrals took a century to complete. Can you imagine the grandeur and scope of a program that would take as long?
    • SIGPLAN, Association for Computing Machinery (1992) "Epigrams in Programming", September 1982
  • A refund for defective software might be nice, except it would bankrupt the entire software industry in the first year.
Main article: Video games
 
I was fascinated by the power of the computer to not only calculate, but also to interact with written language. I had been thinking about writing a program to interact with a human through language, but the content of such a program remained a mystery to me. ~ Bill Heineman
  • The computer was a fairly new thing in 1971, and most people were not well acquainted with it. There was only one five-week course in programming at Carleton. I had taken the class when I was a junior, and the next trimester, I was recruited to be a lab assistant to help other students who were taking the course. I was fascinated by the power of the computer to not only calculate, but also to interact with written language. I had been thinking about writing a program to interact with a human through language, but the content of such a program remained a mystery to me.
  • The only legitimate use of a computer is to play games.
    • Eugene Jarvis, Supercade, MIT Press, p.14 ISBN: 0-262-02492-6
Main article: Programming


Misattributed

edit
  • I think there is a world market for maybe five computers.
    • Commonly attributed to Thomas J. Watson (1874–1956), general manager and chairman of IBM, but there is no evidence he ever said or wrote this. The earliest known citations occurred in the 1980s.[2]

See also

edit
edit
 
Wikipedia
Wikipedia has an article about: