* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download History of AI
Chinese room wikipedia , lookup
Human–computer interaction wikipedia , lookup
Embodied cognitive science wikipedia , lookup
Existential risk from artificial general intelligence wikipedia , lookup
Computer Go wikipedia , lookup
Ethics of artificial intelligence wikipedia , lookup
Knowledge representation and reasoning wikipedia , lookup
History of AI Foundations from related fields Philosophy (400 B.C-) • Socrates->Plato->Aristotle – Socrates (socratic definitions): “I want to know what is characteristic of piety which makes all actions pious...that I may have it to turn to, and to use as a standard whereby to judge your actions and those of other men” (algorithm) – Aristotle: Try to formulate laws of rational part of the mind. Believed in another part, intuitive reason Dualism vs. Materialism • Dualism: The belief that mind, consciousness, cognition, and intelligence are separate and distinct from the material world and cannot be explained as purely physical processes. – Rene Descartes was a dualist who believed that the mind made contact with the physical body through the pineal gland at the back of the brain. • Materialism: The belief that mind, consciousness, cognition, and intelligence are physical processes that can be explained through normal scientific investigation of the material world. 3 Descartes • Strong AI: goal is to produce machines that understand in the sense that humans do • Weak AI: goal is only to produce machines that can act intelligently • Ray Mooney (a current AI researcher): – Strong AI seems to imply materialism • If purely physical machines can be intelligent, then mind is a physical phenomenon – Materialism seems to imply strong AI • If mind is a physical process and computers can emulate any physical process (strong Church-Turing thesis), then AI must be possible Philosophy: Source of knowledge • Empiricism (Francis Bacon 1561-1626) – John Locke (1632-1704): “Nothing is in the understanding which was not in the senses” – David Hume (1711-1776): Principle of induction: General rules from repeated associations between their elements – Bertrand Russell (1872-1970): Logical positivism: All knowledge can be characterized by logical theories connected, ultimately, to observed sentences that correspond to sensory inputs Mathematics • Logic – George Boole (1815-1864): formal language for making logical inference – Gottlob Frege (1848-1925): First-order logic (FOL) – Computability • David Hilbert (1862-1943): is there an algorithm for deciding the truth of any logical proposition involving the natural numbers? • Kurt Godel (1906-1978): No: undecidability. • Alan Turing (1912-1954): which functions are computable? – Church-Turing thesis: any computable function is computable via a Turing machine – No machine can tell in general whether a given program will return an answer on a given input, or run forever Mathematics… • Intractability – Polynomial vs. exponential (Cobham 1964; Edmonds 1965) – Reduction of one class of problems to another (Dantzig, 1960; Edmonds, 1962) – NP-completeness (Steven Cook 1971, Richard Karp 1972) – “Electronic Super-Brain” – in the 1950s, many felt computers had unlimited potential for intelligence; intractability results were sobering. • Interesting article: http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=00476631 Interesting article Mathematics… • Probability – Gerolamo Cardano (1501-1576): probability in gambling – Pierre Fermat (1601-1665), Blaise Pascal (1623-1662), James Bernoulli (1654-1705), Pierre Laplace (17491827): new methods – Thomas Bayes (1702-1761): updating rule • Decision theory = probability theory + utility theory – John Von Neumann & Oskar Morgenstern 1944 • Game theory Psychology (1879-) • Scientific methods for studying human vision – Hermann von Helmholtz (1821-1894), Wilhelm Wundt (1832-1920) • Introspective experimental psychology – Wundt – Results were biased to follow hypotheses • Behaviorism (prevailed 1920-1960) – – – – John Watson (1878-1958), Edward Lee Thorndyke (1874-1949) Against introspection Stimulus-response studies Rejected knowledge, beliefs, goals, reasoning steps Psychology • Cognitive psychology – Brain posesses and processes information – Kenneth Craik 1943: knowledge-based agent: • Stimulus translated to representation • Representation is manipulated to derive new representations • These are translated back into actions – Widely accepted now – Anderson 1980: “A cognitive theory should be like a computer program” Computer Engineering Roman Abacus Schikard’s Adding Machine, 1623 Blaze Pascal: Pascaline adding machine about 1650 Charles Babbage: Analytical and Difference engines 1840’s Computer engineering • Abacus (7000 years old) • Pascaline: mechanical adder & substractor (Pascal; mid 1600’s) – Leibniz added multiplication, 1694 • Analytic Engine: universal computation; never completed (ideas: addressable memory, stored programs, conditional jumps) – Charles Babbage (1792-1871), Ada Lovelace Computer engineering… [See Wired magazine late Fall 1999] • Heath Robinson: digital electronic computer for cracking German codes – Alan Turing 1940, England • Z-3: first programmable computer – Konrad Zuse 1941, Germany • ABC: first electronic computer – John Atanasoff 1940-42, US • ENIAC: first general-purpose, electronic, digital computer – John Mauchy & John Eckert (1946) • IBM 701, 1952 – the first computer to yield a profit Linguistics (1957-present) • Noam Chomsky (against B.F Skinner’s behaviorist approach to language learning): behaviorist theory does not address creativity in language. Chomsky’s theory was formal enough it could in principle be programmed. Chomsky hierarchy of grammars for formal languages very important in CS (regular grammars, context-free grammars, and so on). • Much of the early work in KR was tied to language and informed by research in linguistics History of AI Birth of AI (1943-56) • Warren McCulloch & Walter Pitts (1943): ANN with onoff neurons – Neurons triggered by sufficient #neighbors – Showed that any computable function computable with some network like this – Logical connectives implementable this way – Donald Hebb’s 1949 learning rule – Arguable forerunner of both logicist and connection traditions in AI • Turing & Shannon chess programs, 1950s (not implemented) • SNARC, first ANN computer, Minsky & Edmonds, 1951(3000 vacuum tubes and a surplus automatic pilot mechanism from a B-24 bomber) Birth of AI... • Dartmouth 1956 workshop for 2 months – Term “artificial intelligence” – Fathers of the field introduced (McCarthy, Minsky, Shannon, Samuel, Selfridge, Newell and Simon – “Carnegie Tech”) • Logic Theorist: program for proving theorems by Alan Newell & Herbert Simon Skakey: SRI 1966-72 Early enthusiasm (1952-69) • Claims: computers can do X • General Problem Solver, Newell & Simon – Intentionally solved puzzles in a similar way as humans do (order of subgoals, etc) • • • • • • Geometry Theorem Prover, Herbert Gelernter, 1959 Arthur Samuel’s learning checkers program 1952 LISP, time sharing, Advice taker: McCarthy 1958 Integration, IQ geometry problems, algebra stories Blocks world: vision, learning, NLP, planning Adalines [Widrow & Hoff 1960], perceptron convergence theorem [Rosenblatt 1962] Early AI Researchers were not shy about claims • It is not my aim to surprise or shock you—but the simplest way I can summarize is to say that there are now in the world machines that think, that learn and that create. Moreover, their ability to do these things is going to increase rapidly until—in a visible future—the range of problems they can handle will be coextensive with the range to which human mind has been applied. (Herb Simon, 1957) • In 1958, he predicted that within 10 years a computer would be chess champion, and an important new mathematical theorem would be proved by machine. Claims such as these turned out to be wildly optimistic. • Methods were demonstrated on 1 or 2 simple examples. Failed on others. A dose of reality (1966-74) • Simple syntactic manipulation did not scale – – ELIZA (example rule: if sentence contains “mother”, then say: “tell me more about your family”) • However, sometimes such bots (e.g. Julia) can fool humans • In 1966, a report by an advisory committee found that “there has been no machine translation of general scientific text, and none is in immediate prospect.” All U.S. government funding for academic translation projects was cancelled. “the spirit is willing but the flesh is weak” -> “the vodka is good but the meat is rotten” (simple syntactic transformations and word replacement) • • Intractability (can’t just add more machines) Machine evolution (genetic algorithms) did not scale • Perceptrons book with negative result on representation capability of 1-layer ANNs [Minsky & Papert] • Failure to come to grips with the “combinatorial explosion” was one of the main criticisms of AI contained in the Lighthill report (Lighthill, 1973), which formed the basis for the decision by the British government to end support for AI research in all but two universities. – – Funding dwindled Ironically, back-prop for multi-layer networks caused resurgence in the 80’s. It was actually discovered first in 1969 (Bryson and Ho) ELIZA [Weizenbaum 1966] Men are all alike. IN WHAT WAY They're always bugging us about something or other. CAN YOU THINK OF A SPECIFIC EXAMPLE Well, my boyfriend made me come here. YOUR BOYFRIEND MADE YOU COME HERE He says I'm depressed much of the time. I AM SORRY TO HEAR THAT YOU ARE DEPRESSED It's true that I am unhappy. DO YOU THINK COMING HERE WILL HELP YOU NOT BE UNHAPPY I need some help, that much seems certain. WHAT WOULD IT MEAN TO YOU IF YOU GOT SOME HELP Perhaps I could learn to get along with my mother. TELL ME MORE ABOUT YOUR FAMILY My mother takes care of me. WHO ELSE IN YOUR FAMILY TAKES CARE OF YOU ELIZA… HOW DO YOU DO. PLEASE STATE YOUR PROBLEM I'm wondering whether computers can think. WHY DO YOU MENTION COMPUTERS Since you are one I thought I would ask you - can you think? DID YOU THINK THEY MIGHT NOT BE ONE YOU THOUGHT YOU WOULD ASK I -CAN I THINK Let's talk just about you - not me. Can you think? YOU LIKE TO THINK I - NOT YOU - DON'T YOU Pardon? WHAT DOES THAT SUGGEST TO YOU? What does it suggest to you? TELL ME ABOUT WHAT DOES IT SUGGEST TO I That's nonsense - which of us do you mean by I? WHY DO YOU THINK I MEAN BY YOU The question meant to which of us is the word referring. SUPPOSE THE QUESTION MEANT TO WHICH OF US WERE NOT THE WORD REFERRING I think you are an extremely stupid program. PERHAPS YOU WOULD LIKE TO BE AN EXTREMELY STUPID PROGRAM Knowledge-based systems (1969-79) • DENDRAL: molecule structure identification [Buchanan et al.] – Knowledge intensive • Mycin: medical diagnosis [Feigenbaum, Buchanan, Shortliffe] – 450 rules; knowledge from experts; no domain theory – Better than junior doctors – Certainty factors • PROSPECTOR: drilling site choice [Duda et al] – found a large deposit • Domain knowledge in NLP (Winograd’s SHRDLU); Roger Schank and his students; Wood’s LUNAR system. • Knowledge representation: logic, frames... AI becomes an industry (1980-88) • R1: first successful commercial expert system, configured computer systems at DEC; saved 40M$/year • 1988: DEC had 40 expert systems, DuPont 100... • Nearly all major US corps had its own AI group working on expert systems • 1981: Japan’s 5th generation project • Software tools for expert systems: Carnegie Group, Inference, Intellicorp, Teknowledge • LISP-specific hardware: LISP Machines Inc, TI, Symbolics, Xerox • Industry: few M$ in 1980 -> 2B$ in 1988 Return of ANNs (1986-) • Mid-1980s, different research groups reinvented backpropagation (originally from 1969) • Disillusionment on expert systems • Fear of AI winter Recent events (1987-) • Rigorous theorems and experimental work rather than intuition • Real-world applications rather than toy domains • Building on existing work – E.g. speech recognition • Ad hoc, fragile methods in 1970s • Hidden Markov models now – E.g. planning (unified framework helped progress) Intelligent Agents (1995-present) • SOAR – complete agent architecture (Newell, Laird, Rosenbloom). • Intelligent agents on the Web. So common that “.bot” has entered everyday language • AI technol0gies underlie many internet tasks, such as search engines, recommender systems, site aggregators. • Can’t work on subfields of AI in isolation. E.g., reasoning and planning systems must handle uncertainty, since sensors are not perfect Availability of very large data sets (2001-present) • Data rather than which algorithm sometimes more important • Billions of words, pictures, base pairs of genomic sequences, … • Yarowsky (1995) showed that a simple bootstrapping approach over a very large corpus could be effective for WSD • Perhaps the knowledge bottleneck will be solved by learning methods over very large datasets rather than by hand-coding knowledge. Arguments against strong AI • • • • Theological objectives “It’s simply not possible for a machine” “machines cannot feel emotions” (Why?) Dreyfus (1972, 1986, 1992): background commonsense knowledge, the qualification problem, uncertainty, learning, compiled forms of decision making. Actually, his work was helpful. AI has made progress in all of these areas. Arguments against strong AI • Theological objectives • “It’s simply not possible for a machine” • Godel’s incompleteness theorem: vast literature. Responses: • R&N: applies onto to formal systems that are powerful enough to do arithmetic, such as Turing Machines. But TMs are infinite and computers are finite. So, any computer can be viewed as a large system in propositional logic, which is not subject to Godel’s IT. • R&N: Humans were behaving intelligently for 1000’s of years before they invtend mathematics, so it is unlikely that formal mathematic reasoning plays more than a peripheral role in what it means to be intelligent • R&N: Even if we grant that computers have limitations on what they can prove, we have no evidence that humans are immune from those limitations. “It’s impossible to prove that humans are not subject to Godel’s incompleteness theorem because any rigorous proof would require a formalization of the claimed unformalizable human talent, and hence refute itself. So, we are left with an appeal to intuition that humans can somehow perform superhuman feats of mathematical insight.” Arguments against strong AI • Machines just do what we tell them (Maybe people just do what their neurons tell them to do?) • Machines are digital; people are analog • The Chinese Room argument … John Searle’s Chinese Room R&N p. 1033 • “Searle appeals to intutition, not proof, for this part: just look at the room; what’s there to be a mind? But one could make the same argument about the brain: just look at this collection of cells (or of atoms), blindly operating according to the laws of biochemistry (or of physics) – what’s there to be a mind? Why can a hynk of brain be a mind while a hunk of liver cannot? That remains a great mystery”