* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download cl11_oct9
Survey
Document related concepts
Wizard of Oz experiment wikipedia , lookup
Alan Turing wikipedia , lookup
Human–computer interaction wikipedia , lookup
Intelligence explosion wikipedia , lookup
Kevin Warwick wikipedia , lookup
Chinese room wikipedia , lookup
Visual Turing Test wikipedia , lookup
Embodied cognitive science wikipedia , lookup
Existential risk from artificial general intelligence wikipedia , lookup
Turing test wikipedia , lookup
History of artificial intelligence wikipedia , lookup
Transcript
CMSC 671 Fall 2001 Class #11 – Tuesday, October 9 1 Today’s class • Philosophy of AI – Can we build intelligent machines? • If we do, how will we know they’re intelligent? – Should we build intelligent machines? • If we do, how should we treat them… • …and how will they treat us? 2 Philosophy of AI Alan M. Turing, “Computing Machinery and Intelligence” John R. Searle, “Minds, Brains, and Programs” J. Storrs Hall, “Ethics for Machines” (supplementary: Russell & Norvig Ch. 27) 3 Philosophical debates • What is AI, really? – What does an intelligent system look like? – Do we need, and can we have, emotions, consciousness, empathy, love? • Can we ever achieve AI, even in principle? • How will we know if we’ve done it? • If we can do it, should we? 4 Turing test • Basic test: – – – – Interrogator in one room, human in another, system in a third Interrogator asks questions; human and system answer Interrogator tries to guess which is which If the system wins, it’s passed the Turing Test • The system doesn’t have to tell the truth (obviously…) 5 Turing test objections • Objections are basically of two forms: – “No computer will ever be able to pass this test” – “Even if a computer passed this test, it wouldn’t be intelligent” 6 “Machines can’t think” • Theological objections • “It’s simply not possible, that’s all” • Arguments from incompleteness theorems – But people aren’t complete, are they? • Machines can’t be conscious or feel emotions – Reductionism doesn’t really answer the question: why can’t machines be conscious or feel emotions?? • Machines don’t have Human Quality X • Machines just do what we tell them to do – Maybe people just do what their neurons tell them to do… • Machines are digital; people are analog 7 “The Turing test isn’t meaningful”: Chinese Room argument 8 “The Turing test isn’t meaningful” • Maybe so, but… If we don’t use the Turing test, what measure should we use? • Very much an open question… 9 Ethical concerns: Robot behavior • • • How do we want our intelligent systems to behave? How can we ensure they do so? Asimov’s Three Laws of Robotics: 1. A robot may not injure a human being or, through inaction, allow a human being to come to harm. 2. A robot must obey orders given it by human beings except where such orders would conflict with the First Law. 3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law. 10 Ethical concerns: Human behavior • Is it morally justified to create intelligent systems with these constraints? – As a secondary question, would it be possible to do so? • Should intelligent systems have free will? Can we prevent them from having free will?? • Will intelligent systems have consciousness? (Strong AI) – If they do, will it drive them insane to be constrained by artificial ethics placed on them by humans? • If intelligent systems develop their own ethics and morality, will we like what they come up with? 11