Transcript cl11_oct9
CMSC 671
Fall 2001
Class #11 – Tuesday, October 9
1
Today’s class
• Philosophy of AI
– Can we build intelligent machines?
• If we do, how will we know they’re intelligent?
– Should we build intelligent machines?
• If we do, how should we treat them…
• …and how will they treat us?
2
Philosophy of AI
Alan M. Turing, “Computing Machinery and
Intelligence”
John R. Searle, “Minds, Brains, and Programs”
J. Storrs Hall, “Ethics for Machines”
(supplementary: Russell & Norvig Ch. 27)
3
Philosophical debates
• What is AI, really?
– What does an intelligent system look like?
– Do we need, and can we have, emotions, consciousness, empathy,
love?
• Can we ever achieve AI, even in principle?
• How will we know if we’ve done it?
• If we can do it, should we?
4
Turing test
• Basic test:
–
–
–
–
Interrogator in one room, human in another, system in a third
Interrogator asks questions; human and system answer
Interrogator tries to guess which is which
If the system wins, it’s passed the Turing Test
• The system doesn’t have to tell the truth (obviously…)
5
Turing test objections
• Objections are basically of two forms:
– “No computer will ever be able to pass this test”
– “Even if a computer passed this test, it wouldn’t be intelligent”
6
“Machines can’t think”
• Theological objections
• “It’s simply not possible, that’s all”
• Arguments from incompleteness theorems
– But people aren’t complete, are they?
• Machines can’t be conscious or feel emotions
– Reductionism doesn’t really answer the question: why can’t
machines be conscious or feel emotions??
• Machines don’t have Human Quality X
• Machines just do what we tell them to do
– Maybe people just do what their neurons tell them to do…
• Machines are digital; people are analog
7
“The Turing test isn’t meaningful”:
Chinese Room argument
8
“The Turing test isn’t meaningful”
• Maybe so, but…
If we don’t use the Turing test, what
measure should we use?
• Very much an open question…
9
Ethical concerns: Robot behavior
•
•
•
How do we want our intelligent systems to behave?
How can we ensure they do so?
Asimov’s Three Laws of Robotics:
1. A robot may not injure a human being or, through inaction, allow a
human being to come to harm.
2. A robot must obey orders given it by human beings except where
such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such protection
does not conflict with the First or Second Law.
10
Ethical concerns: Human behavior
• Is it morally justified to create intelligent systems with these
constraints?
– As a secondary question, would it be possible to do so?
• Should intelligent systems have free will? Can we prevent
them from having free will??
• Will intelligent systems have consciousness? (Strong AI)
– If they do, will it drive them insane to be constrained by artificial
ethics placed on them by humans?
• If intelligent systems develop their own ethics and morality,
will we like what they come up with?
11