Diapositiva 1 - uni
Download
Report
Transcript Diapositiva 1 - uni
Artificial Stupidity
The Myth of the Intelligent Agent
Richard Walker
Koeln, November 29, 2005
Artificial Stupidity (Examples)
An old paper (good version)
An old paper (bad version)
A nice drawing
A spread sheet
Koeln
November, 29, 2005
Artificial Stupidity
In each of these cases, the software performs a task
which could easily have been performed by a
human being.
This introduces Artificial Stupidity
Definition: ‘Artificial stupidity’ is the stupidity
produced by attempts to replace complex human
decision-making with so-called ‘intelligent’ software
Koeln
November, 29, 2005
The Argument /1
Software designers want to build “intelligent” systems in which
the computer takes the initiative on behalf of the user
(“intelligent agents”)
Intelligent Agents systematically fail – Artificial Stupidity
There exists a (very large) set of decision-making problems,,
where computers cannot, in principle, replace human beings
The limitations have nothing to do with technology
Even they were based on a perfect simulation of the brain,
“intelligent agents” would not be able to take decisions in the same
way as a human being
This depends on the “computer’s position in the world” – ecology
If we do build intelligent agents they will have an “alien
intelligence”
Koeln
November, 29, 2005
The Argument /2
BUT designers continue in the attempt to build
“intelligent software”
Many of these attempts are ergonomically
disastrous, particularly when they mimic human
intelligence
Intelligent agents are socially and culturally
dangerous
An alternative design strategy
Computers as a tool
Consequences for design
A note of caution…
Koeln
November, 29, 2005
Herbert Simon (1963)
“Machines will be capable, within twenty years,
of doing any work that a man can do”
The Shape of Automation for Men and Management
Koeln
November, 29, 2005
Martha Pollack (1991)
“We want to build intelligent actors, not just
intelligent thinkers. Indeed, it is not even clear
how one could assess intelligence in a
system that never acted -- or, put otherwise,
how a system could exhibit intelligence in the
absence of action”
‘Computers and Thought Lecture’, IJCAI-91
Koeln
November, 29, 2005
Intelligent Agents
“Our product contains an intelligent agent”
Semantic debate
Definition: An Intelligent Agent as a piece of
software that:
Acts/takes decisions (sends an email, makes a
recommendation, concludes a purchase)
Acts autonomously (does not merely respond to user input)
Acts in ways which benefit users
Is “intelligent”
Koeln
November, 29, 2005
Examples
Information management
E-Commerce
Microsoft paperclip
Chat bots
Security
Buying selling (e.g. on stock market)
Choosing hotels/restaurants
Chat-bots
Help systems/ technical support
Searching for, filtering, categorizing, distributing information
(personal news pages, anti-spam)
Detecting terrorists
Education
Intelligent tutors
Koeln
November, 29, 2005
Testing an Intelligent Agent
Turing Criteria (can it match a human being?)
Error rate at least equal to minimum acceptable human performance
Error mode comparable to human mode
Usefulness criteria (easier)
Net benefit must be greater than net benefit with a human agent
If the cost of a human agent infinite (task not possible) or very high, the
artificial agent must give tangible benefits to the user
Benefitsagent Costsagent Benefitshuman Costshuman
Benefitsagent Costsagent 0, Costshuman
Koeln
November, 29, 2005
How Well do they Work?
Many systems not testable because not
available
No systems for choosing flights/hotels/restaurants
No intelligent tutors
Koeln
November, 29, 2005
Detecting Terrorists
CAPPSII
Benefits
Cost savings (fewer trained staff)
Simple technological fix
Reduces staff responsibility
Costs
US system for vetting aircraft passenger list for potential terrorists
Secret algorithm
Data mining
Fewer trained staff
False security
Passengers not flown
Pilots lost jobs
Vulnerability
Balance: zero benefits – many costs
Koeln
November, 29, 2005
Chat Bots
Used in call centers, technical support etc.
Example
Human: I wanted to use you in a demonstration of what a chatbot
can do
ALICE: Interesting. I want to find out more about you.
Human: I'm a professor at Cologne University
ALICE: You are a professor at Cologne University? Do you get to
meet a lot of people?
Human: A fair number
ALICE: Are you telling the truth?
Human: I don't see why I should lie
ALICE: Look harder.
Koeln
November, 29, 2005
Costs and Benefits
Benefits
Cost savings (fewer trained staff)
Costs
Inadequate information to customers
Loss of customers
Simpler versions have interface to Call Center
Agent – Artificial Stupidity remains
Koeln
November, 29, 2005
Conclusions
In the real world we use very few intelligent
agents
Those we do use are not very good
The reasons have nothing to do with
technology
Koeln
November, 29, 2005
Software can be Autonomous
Washing machine
ABS
Autopilot system
Collision avoidance
Automatic defibrillators
Buying or selling on stock market
Not perfect – but meets usefulness criteria
Koeln
November, 29, 2005
Agents which Work
Limited number of input parameters
Context-independent
Path-independent
Given the input parameters, the procedure can
always be executed in the same way
Previous executions of procedure irrelevant
Algorithm simple (easy to verify)
Algorithm uses little or no background
information
Koeln
November, 29, 2005
The Washing Machine (Decision to Wash)
Decision to wash
Context-independent
Path-independent
Is there enough water?
Water temperature
Desired temperature
Simple algorithm
Doesn’t learn from previous attempts gone wrong
Limited number of Parameters
Procedure doesn’t need to take account of anything outside washing
machine
If( enough_water AND temp>=desired_temp) THEN wash
No background information
Does not know anything about what kind of clothes it is washing or how they
should be washed
Koeln
November, 29, 2005
More Complex Problems
Context dependency
Current state of user (mood, goals, desires, comfort, health etc.)
Current state of world (including other humans)
Path dependency
User memory (declarative, procedural, autobiographical)
Reflects past states of user and world
Potentially infinite number of parameters
Potentially any aspect of user or world, present or past may be
relevant to problem
Different parameters relevant in different contexts
Problem of how to select relevant parameters
Complex algorithm
Algorithm requires complex background information
Koeln
November, 29, 2005
Choice of Restaurant (Business Dinner)/1
Context dependency
My Goals
Requirements of guest
I am very hungry,
This is the third business dinner this week
Emotions
What does my boss want
How much can I put on the expense account
What are the ‘social rules’ for the situation)
Body state
What sort of dinner would please him/her
Cultural knowledge
Constraints
Show him how big and powerful we are
Show him we don’t waste money
Am I in an exploratory or a conservative mood?
Other parties: opinions of colleagues/family
…
Koeln
November, 29, 2005
Choice of Restaurant (Business Dinner)/2
Path dependency
Previous experience with customer
Previous experience in unknown countries
Previous experience with business dinners
Background knowledge
Experience with restaurant advertisements
Language knowledge
Knowledge of local cuisine
…
Koeln
November, 29, 2005
Chaitin/ Kolmogoroff Complexity
A problem characterized by length of shortest
program capable of coding solution
Agents which work have low C/K complexity
Agents which do not work have high C/K
complexity
Koeln
November, 29, 2005
Artificial Intelligence
Goal: to imitate human cognition (more recently: the
brain)
Technologies
Good Old Fashioned Artificial Intelligence (GOFAI)
Artificial Neural Networks (ANNs)
CYC
Supervised learning (back-prop)
Unsupervised learning (Kohonen)
Reinforcement Learning
“Conscious machines”
Evolutionary Computing
Koeln
November, 29, 2005
The Feasibility of an Artificial Brain
Brain is a physical-chemical system.
Nothing in principle prevents us from simulating it
Even if very large our computing power is catching up
Human genome <=Microsoft Office
Whole brain simulations 2015-20
Long term possibility of artificial cells
BUT
Even if we could do full-scale brain simulation in real-time
OR build a system which grows
Even if it could learn
Even if machines had ‘self-awareness’ , ‘emotions’, ‘feelings’
We would still not be able to build useful intelligent agents
Koeln
November, 29, 2005
Evolution
Many human capabilities are “built in” by evolution
Emotions (representation of body state)
Feelings (ability to represent and think about emotions)
Low level perception
Automatic fear snakes, spiders…
Automatic behavior
E.g. movement -->saliency
High level perception
Mood affects cognition
Baby cries for attention
This not very complex (genome c. 3 Gb)
Koeln
November, 29, 2005
Development
Genome codes for a system that develops
(incorporates information) during its interaction with
the environment
As cells duplicate different cells express different
genes
Cell duplication and gene expression
environmentally controlled
6*1013 cells in human body (6 thousand bullion)
Of which 1011 are neurons
Each cell has 25.000-30.000 genes which can be on or off
Expression (activation) of gene depends on internal and
external environment
Koeln
November, 29, 2005
Development /2
During development (interaction with environment)
huge amount of information incorporated in body
Brain (not just cognitive parts but also procedural memory
– motor procedures – social procedures)
Immune system
Morphology (skeleton, muscles, CV system)
Development itself is deeply context-dependent
People who develop in different societies/physical
environments/families develop in different ways
Koeln
November, 29, 2005
Human Decisions
All aspects of context present
Saliency
Determined by “inbuilt” and developed knowledge
Humans response
Automatic response
Brain-body system automatically produces action (balancing while
riding a bike)
“Moral”, “Ethical” decisions
Mediated response
Mental simulation of actions (includes motor and emotional areas)
Responses often wrong but much better than random response
Humans respond to any situation – even if they have had no
previous experience of it
In their responses humans use both “built in” and developed
capabilities
Koeln
November, 29, 2005
Non-Transparency of Decision-Making
Humans not fully aware of reasons for
decision-making
Emotional memory can work in absence of
awareness (Damasio experiments)
Main way of finding out through selfobservation and mental simulation
Understanding of reasons for
decisions/behavior extremely poor
Impossible to articulate in language
Koeln
November, 29, 2005
The Critical Obstacle
An agent with genuinely human intelligence would need the full
range of information incorporated in the genotype and the
phenotype
Self report does not work: humans do not understand reasons for
own decisions
And even if they did, the volume of information would be too
large for practical self-reporting
Observing actual behavior not enough – actual behavior is only
a small sample of potential behavior
No way of knowing reactions to rare (but critical) situations
The only way to incorporate the information normally used by a
human agent would be to:
Build a (growing) system which functions as same way as human baby
Bring it up as human being (in same culture as human beings)
Koeln
November, 29, 2005
The Impossibility of Intelligent agents
No way of transferring/
incorporating
information required for
human-like decisionmaking
Therefore impossible…
Failures have led to a
loss of enthusiasm in
academic community
The Rise and Fall of the Intelligent Agent
2000
1800
1600
Citations (Google Scholar)
1400
Papers (Google Scholar)
1200
Note: Survey on November 20,
2005.
Data for 2005 extrapolated from
partial annual figures
1000
800
600
400
200
0
1995
1996
1997
1998
1999
2000
Year
Koeln
November, 29, 2005
2001
2002
2003
2004
2005
John Maynard Keynes
“Practical men, who believe themselves to
be quite exempt from any intellectual
influence, are usually the slaves of some
defunct economist.“
The General Theory of Employment, Interest and Money (1935) Ch.
24 "Concluding Notes"
Koeln
November, 29, 2005
‘Alien intelligence’
Loss of academic enthusiasm does not imply loss of industrial
interest
Small scale agents still implemented on large scale (examples at
beginning of lecture)
To extent system has goals, values, emotions etc. they will not
be human goals
Example of video game
For foreseeable future will be much simpler than human
cognition
In most cases problem simply ignored
Systems with no knowledge, goals, emotions, bodily state,
context of user
Koeln
November, 29, 2005
Ergonomics
Designers and marketing managers still believe that
computers should be intelligent
Building intelligence into a computers a ‘good thing’
Positive marketing point (the “intelligent washing machine”)
Most attempts to make computers/machines
intelligent systematically makes them harder to use
(examples at beginning)
Koeln
November, 29, 2005
Broader Effects
Powerful economic logic in favor of intelligent agents
Much cheaper – less demanding than humans
Intelligent agents sold (and bought) as replacement for
humans
Autonomous help agents
‘Knowledge management’ for call centers
Chat bots for e-commerce (fortunately rare)
Destroys employment in jobs requiring human skills
Koeln
November, 29, 2005
Broader Effects /2
Forces us to interact with alien, low-grade intelligence
Extreme plasticity of the human brain
The “intentional stance”
Tamagochi
Kids and videogames
Adults and computer applications
At work
In services
Koeln
November, 29, 2005
Herbert Dreyfuss
“What we need to be afraid of is not computers
with superhuman intelligence, but humans
with subhuman intelligence”
What Computers can’t do, 1972
Koeln
November, 29, 2005
Design and Artificial Stupidity
An apparent contradiction
Systems should be simple
Goal of “intelligence” is to eliminate need for
unnecessary user actions/knowledge
E.g. make it possible to configure a network without
understanding protocols etc.
Eliminating “intelligent agents” seems to make this
impossible
Koeln
November, 29, 2005
Albert Einstein
“The supreme goal of all theory is to make the
irreducible basic elements as simple and as few
as possible without having to surrender the
adequate representation of a single datum of
experience.”
(actual quotation)
"Everything should be made as simple as
possible, but not simpler"
(usual paraphrase)
Koeln
November, 29, 2005
Design Guidelines
Examine each possibility for automation
Implement system context-dependent system only
when
Does there exist a context-free procedure which will give
guaranteed benefits
If so – use it
Can be shown to give benefits (user tests)
No human alternative (e.g. Google)
Where decision requires human intelligence, provide
information and options, not decision
Amazon
Koeln
November, 29, 2005
A Doubt
Not always clear whether there exists a context-free
solution to problem
Many problems though to require “intelligence” have
useful context-free solutions
Anti-SPAM
Anti-Virus
Amazon recommendations
Google
The final test can only be experimental
BUT experiments will often fail
Koeln
November, 29, 2005
Shakespeare
Glendower: I can call spirits from the vasty
deep
Hotspur: Why, so can I, or so can any man;
But will they come when you do call for
them?
W. Shakespeare, Henry IV Part I
Koeln
November, 29, 2005
Bibliography
Damasio, A. R. (1994). Descartes' Error. New York, G.P. Putnam's Sons.
Damasio, A. R. (1999). The feeling of what happens. New York, Harcourt Brace.
Dreyfuss, H. L. (1972). What Computers Can't Do, A Critique of Artificial Reason.
New York.
Dreyfuss, H. L. (1986). Mind over Machine. New York, NY, USA, The Free Press.
Jain, L. C., Z. Chen, et al., Eds. (2002). Intelligent Agents and Their Applications
(Studies in Fuzziness and Soft Computing, Vol. 98). Heidelberg, Germany, Physica
Verlag.
Pfeifer, R. and C. Scheier (1999). Understanding Intelligence. Cambridge, MA, MIT
Press.
Pollack, M. (1991). Computers and Thought Lecture. International Joint Conference
on Artificial Intelligence (IJCAI 91), Sydney, Australia, Morgan Kauffman.
Simon, H. A. (1965). The Shape of Automation for Men and Management. New York,
NY, USA, Harper and Row.
West-Eberhard, M. J. (2003). Developmental Plasticity and evolution. New York,
Oxford University Press.
Winograd, T. and F. Flores (1987). Understanding computers and cognition - a new
foundation for design. Norward, NY, USA, Ablex Publishing Corporation.
Koeln
November, 29, 2005