Computational Discovery of Communicable Knowledge

Download Report

Transcript Computational Discovery of Communicable Knowledge

Intelligent Behavior in
Humans and Machines
Pat Langley
Computational Learning Laboratory
Center for the Study of Language and Information
Stanford University, Stanford, California USA
http://cll.stanford.edu/
Thanks to Herbert Simon, Allen Newell, John Anderson, David Nicholas, John Laird,
Randy Jones, and many others for discussions that led to the ideas presented in this talk.
Basic Claims
Early AI was closely linked to the study of human cognition.
This alliance produced many ideas that have been crucial to the
field’s long-term development.
Over the past 20 years, that connection has largely been broken,
which has hurt our ability to pursue two of AI's original goals:
 to understand the nature of the human mind
 to achieve artifacts that exhibit human-level intelligence
Re-establishing the connection to psychology would help achieve
these challenging objectives.
Outline of the Talk
 Review of early AI accomplishments that benefited from
connections to cognitive psychology
 Examples of AI's current disconnection from psychology
and some reasons behind this unfortunate development
 Ways that AI can benefit from renewed links to psychology
 Research on cognitive architectures as a promising avenue
 Steps we can take to encourage research along these lines
Early Links Between AI and Psychology
As AI emerged in the 1950s, one central insight was that
computers might reproduce the complex cognition of humans.
Some took human intelligence as an inspiration without trying
to model the details.
Others, like Herb Simon and Allen Newell, viewed themselves
as psychologists aiming to explain human thought.
This paradigm was pursued vigorously at Carnegie Tech, and it
was respected elsewhere.
The approach was well represented in the early edited
volume Computers and Thought.
Early Research on Knowledge Representation
Much initial work on representation dealt with the structure
and organization of human knowledge:
 Hovland and Hunt's (1960) CLS
 Feigenbaum's (1963) EPAM
 Quillian's (1968) semantic networks
 Schank and Abelson's (1977) scripts
 Newell's (1973) production systems
Not all research was motivated by concerns with
psychology, but it had a strong impact on the field.
Early Research on Problem Solving
Studies of human problem solving also influenced early AI
research:
 Newell, Shaw, and Simon’s (1958) Logic Theorist
 Newell, Shaw, and Simon’s (1961) General Problem Solver
 DeGroot’s (1965) discovery of progressive deepening
 VanLehn’s (1980) analysis of impasse-driven errors
Psychological studies led to key insights about both state-space
and goal-directed heuristic search.
Initial Paper on the Logic Theorist
Early Research on Knowledge-Based Reasoning
The 1980s saw many developments in knowledge-based
reasoning that incorporated ideas from psychology:
 expert systems (e.g., Waterman, 1986)
 qualitative physics (e.g., Kuipers, 1984; Forbus, 1984)
 model-based reasoning (e.g., Gentner & Stevens, 1983)
 analogical reasoning (e.g., Gentner & Forbus, 1991)
Research on natural language also borrowed many ideas from
studies of structural linguistics.
Early Research on Learning and Discovery
Many AI systems also served as models of human learning and
discovery processes:
 categorization (Hovland & Hunt, 1960; Feigenbaum, 1963;
Fisher, 1987)
 problem solving (Anzai & Simon, 1979; Anderson, 1981;
Minton et al., 1989; Jones & VanLehn, 1994)
 natural language (Reeker, 1976; Anderson, 1977; Berwick,
1979, Langley, 1983)
 scientific discovery (Lenat, 1977; Langley, 1979)
This work reflected the diverse forms of knowledge supported
by human learning and discovery.
The Unbalanced State of Modern AI
Unfortunately, AI has moved away from modeling human
cognition and become unfamiliar with results from psychology.
Despite the historical benefits, many AI researchers now believe
psychology has little to offer it.
Similarly, few psychologists believe that results from AI are
relevant to modeling human behavior.
This shift has taken place in a number of research areas, and it
has occurred for a number of reasons.
Current Emphases in AI Research
 Knowledge representation
 focus on restricted logics that guarantee efficient processing
 less flexibility and power than observed in human reasoning
 Problem solving and planning
 partial-order and, more recently, disjunctive planners
 bear little resemblance to problem solving in humans
 Natural language processing
 statistical methods with few links to psycho/linguistics
 focus on tasks like information retrieval and extraction
 Machine learning
 statistical techniques that learn far more slowly than humans
 almost exclusive focus on classification and reactive control
Technological Reasons for the Shift
One reason revolves around faster computer processors and
larger memories, which have made possible new methods for:
 playing games by carrying out far more search than humans
 finding complicated schedules that trade off many factors
 retrieving relevant items from large document repositories
 inducing complex predictive models from large data sets
These are genuine scientific advances, but AI might fare even
better by incorporating insights from human behavior.
Formalist Trends in Computer Science
Another factor involves AI’s typical home in departments of
computer science:
 which often grew out of mathematics departments
 where analytical tractability is a primary concern
 where guaranteed optimality trumps heuristic satisficing
 even when this restricts work to narrow problem classes
Many AI faculty in such organizations view connections to
psychology with intellectual suspicion.
Commercial Success of AI
Another factor has been AI’s commercial success, which has:
 led many academics to study narrowly defined tasks
 produced a bias toward near-term applications
 caused an explosion of work on “niche AI”
Moreover, component algorithms are much easier to evaluate
experimentally, especially given available repositories.
Such focused efforts are appropriate for corporate AI labs, but
academic researchers should aim for higher goals.
Benefits: Understanding Human Cognition
One reason for renewed interchange between the two fields is
to understand the nature of human cognition:
 because this would have important societal applications in
education, interface design, and other areas;
 because human intelligence comprises an important set of
phenomena that demand scientific explanation.
This remains an open and challenging problem, and AI systems
remain the best way to tackle it.
Benefits: Source of Challenging Tasks
Another reason is that observations of human abilities serve as
an important source of challenges, such as:
 understanding language at a deeper level than current systems
 interleaving planning with execution in pursuit of many goals
 learning complex knowledge structures from few experiences
 carrying out creative activities in art and science
Most work in AI sets its sights too low by focusing on tasks that
hardly involve intelligence.
Psychological studies reveal the impressive abilities of human
cognition and pose new problems for AI research.
Benefits: Constraints on Intelligent Artifacts
To develop intelligent systems, we must constrain their design,
and findings about human behavior can suggest:
 how the system can represent and organize knowledge;
 how the system can use that knowledge in performance;
 how the system can acquire knowledge from experience.
Some of the most interesting AI research uses psychological
ideas as design heuristics, including abilities we do not need
(e.g., to carry out rapid and extensive search).
Humans remain our only example of general intelligent systems,
and insights about their operation deserve serious consideration.
AI and Cognitive Systems
In 1973, Allen Newell argued “You can’t play twenty questions
with nature and win”. Instead, he proposed that we:
 move beyond isolated phenomena and capabilities to develop
complete models of intelligent behavior;
 develop cognitive systems that make strong theoretical claims
about the nature of the mind;
 view cognitive psychology and artificial intelligence as close
allies with distinct but related goals.
Newell claimed that a successful framework should provide a
unified theory of intelligent behavior.
He associated these aims with the idea of a cognitive architecture.
Assumptions of Cognitive Architectures
Most cognitive architectures incorporate a variety of assumptions
from psychological theories:
1. Short-term memories are distinct from long-term stores
2. Memories contain modular elements cast as symbolic structures
3. Long-term structures are accessed through pattern matching
4. Cognition occurs in retrieval/selection/action cycles
5. Performance and learning compose elements in memory
These claims are shared by a variety of architectures, including
ACT-R, Soar, Prodigy, and ICARUS.
Ideas about Representation
Cognitive psychology makes important representational claims:
 each element in a short-term memory is an active version of
some structure in long-term memory;
 many mental structures are relational in nature, in that they
describe connections or interactions among objects;
 concepts and skills encode different aspects of knowledge
that are stored as distinct cognitive structures;
 long-term memories have hierarchical organizations that
define complex structures in terms of simpler ones.
Many architectures adopt these assumptions about memory.
Architectural Commitment to Processes
In addition, a cognitive architecture makes commitments about:
 performance processes for:
 retrieval, matching, and selection
 inference and problem solving
 perception and motor control
 learning processes that:
 generate new long-term knowledge structures
 refine and modulate existing structures
In most cognitive architectures, performance and learning are
tightly intertwined, again reflecting influence from psychology.
Ideas about Performance
Cognitive psychology makes clear claims about performance:
 humans often resort to problem solving and search to solve
novel, unfamiliar problems;
 problem solving depends on mechanisms for retrieval and
matching, which occur rapidly and unconsciously;
 people use heuristics to find satisfactory solutions, rather
than algorithms to find optimal ones;
 problem solving in novices requires more cognitive resources
than experts’ use of automatized skills.
Many architectures embody these ideas about performance.
Ideas about Learning
Cognitive psychology has also developed ideas about learning:
 efforts to overcome impasses during problem solving can lead
to new cognitive structures;
 learning can transform backward-chaining heuristic search
into forward-chaining behavior;
 learning is incremental and interleaved with performance;
 structural learning involves monotonic addition of symbolic
elements to long-term memory;
 transfer to new tasks depends on the amount of structure
shared with previously mastered tasks.
Architectures often incorporate these ideas into their operation.
Architectures as Programming Languages
Cognitive architectures come with a programming language that:
 includes a syntax linked to its representational assumptions
 inputs long-term knowledge and initial short-term elements
 provides an interpreter that runs the specified program
 incorporates tracing facilities to inspect system behavior
Such programming languages ease construction and debugging
of knowledge-based systems.
Thus, ideas from psychology can support efficient development
of software for intelligent systems.
Responses: Broader AI Education
Most current AI courses ignore the field’s history; we need a
broader curriculum that covers its connections to:
 cognitive psychology
 structural linguistics
 logical reasoning
 philosophy of mind
These areas are more important to AI’s original agenda than are
ones from mainstream computer science.
For one example, see http://cll.stanford.edu/reason-learn/ , a
course I have offered for the past three years.
Responses: Funding Initiatives
We also need funding to support additional AI research that:
 makes contact with ideas from computational psychology
 addresses the same range of tasks that humans can handle
 develops integrated cognitive systems that move beyond
component algorithms
In recent years, DARPA and NSF have taken promising steps
in this direction, with clear effects on the community.
However, we need more funding programs along these lines.
Responses: Publication Venues
We also need places to present work in this paradigm, such as:
 AAAI’s new track for integrated intelligent systems
 this year’s Spring Symposium on AI meets Cognitive Science
 the special issue of AI Magazine on human-level intelligence
We need more outlets of this sort, but recent events have been
moving the field in the right direction.
Closing Remarks
In summary, AI’s original vision was to understand the basis of
intelligent behavior in humans and machines.
Many early systems doubled as models of human cognition,
while others made effective use of ideas from psychology.
Recent years have seen far less research in this tradition, with
AI becoming a set of narrow, specialized subfields.
Re-establishing contact with ideas from psychology, including
work on cognitive architectures, can remedy this situation.
The next 50 years must see AI return to its psychological roots
if it hopes to achieve human-level intelligence.
Closing Dedication
I would like to dedicate this talk to two of AI’s founding fathers:
Allen Newell (1927 – 1992)
Herbert Simon (1916 – 2001)
Both contributed to the field in many ways: posing new problems,
inventing methods, writing key papers, and training students.
They were both interdisciplinary researchers who contributed not
only to AI but to other disciplines, including psychology.
Allen Newell and Herb Simon were excellent role models who we
should all aim to emulate.