COMP201 Java Programming - Department of Computer Science

Download Report

Transcript COMP201 Java Programming - Department of Computer Science

COMP 538
Reasoning and Decision under Uncertainty
Introduction
Readings: Pearl (1998, Chapter 1
Shafer and Pearl, Chapter 1
COMP 538 Introduction / Slide 2
Objectives


Course objectives
Course contents
COMP 538 Introduction / Slide 3
Uncertainty
 Uncertainty: the quality or state of being not clearly known.
 Uncertainty appears in many tasks




Partial knowledge of the state of the world
Noisy observations
Phenomena that is not covered by our models
Inherent randomness
COMP 538 Introduction / Slide 4
Probability and Decision Theory
 Well-known and well-understood framework for
uncertainty
 Clear semantics
 Provides principled answers for:
 Combining evidence
 Predictive & Diagnostic reasoning
 Incorporation of new evidence
 Intuitive (at some level) to human experts
 Can be learned
COMP 538 Introduction / Slide 5
Course Objectives

When applied to real-world problems, probability theory and
decision theory suffer from

Complexity of model construction
 Complexity of problem solving

This course covers methodologies developed recently in AI
community for dealing with those complexity problems.

The methodologies combine ideas from several disciplines
Artificial Intelligence, Machine Learning
Decision Theory, Theory of Computer Science
Statistics, Information Theory, Operations Research
COMP 538 Introduction / Slide 6
Complexity Problem of Applying Probability Theory

Example:


Patients in hospital are described by several attributes:

Background: age, gender, history of diseases, …

Symptoms: fever, blood pressure, headache, …

Diseases: pneumonia, heart attack, …
A joint probability distribution needs to assign a number to each
combination of values of these attributes, exponential model
size.


20 attributes require 2020 ( roughly 106 ) numbers
Real applications usually involve hundreds of attributes
COMP 538 Introduction / Slide 7
Complexity Problem of Applying Probability Theory

Because of the exponential model size problem, it was believed
that probability theory is not practical for dealing with uncertainty
in AI.

Alternative uncertainty calculi were introduced: uncertainty
factors, non-monotonic logic, fuzzy logic, etc.
COMP 538 Introduction / Slide 8
Complexity Problem of Applying Probability Theory



Bayesian networks alleviate the exponential model size problem.
Key idea: use conditional independence to factorize model into
smaller parts.
MINVOLSET
Example: Alarm network PULMEMBOLUS INTUBATION
KINKEDTUBE
VENTMACH
DISCONNECT
37 variables
PAP
SHUNT
VENTLUNG
VENITUBE
237
Model size:
Size of factored model: 509
PRESS
MINOVL
ANAPHYLAXIS
Model construction and
problem solving
becomes possible
SAO2
TPR
HYPOVOLEMIA
LVEDVOLUME
CVP
PCWP
LVFAILURE
STROEVOLUME
FIO2
VENTALV
PVSAT
ARTCO2
INSUFFANESTH
CATECHOL
HISTORY
ERRBLOWOUTPUT
CO
HR
HREKG
HRBP
BP
EXPCO2
ERRCAUTER
HRSAT
COMP 538 Introduction / Slide 9
Advantages of Bayesian Networks

Semantics



Model construction by expert


Probability theory provides the glue whereby the parts are
combined, ensuring that the system as a whole is consistent.
Alternative approaches suffer from several semantic deficiencies
(Pearl 1988, Chapter 1).
Appealing graphical interface allows experts to build model for
highly interacting variables.
Model construction from data

Probability foundation allows model construction from data by well
established statistical principles such as maximum likelihood
estimation and Bayesian estimation.
COMP 538 Introduction / Slide 10
Fielded Applications

Expert systems



Monitoring



Space shuttle engines (Vista project)
Freeway traffic
Sequence analysis and classification



Medical diagnosis
Fault diagnosis (jet-engines, Windows 98)
Speech recognition
Biological sequences
Information access
 Collaborative filtering
 Information retrieval
See tutorial by Breese and Koller (1997) and online resources for application
samples.
COMP 538 Introduction / Slide 11
Course Content/Bayesian Networks

Concept and semantics of Bayesian networks

Inference: How to answer queries efficiently

Learning: How to learn/adapt Bayesian network models from
data

Causal models: How to learn causality from statistical data
Detailed discussion of those topics can take one whole course
(DUKE). We will focus on the main ideas and skip the details so
that we can study other related topics.
COMP 538 Introduction / Slide 12
Bayesian networks and classical multivariate models

Special cases of Bayesian networks: many of the classical
multivariate models from



statistics, systems engineering, information theory, pattern
recognition and statistical mechanics
Examples: mixture models, factor analysis, hidden Markov
models, Kalman filters, Ising models.
Bayesian networks provide a way to view all of these models
as instances of a common underlying formalism.
COMP 538 Introduction / Slide 13
Course Content/Special models

Latent class analysis:



Statistical method for finding subtypes of related cases.
With proper generalization, might provide a statistical
foundation for Chinese medicine diagnosis.
Hidden Markov models:

A temporal model widely used in pattern recognition:
handwriting recognition, speech recognition.
COMP 538 Introduction / Slide 14
Decision Making under Uncertainty

Typical scenario: whether to take umbrella

Decision theory provides basis
for rational decision making:
Maximum expected utility principle.
Decision analysis: applying of decision theory.
Suffers from exponential model size.


COMP 538 Introduction / Slide 15
Course Content/Simple Decision Making

Influence diagrams



Generalization of Bayesian network.
Alleviate the complexity problem of decision analysis
Topics:


Evaluation: How to compute optimal decisions in an
influence diagram.
Value of information: Whether it is worthwhile to collect more
information to reduce uncertainty.
COMP 538 Introduction / Slide 16
Sequential Decision Making under Uncertainty

Agent/robot needs to execute multiple actions in
order to achieve a goal

Uncertainty originates from noisy sensor and
inaccurate actuators/uncontrollable enviroment
factors

What is the best way to reach goal with minimum
cost/time?
COMP 538 Introduction / Slide 17
Course Content/Sequential Decision Making

Models:

Markov decision processes, consider only uncertainty in
actuators/environment factors

Partially observable decision processes, consider uncertainty
in sensors and actuators/environment factors.

Solution methods:




Value iteration
Policy iteration
Dealing with model complexity using dynamic Bayesian networks
Learning sequential decision models

Model-based reinforcement learning
COMP 538 Introduction / Slide 18
Course Content/Summary

Bayesian networks

Concept and semantics, inference, learning, causilty

Special models: Hidden Markov models, latent class
analysis

Influence diagrams:


Evaluation, value of information
Markov decision processes and partially observable Markov
decision processes


Solution methods: value iteration, policy iteration, dynamic
Bayesian networks
Model-based reinforcement learning