CIS730-Lecture-40

Download Report

Transcript CIS730-Lecture-40

Lecture 40 of 42
Final Review Part 1 of 2
Monday, 05 December 2005
William H. Hsu
Department of Computing and Information Sciences, KSU
http://www.kddresearch.org
http://www.cis.ksu.edu/~bhsu
Reading:
None
Final Review: Chapters 1-15, 18-19, 23, 24 R&N
(emphasis on 14-15, 18-19)
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 1:
The Intelligent Agent Framework
•
Artificial Intelligence (AI)
– Operational definition: study / development of systems capable of “thought
processes” (reasoning, learning, problem solving)
– Constructive definition: expressed in artifacts (design and implementation)
•
•
Intelligent Agents
Topics and Methodologies
– Knowledge representation
• Logical
• Uncertain (probabilistic)
• Other (rule-based, fuzzy, neural, genetic)
– Search
– Machine learning
– Planning
•
Applications
–
–
–
–
Problem solving, optimization, scheduling, design
Decision support, data mining
Natural language processing, conversational and information retrieval agents
Pattern recognition and robot vision
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 2:
Agents and Problem Solving
•
Agent Frameworks
– Reactivity vs. state
– From goals to preferences (utilities)
•
Applications and Automation Case Studies
–
–
–
–
•
Search: game-playing systems, problem solvers
Planning, design, scheduling systems
Control and optimization systems
Machine learning: pattern recognition, data mining (business decision support)
Things to Check Out Online
– Resources page:
www.kddresearch.org/Courses/Fall-2001/CIS730/Resources
– Yahoo! Group discussions: groups.yahoo.com/group/ksu-cis730-fall2001
– Suggested project topics, resources – posted in YG
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 3:
Search and Constraints
•
Today’s Reading: Sections 3.5-3.8, Russell and Norvig
•
Solving Problems by Searching
– Problem solving agents: design, specification, implementation
– Specification components
• Problems – formulating well-defined ones
• Solutions – requirements, constraints
– Measuring performance
•
Formulating Problems as (State Space) Search
•
Example Search Problems
– Toy problems: 8-puzzle, 8-queens, cryptarithmetic, toy robot worlds, constraints
– Real-world problems: layout, scheduling
•
Data Structures Used in Search
•
Uninformed Search Algorithms: BFS, DFS, Branch-and-Bound
•
Next Class: Informed Search Strategies
– State space search handout (Winston)
– Search handouts (Ginsberg, Rich and Knight)
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 4:
Uninformed Search Algorithms
•
Search
– Problem formulation: state space (initial / operator / goal test / cost), graph
– State space search approaches
• Blind (uninformed)
• Heuristic (informed)
•
Applications
– Problem solving
• Optimization
• Scheduling
• Design
– Machine learning (hypothesis space search)
•
More Resources Online
– http://www-jcsu.jesus.cam.ac.uk/~tdk22/project
– See also http://groups.yahoo.com/group/ksu-cis730-fall2001 (“REFERENCES”)
•
Course Project Guidelines Posted in YG
– Part I: format
– Part II: writing quality and criteria
– Part III: resources and suggested topics
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 5:
Heuristic Search Algorithms – Greedy, A*
•
More Heuristic Search
– Best-First Search
• Greedy
• A/A*
– Search as function maximization
• Problems: ridge; foothill; plateau, jump discontinuity
• Solutions: macro operators; global optimization
•
Constraint Satisfaction Search
•
Next Class: IDA*, Hill-Climbing, Iterative Improvement
– Gradient descent
– Global search
• MCMC: intuition
• Some examples of state-of-the-art applications
• Properties and tradeoffs
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 6:
More Heuristic Search – A*, Hill-Climbing / SA
•
More Heuristic Search
– Best-First Search: A/A* concluded
– Iterative improvement
• Hill-climbing
• Simulated annealing (SA)
– Search as function maximization
• Problems: ridge; foothill; plateau, jump discontinuity
• Solutions: macro operators; global optimization (genetic algorithms / SA)
•
Next Class: Constraint Satisfaction Search, Heuristic Search
•
Next Week: Adversarial Search (e.g., Game Tree Search)
– Competitive problems
– Minimax algorithm
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 7:
Constraint Satisfaction Problems
•
Constraint Satisfaction Problems (CSPs)
– Problem definition
• Domain
• Constraints
– Examples: N-queens, cryptarithmetic, etc.
•
Issues to be Covered Later
– Knowledge representation: how to express domain, constraints
– Relational constraints
• In classical logic (propositional, predicate, first-order)
• In uncertain reasoning
•
Solving CSPs
– Propositional constraints: satisfiability solver
– First-order relational constraints: difficulties – later
– Speeding up CSPs: iterative improvement
• Gradient (hill-climbing) optimization
• Simulated annealing
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 8:
Game Tree Search: Minimax
•
Game Graph Search
– Frameworks
• Two-player versus multi-player
• Zero-sum versus cooperative
• Perfect information versus partially-observable (hidden state)
– Concepts
• Utility and representations (e.g., static evaluation function)
• Reinforcements: possible role for machine learning
• Game tree: node/move correspondence, search ply
•
Family of Algorithms for Game Trees: Minimax
– Propagation of credit
– Imperfect decisions
– Issues
• Quiescence
• Horizon effect
– Need for (alpha-beta) pruning
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 9:
More Game Tree Search: -, Expectiminimax
•
Games as Search Problems
– Frameworks
– Concepts: utility, reinforcements, game trees
– Static evaluation under resource limitations
•
Family of Algorithms for Game Trees: Minimax
– Static evaluation algorithm
• To arbitrary ply
• To fixed ply
• Sophistications: iterative deepening, pruning
– Credit propagation
• Intuitive concept
• Basis for simple (delta-rule) learning algorithms
•
State of The Field
•
Uncertainty in Games: Expectiminimax and Other Algorithms
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 10:
Logical Agents and Knowledge Representations
•
Logical Agents
– Knowledge Bases (KB)
– Logic in general
• Representation languages, syntax
• Inference systems
– Calculi
• Propositional
• First-order (FOL, FOPC)
•
Possible Worlds
– Entailment
– Models
•
IA Toy Worlds
– Wumpus world
– Blocks world
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 11:
Propositional and Predicate Logic
•
Logical Frameworks
– Knowledge Bases (KB)
– Logic in general: representation languages, syntax, semantics
– Propositional logic
– First-order logic (FOL, FOPC)
– Model theory, domain theory: possible worlds semantics, entailment
•
Normal Forms
– Conjunctive Normal Form (CNF)
– Disjunctive Normal Form (DNF)
– Horn Form
•
Proof Theory and Inference Systems
– Sequent calculi: rules of proof theory
– Derivability or provability
– Properties
• Soundness (derivability implies entailment)
• Completeness (entailment implies derivability)
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 12:
Foundations of First-Order Logic
•
FOL in Practice
– FOL agents
– Example: Wumpus World in FOL
– Situation calculus
– Frame problem and variants (see R&N sidebar)
• Representational vs. inferential frame problems
• Qualification problem: “what if?”
• Ramification problem: “what else?” (side effects)
– Successor-state axioms
•
Logical Languages
– Propositional logic
– Predicates, terms, functions, atoms (atomic sentences / atomic WFFs), WFFs
– First-order logic (FOL, FOPC): universal and existential quantification
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 13:
First-Order Knowledge Bases
•
Properties of Knowledge Bases (KBs)
– Satisfiability and validity
– Entailment and provability
•
Properties of Proof Systems: Soundness and Completeness
•
Normal Forms: CNF, DNF, Horn; Clauses vs. Terms
•
Frame, Ramification, Qualification Problems
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 14:
Resolution Theorem Proving
•
Resolution Theorem Proving
– Conjunctive Normal Form (clausal form)
– Inference rule
• Single-resolvent form
• General form
– Proof procedure: refutation
– Decidability properties
• FOL-SAT
• FOL-NOT-SAT (language of unsatisfiable sentences; complement of FOL-SAT)
• FOL-VALID
• FOL-NOT-VALID
•
Next Class
– More Prolog
– Implementing unification
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 15:
Logic Programming Techniques
•
Properties of Proof Systems (Again)
– Soundness and completeness
– Decidability, semi-decidability, undecidability
•
Resolution
•
Refutation
•
Satisfiability, Validity
•
Unification
– Occurs check
– Most General Unifier
•
Prolog: Tricks of The Trade
– Demodulation, paramodulation
– Unit resolution, set of support, input / linear resolution, subsumption
– Indexing (table-based, tree-based)
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 16:
Classical Planning
•
Classical Planning
– Planning versus search
– Problematic approaches to planning
• Forward chaining
• Situation calculus
– Representation
• Initial state
• Goal state / test
• Operators
•
Efficient Representations
– STRIPS axioms
• Components: preconditions, postconditions (ADD, DELETE lists)
• Clobbering / threatening
– Reactive plans and policies
– Markov decision processes
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 17:
Partial-Order Planning
•
Classical Planning Framework
– Planning versus search
– Representation: initial state, goal state / test, operators
•
STRIPS Operators
– Components: preconditions, postconditions (ADD, DELETE lists)
– STRIPS and interference
• Clobbering / threatening
• Promotion / demotion
– Partial-Order Planners (POP systems)
•
Next Week
– Hierarchical abstraction planning: ABSTRIPS
– Conditional plans
– Reactive plans and policies
– Markov decision processes
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 18:
STRIPS and ABSTRIPS
•
Classical Planning Framework
– Planning versus search
– Representation: initial state, goal state / test, operators
•
STRIPS Operators
– Components: preconditions, postconditions (ADD, DELETE lists)
– STRIPS and interference
• Clobbering / threatening
• Promotion / demotion
– Partial-Order Planners (POP systems)
•
Next Week
– Hierarchical abstraction planning: ABSTRIPS
– Conditional plans
– Reactive plans and policies
– Markov decision processes
Adapted from slides by S. Russell, UC Berkeley
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 19:
Reaction and Replanning
•
Classical Planning Framework
– Planning versus search
– Representation: initial state, goal state / test, operators
– STRIPS operators
– Partial versus total-order: property of plans
– Interleaved vs. noninterleaved: property of planners
•
Last Week
– Hierarchical abstraction planning: ABSTRIPS
– Conditional plans
•
This Week
– Monitoring and replanning
– Reactive plans and policies
•
Later
– Decision theory
– Markov decision processes
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 20:
Reasoning under Uncertainty
•
Introduction to Probabilistic Reasoning
– Framework: using probabilistic criteria to search H
– Probability foundations
• Definitions: subjectivist, objectivist; Bayesian, frequentist, logicist
• Kolmogorov axioms
•
Bayes’s Theorem
– Definition of conditional (posterior) probability
– Product rule
•
Maximum A Posteriori (MAP) and Maximum Likelihood (ML) Hypotheses
– Bayes’s Rule and MAP
– Uniform priors: allow use of MLE to generate MAP hypotheses
– Relation to version spaces, candidate elimination
•
Next Week: Chapter 15, Russell and Norvig
– Later: Bayesian learning: MDL, BOC, Gibbs, Simple (Naïve) Bayes
– Categorizing text and documents, other applications
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 21:
Introduction to Bayesian Networks
•
Graphical Models of Probability
– Bayesian belief networks (BBNs) aka belief networks aka causal networks
– Conditional independence, causal Markovity
– Inference and learning using Bayesian networks
• Representation of distributions: conditional probability tables (CPTs)
• Learning polytrees (singly-connected BBNs) and tree-structured BBNs (trees)
•
BBN Inference
– Type of probabilistic reasoning
– Finds answer to query about P(x) - aka QA
•
Learning in BBNs: In Two Weeks
– Known structure
– Partial observability
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 22:
Introduction to Machine Learning
•
Taxonomies of Learning
•
Definition of Learning: Task, Performance Measure, Experience
•
Concept Learning as Search through H
– Hypothesis space H as a state space
– Learning: finding the correct hypothesis
•
General-to-Specific Ordering over H
– Partially-ordered set: Less-Specific-Than (More-General-Than) relation
– Upper and lower bounds in H
•
Version Space Candidate Elimination Algorithm
– S and G boundaries characterize learner’s uncertainty
– Version space can be used to make predictions over unseen cases
•
Learner Can Generate Useful Queries
•
Next Tuesday: When and Why Are Inductive Leaps Possible?
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 23:
Decision Trees
•
•
(Inductive) Bias: Preference for Some h  H (Not Consistency with D Only)
Decision Trees (DTs)
– Boolean DTs: target concept is binary-valued (i.e., Boolean-valued)
– Building DTs
• Histogramming: a method of vector quantization (encoding input using bins)
• Discretization: continuous input  discrete (e.g.., by histogramming)
•
Entropy and Information Gain
– Entropy H(D) for a data set D relative to an implicit concept c
– Information gain Gain (D, A) for a data set partitioned by attribute A
– Impurity, uncertainty, irregularity, surprise
•
Heuristic Search
– Algorithm Build-DT: greedy search (hill-climbing without backtracking)
– ID3 as Build-DT using the heuristic Gain(•)
– Heuristic : Search :: Inductive Bias : Inductive Generalization
•
MLC++ (Machine Learning Library in C++)
– Data mining libraries (e.g., MLC++) and packages (e.g., MineSet)
– Irvine Database: the Machine Learning Database Repository at UCI
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 24:
Perceptrons and Artificial Neural Networks
•
Neural Networks (NNs): Parallel, Distributed Processing Systems
– Biological NNs and artificial NNs (ANNs)
– Perceptron aka Linear Threshold Gate (LTG), Linear Threshold Unit (LTU)
• Model neuron
• Combination and activation (transfer, squashing) functions
•
Multi-Layer ANNs
– Focused on one species: (feedforward) multi-layer perceptrons (MLPs)
– Input layer: an implicit layer containing xi
– Hidden layer: a layer containing input-to-hidden unit weights and producing hj
– Output layer: a layer containing hidden-to-output unit weights and producing ok
– n-layer ANN: an ANN containing n - 1 hidden layers
– Epoch: one training iteration
•
Overfitting
– Overfitting: h does better than h’ on training data and worse on test data
– Prevention, avoidance, and recovery techniques
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences
Lecture 25:
Introduction to Bayesian Learning
•
Minimum Description Length (MDL)
– Bayesian Information Criterion (BIC)
– BIC = additive inverse of MDL (i.e., BIC(h) = -MDL(h))
•
Bayesian Classification: Finding Most Probable v Given Examples x
•
Bayes Optimal Classifier (BOC)
– Probabilistic learning criteria: measures of P(prediction | D) or P(hypothesis | D)
– BOC: a gold standard for probabilistic learning criteria
•
Gibbs Classifier
– Randomly sample h according to P(h | D), then use to classify
– Ratio bound: error no worse than 2 • Bayes optimal error
– MCMC methods (Gibbs sampling): Monte Carlo integration over H
•
Simple Bayes aka Naïve Bayes
– Assumption of conditional independence of attributes given classification
– Naïve Bayes classifier: factors conditional distribution of x given label v
v NB  argmax P v j  P x i | v j 
v j V
CIS 730: Introduction to Artificial Intelligence
i
Kansas State University
Department of Computing and Information Sciences
Lecture 28:
NLP Survey
•
More on Simple Bayes, aka Naïve Bayes
•
Learning in Natural Language Processing (NLP)
– Learning over text: problem definitions
– Bayesian approaches to NLP
• Issues: word sense disambiguation, part-of-speech tagging
• Applications: spelling; reading/posting news; web search, IR, digital libraries
•
Layers: Syntax, Semantics, Pragmatics, Discourse
•
Problems: Scanning, Parsing, Typing (POS Tagging), Pragmatics, Discourse
•
Thursday: Final Exam Review
CIS 730: Introduction to Artificial Intelligence
Kansas State University
Department of Computing and Information Sciences