Introduction_to_Machine_Learning_Lec1x

Download Report

Transcript Introduction_to_Machine_Learning_Lec1x

Foundations of Machine
Learning (CS725)
Autumn 2011
Instructor: Prof. Ganesh Ramakrishnan
TAs: Ajay Nagesh, Amrita Saha,
Kedharnath Narahari
The grand goal
From the movie 2001: A Space Odyssey (1968)
Outline
• Introduction to Machine Learning
– What is machine learning?
– Why machine learning
– How machine learning relates to other fields
• Real world applications
• Machine Learning : Models and methods
–
–
–
–
Supervised
Unsupervised
Semi-supervised
Active learning
• Course Information
– Tools and software
– Pre-requisites
INTRODUCTION TO
MACHINE LEARNING
Intelligence
• Ability for abstract thought, understanding,
communication, reasoning, planning, emotional
intelligence, problem solving, learning
• The ability to learn and/or adapt is generally
considered a hallmark of intelligence
Learning and Machine Learning
• ``Learning denotes changes in the system that
are adaptive in the sense that they enable the
system to do the task(s) drawn from the same
population more efficiently and more
effectively the next time.''--Herbert Simon
• Machine Learning is concerned with the
development of algorithms and techniques
that allow computers to learn.
Machine Learning
• “Machine learning studies the process of
constructing abstractions (features, concepts,
functions, relations and ways of acting)
automatically from data.”
E.g.: Learning concepts and words
“tufa”
“tufa”
“tufa”
Can you pick out the tufas?
Source: Josh Tenenbaum
Why Machine Learning ?
• Human expertise does not exist (e.g. Martian
exploration)
• Humans cannot explain their expertise or reduce
it to a rule set, or their explanation is incomplete
and needs tuning (e.g. speech recognition)
• Situation changing in time (e.g. spam/junk email)
• Humans are expensive to train up (e.g. zipcode
recognition)
• There are large amounts of data (e.g. discover
astronomical objects)
APPLICATIONS OF
MACHINE LEARNING
Data Data Everywhere …
• Library of Congress text database of ~20 TB
• AT&T 323 TB, 1.9 trillion phone call records.
• World of Warcraft utilizes 1.3 PB of storage to
maintain its game.
• Avatar movie reported to have taken over 1 PB of
local storage at WetaDigital for the rendering of the
3D CGI effects.
• Google processes ~24 PB of data per day.
• YouTube: 24 hours of video uploaded every minute.
More video is uploaded in 60 days than all 3 major
US networks created in 60 years. According to Cisco,
internet video will generate over 18 EB of traffic per
month in 2013.
Information Overload
Machine Learning to the rescue
• Machine Learning is one of the front-line
technologies to handle Information Overload
• Business
– Mining correlations, trends, spatio-temporal predictions.
– Efficient supply chain management.
– Opinion mining and sentiment analysis.
– Recommender systems.
Fields related to Machine Learning
Fields related to Machine Learning
• Artificial Intelligence: computational intelligence
• Data Mining: searching through large volumes of
data
• Neural Networks: neural/brain inspired methods
• Signal Processing: signals, video, speech, image
• Pattern Recognition: labeling data
• Robotics: building autonomous robots
Application of Machine Learning
Deep Blue
and the chess
Challenge
RoboCup
Online Poker
Application of Machine Learning
• Computation Biology
(Structure learning)
• Animation and Control
• Tracking and activity
recognition
Application of Machine Learning
• Application in speech and
Natural Language processing
• Probabilistic Context Free
Grammars
• Graphical Models
• Social network graph
analysis, causality
analysis
Deep Q and A: IBM Watson
• Deep Question and Answering : Jeopardy challenge
• Watson emerged winner when pitted against all time
best rated players in the history of Jeopardy
Source: IBM Research
MACHINE LEARNING
MODELS AND METHODS
Machine Learning Process

How to do the learning actually?
Learning (Formally)



Task
 To apply some machine learning method to the data
obtained from a given domain (Training Data)
 The domain has some characteristics, which we are trying
to learn (Model)
Objective
 To minimise the error in prediction
Types of Learning
 Supervised Learning
 Unsupervised Learning
 Semi-Supervised Learning
 Active Learning
Supervised Learning


Classification / Regression problem
Where some samples of data (Training data) with the
correct class labels are provided.


Using knowledge from training data, the classifier/ regressor
model is learnt




i.e. Some correspondence between input (X) & output (Y) given
i.e. Learn some function f : f(X) = Y
f may be probabilistic/deterministic
Learning the model ≡ Fitting the parameters of model to
minimise prediction error
Model can then be tested on test-data
Regression


Linear regression
Uses
 Stock Prediction
 Outlier detection
Regression
Regression

Non Linear regression
All models are not good

Constrain the parameters
Classification
Supervised Classification example
f1
f2
f3
f4
Class label
d1
BearHead
d2
???
d3
DuckHead
LionHead
Source: LHI Animal Faces Dataset
Classification






Example:

Credit Scoring
Goal:

Differentiating between high-risk and lowrisk customers based on their income and
savings
Discriminant:

IF income > θ1 AND savings > θ2 THEN
low-risk ELSE high-risk
Discriminant is called 'hypothesis'
Input attribute space is called 'Feature Space'
Here Input data is 2-dimensional and the
output is binary
Other applications
Building non-linear classifiers

Curse of dimensionality
Application
What is the right hypothesis?
What is the right hypothesis for this
classification problem
What is the right hypothesis for this
regression problem
Which linear hypothesis is better

Max – Margin
Classifier
Other considerations


Feature extraction: which are the good features that
characterise the data
Model selection: picking the right model using some
scoring/fitting function:



It is important not only to provide a good predictor, but
also to assess accurately how “good” the model is on
unseen test data
So a good performance estimator is needed to rank the
model
Model averaging: Instead of picking a single model, it
might be better to do a weighted average over the bestfit models
Which hypothesis is better?



Unless you know something about the
distribution of problems your learning
algorithm will encounter, any hypothesis that
agrees with all your data is as good as any
other.
You have to make assumptions about the
underlying features.
Hence learning is inductive, not deductive.
Unsupervised Learning


Labels may be too expensive to generate or
may be completely unknown
There is lots of training data but with no class
labels assigned to it
???
Source: LHI Animal Faces Dataset
Unsupervised Learning


For example clustering
Clustering –


grouping similar objects
Similar in which way?
Clustering
Clustering Problems

How to tell which type of clustering is
desirable?
Semi-Supervised Learning

Supervised learning + Additional unlabeled data
Unsupervised learning + Additional labeled data

Learning Algorithm:




Start from the labeled data to build an initial classifier
Use the unlabeled data to enhance the model
Some Techniques:


Co-Training: two or more learners can be trained
using an independent set of different features
Or to model joint probability distribution of the
features and labels
Example

ideally...
Active Learning






Unlabeled data is easy to obtain; but labels may be very
expensive
For e.g. Speech recognizer
Active Learning
 Initially all data labels are hidden
 There is some charge for revealing every label
Active Learner will interactively query the user for labels
By intelligent querying, a lot less number of labels will be
required than in usual supervised training
But a bad algorithm might focus on unimportant or invalid
examples
Active Learning: Example

Ideally,
Active Learning: Example

Suppose data lies on a real line and the classifier discriminant
looks like


Theoretically we can prove that if the actual data distribution P can
be classified using some hypothesis hw in H






H= {hw}: hw(x) = 1 if x > w, 0 otherwise
Then to get a classifier with error 'e', we just need O(1/e) random
labeled samples from P
Now labels are sequences of 0s and 1s
Goal is to discover the pt 'w' where transition occurs
Find that using binary search
So only log (1/e) samples queried
Exponential improvement in terms of number of samples required
Active Learning and survelliance
Active Learning and sensor networks
How learning happens
Human
Machine
Memorize
k-Nearest Neighbours,
Case/Example-based learning
Observe someone else,
then repeat
Supervised Learning, Learning by
Demonstration
Keep trying until it works
(riding a bike)
Reinforcement Learning
20 Questions
Active Learning
Pattern matching
(faces, voices, languages)
Pattern Recognition
Guess that current trend will
continue (stock market, real
estate prices)
Regression
COURSE INFORMATION
Tools and Resources










Weka: http://www.cs.waikato.ac.nz/ml/weka
Scilab: http://www.scilab.org/
R-software: http://www.r-project.org/
RapidMiner: http://rapid-i.com/content/view/181/190/
Orange: http://orange.biolab.si/
KNIME: http://www.knime.org/
SVM Light: http://svmlight.joachims.org
ShogunToolbox: http://www.shogun-toolbox.org/
Elefant: http://elefant.developer.nicta.com.au
Google prediction API: http://code.google.com/apis/predict/
Course Info

Pre-requisites for course




Probability & Statistics
Basics of convex optimization
Basics of linear algebra
Online Materials

Online class-notes : http://www.cse.iitb.ac.in/~cs725/notes/classNotes/




Username: cs717
Password: cs717_student
Andrew Ng. Notes http://www.stanford.edu/class/cs229/materials.html and
video lecture series http://videolectures.net/andrew_ng/
Main Text Book: Pattern Recognition and Machine Learning – Christopher
Bishop

Reference: Hastie, Tibshirani, Friedman The elements of Statistical Learning
Springer Verlag