Gaussian Process Tutorial
Download
Report
Transcript Gaussian Process Tutorial
Daphne’s
Approximate
Group of
Students
Outline
Linear Regression
Unregularized
L2 Regularized
What is a GP?
Prediction with a GP
Relationship to SVM
Implications
What does this mean?
Linear Regression
Predicting Y given X
Y = wtx + n
w_ml = argmax
y[m+1] = w_mltx[m+1]
L2 Regularized Lin Reg
L2 Regularized (Gaussian Prior on w)
Y = wtx + n
w ~ N(0,S)
w_map = argmax blah + ||w||^2
What is a random process?
It’s a prior over functions
What is a Gaussian Process?
It’s a prior over functions that generalized a Gaussian
Random Vector
Prior over Y(x) ~ N(0,I)
Alternate Definition
The thing with Euler’s equation
This is weird
Not used to thinking of prior over Ys
Or are we?
We ARE used to thining about prior over w
What prior over y does this induce
Math P(w) -> P(Y)
Wow! This became a Gaussian Process!
Prediction with a GP
Predict y*[m+1] given y[1]…y[m]
We get a covariance = error bars
Wow! This prediction is the same as w_map but
we get error bars!
Generalize that shit - Covariance
Functions
Note that we have a thing here that is defined by
C(x1,x2) which can be kernelized
Has to be pos semidefinite
Is a kernel function
Relationship to SVM
Example
How do we reconcile these
views?
Does this change anything?