Lecture 3 – Bayesian Graphical Models
Advanced Probabilistic Machine Learning Lecture 3 – Bayesian Graphical Models
Riccardo Sven Risuleo Division of Systems and Control Department of Information Technology Uppsala University
[email protected] www.it.uu.se/katalog/ricri923
1 / 39 [email protected] Bayesian Graphical Models Summary of Lecture 2 (I)
Bayesian lLinear regression model
y = wwT x + ε , ε (0, σ2), n = 1, ..., N n n n n ∼ N w p(w). ∼
Present assumptions:
1. yn – observed random variable. 2. w – unknown deterministicw – unknown random variable. (difference from SML)
3. xn – known deterministic variable.
4. εn – unknown random variable. 5. σ – known deterministic variable.
2 / 39 [email protected] Bayesian Graphical Models Summary of Lecture 2 (II)
Remember Bayes’ theorem
p(w, y) p(y w)p(w) p(w y) = = | | p(y) p(y)
Prior distribution: p(w) describes the knowledge we have about • w before observing any data. Likelihood: p(y w) described how “likely” the observed data is • | for a particular parameter value. Posterior distribution: p(w y) summarize all our knowledge • | about w from the observed data and the model.
In Bayesian linear regression we use a Gaussian distribution as prior p(w) = (w; m , Σ ) N 0 0
3 / 39 [email protected] Bayesian Graphical Models Summary of Lecture 2 (III)
p(xa, xb) Thm 1
Thm 2
Thm 3 p(x ) p(x x ) a b| a
Col 1 = Thm 3 + Thm 2
p(x ) Col 2= Thm 3 + Thm 1 p(x x ) b a| b 4 / 39 [email protected] Bayesian Graphical Models Summary of Lecture 2 (IV)
Plot of the situation after one measurement has arrived.
w1 w1 w1
w0 w0 w0 Prior Likelihood Posterior/prior,