Computer Vision Group Prof. Daniel Cremers

4a. Inference in Graphical Models Inference on a Chain (Rep.)

• The first values of µα and µβ are:

• The partition function can be computed at any node:

• Overall, we have O(NK2) operations to compute the marginal

Machine Learning for Dr. Rudolph Triebel Computer Vision 2 Computer Vision Group More General Graphs

The message-passing algorithm can be extended to more general graphs: Directed Undirected Polytree Tree Tree

It is then known as the sum-product algorithm. A special case of this is .

Machine Learning for Dr. Rudolph Triebel Computer Vision 3 Computer Vision Group More General Graphs

The message-passing algorithm can be extended to more general graphs: Undirected Tree An undirected tree is defined as a graph that has exactly one path between any two nodes

Machine Learning for Dr. Rudolph Triebel Computer Vision 4 Computer Vision Group More General Graphs

The message-passing algorithm can be extended to more general graphs: Directed A directed tree has Tree Conversion from only one node a directed to an without parents and undirected tree is all other nodes no problem, have exactly one because no links parent are inserted

The same is true for the conversion back to a directed tree

Machine Learning for Dr. Rudolph Triebel Computer Vision 5 Computer Vision Group More General Graphs

The message-passing algorithm can be extended to more general graphs:

Polytree Polytrees can contain nodes with several parents, therefore moralization can remove independence relations

Machine Learning for Dr. Rudolph Triebel Computer Vision 6 Computer Vision Group Factor Graphs

• The Sum-product algorithm can be used to do inference on undirected and directed graphs. • A representation that generalizes directed and undirected models is the factor graph.

f(x ,x ,x )=p(x )p(x )p(x x ,x ) 1 2 3 1 2 3 | 1 2 Directed graph Factor graph

Machine Learning for Dr. Rudolph Triebel Computer Vision 7 Computer Vision Group Factor Graphs

• The Sum-product algorithm can be used to do inference on undirected and directed graphs. • A representation that generalizes directed and undirected models is the factor graph.

Undirected graph Factor graph

Machine Learning for Dr. Rudolph Triebel Computer Vision 8 Computer Vision Group Factor Graphs

Factor graphs x1 x3 fa • can contain multiple factors for the same nodes fb • are more general than x4 undirected graphs • are bipartite, i.e. they consist of two kinds of nodes and all edges connect nodes of different kind

Machine Learning for Dr. Rudolph Triebel Computer Vision 9 Computer Vision Group Factor Graphs x1 x3 • Directed trees convert to tree-structured factor graphs

• The same holds for x4 undirected trees • Also: directed polytrees

convert to tree-structured x1 x3 factor graphs fa • And: Local cycles in a directed graph can be removed by converting to a x4 factor graph

Machine Learning for Dr. Rudolph Triebel Computer Vision 10 Computer Vision Group The Sum-Product Algorithm Assumptions: • all variables are discrete • the factor graph has a tree structure The factor graph represents the joint distribution as a product of factor nodes:

p(x)= fs(xs) s Y The marginal distribution at a given node x is p(x)= p(x) x x X\

Machine Learning for Dr. Rudolph Triebel Computer Vision 11 Computer Vision Group The Sum-Product Algorithm For a given node x the joint can be written as

p(x)= Fs(x, Xs) s ne(x) 2Y Product of all factors associated p(x)= Fs(x, Xs) Thus, we have with f x x s ne(x) s X\ 2Y Key insight: Sum and product can be exchanged!

p(x)= Fs(x, Xs)= µf x(x) s! s ne(x) Xs s ne(x) 2Y X 2Y “Messages from factors to node x”

Machine Learning for Dr. Rudolph Triebel Computer Vision 12 Computer Vision Group The Sum-Product Algorithm

The factors in the messages can be factorized further:

Fs(x, Xs)=fs(x, x1,...,xM )G1(x1,Xs1 ) ...GM (xM ,XsM )

The messages can then be computed as

µf x(x)= fs(x, x1,...,xM ) Gm(xm,Xs ) s! ··· m x1 xM m ne(fs) x Xs X X 2 Y \ Xm

= fs(x, x1,...,xM ) µx f (xm) ··· m! s x1 xM m ne(fs) x X X 2 Y \ “Messages from nodes to factors”

Machine Learning for Dr. Rudolph Triebel Computer Vision 13 Computer Vision Group The Sum-Product Algorithm

The factors G of the neighboring nodes can again be factorized further:

GM (xm,Xsm )= Fl(xm,Xml )

l ne(xm) fs 2 Y \ This results in the exact same situation as above! We can now recursively apply the derived rules:

µx f (xm)= Fl(xm,Xm ) m! s l l ne(xm) fs Xm 2 Y \ Xl

= µf x (xm) l! m l ne(xm) fs 2 Y \ Machine Learning for Dr. Rudolph Triebel Computer Vision 14 Computer Vision Group The Sum-Product Algorithm Summary marginalization: 1.Consider the node x as a root note 2.Initialize the recursion at the leaf nodes as: µ f x ( x )=1 (var) or µ x f ( x )= f ( x ) (fac) ! ! 3.Propagate the messages from the leaves to the root x 4.Propagate the messages back from the root to the leaves 5.We can get the marginals at every node in the graph by multiplying all incoming messages

Machine Learning for Dr. Rudolph Triebel Computer Vision 15 Computer Vision Group The Max-Sum Algorithm Sum-product is used to find the marginal distributions at every node, but: How can we find the setting of all variables that maximizes the joint probability? And what is the value of that maximal probability? Idea: use sum-product to find all marginals and then report the value for each node x that maximizes the marginal p(x) However: this does not give the overall maximum of the joint probability

Machine Learning for Dr. Rudolph Triebel Computer Vision 16 Computer Vision Group The Max-Sum Algorithm Observation: the max-operator is distributive, just like the multiplication used in sum-product: max(ab, ac)=a max(b, c)ifa 0 Idea: use max instead of sum as above and exchange it with the product Chain example: 1 max p(x)= max ...max[ 1,2(x1,x2) ... N 1,N (xN 1,xN )] x Z x1 1 = max [ 1,2(x1,x2)[...max N 1,N (xN 1,xN )]] Z x1

Message passing can be used as above!

Machine Learning for Dr. Rudolph Triebel Computer Vision 17 Computer Vision Group The Max-Sum Algorithm

To find the maximum value of p(x), we start again at the leaves and propagate to the root. Two problems: • no summation, but many multiplications; this leads to numerical instability (very small values) • when propagating back, multiple configurations of x can maximize p(x), leading to wrong assignments of the overall maximum Solution to the first: Transform everything into log-space and use sums

Machine Learning for Dr. Rudolph Triebel Computer Vision 18 Computer Vision Group The Max-Sum Algorithm Solution to the second problem: Keep track of the arg max in the forward step, i.e. store at each node which value was responsible for the maximum:

(xn) = arg max[ln fn 1,n(xn 1,xn)+µxn 1 fn 1,n (xn)] xn 1 !

Then, in the back-tracking step we can recover the arg max by recursive substitution of: max max xn 1 = (xn )

Machine Learning for Dr. Rudolph Triebel Computer Vision 19 Computer Vision Group Other Inference Algorithms Junction Tree Algorithm: • Provides exact inference on general graphs. • Works by turning the initial graph into a junction tree and then running a sum-product-like algorithm • A junction tree is obtained from an undirected model by triangulation and mapping cliques to nodes and connections of cliques to edges • It is the maximal spanning tree of cliques Problem: Intractable on graphs with large cliques. Cost grows exponentially with the number of variables in the largest clique (“tree width”).

Machine Learning for Dr. Rudolph Triebel Computer Vision 20 Computer Vision Group Other Inference Algorithms Loopy Belief Propagation: • Performs Sum-Product on general graphs, particularly when they have loops • Propagation has to be done several times, until a convergence criterion is met • No guarantee of convergence and no global optimum • Messages have to be scheduled • Initially, unit messages passed across all edges • Approximate, but tractable for large graphs

Machine Learning for Dr. Rudolph Triebel Computer Vision 21 Computer Vision Group Conditional Random Fields

• Another kind of undirected is known as (CRF). • CRFs are used for classification where labels are represented as discrete random variables y and features as continuous random variables x • A CRF represents the conditional probability

where w are parameters learned from training data. • CRFs are discriminative and MRFs are generative

Machine Learning for Dr. Rudolph Triebel Computer Vision 22 Computer Vision Group Conditional Random Fields

Derivation of the formula for CRFs:

In the training phase, we compute parameters w that maximize the posterior: where (x*,y*) is the training data and p(w) is a Gaussian prior. In the inference phase we maximize

Machine Learning for Dr. Rudolph Triebel Computer Vision 23 Computer Vision Group Conditional Random Fields

Typical example: observed variables

xi,j are intensity values of pixels in an image and

hidden variables yi,j are object labels

Note: the definition of xi,j and yi,j is different from the one in C.M. Bishop (pg.389)!

Machine Learning for Dr. Rudolph Triebel Computer Vision 24 Computer Vision Group CRF Training

We minimize the negative log-posterior:

Computing the likelihood is intractable, as we have to compute the partition function for each w. We can approximate the likelihood using pseudo-likelihood:

Markov blanket where Ci: All cliques containing yi

Machine Learning for Dr. Rudolph Triebel Computer Vision 25 Computer Vision Group Pseudo Likelihood

Machine Learning for Dr. Rudolph Triebel Computer Vision 26 Computer Vision Group Pseudo Likelihood

Pseudo-likelihood is computed only on the Markov

blanket of yi and its corresp. feature nodes.

Machine Learning for Dr. Rudolph Triebel Computer Vision 27 Computer Vision Group Potential Functions

• The only requirement for the potential functions is that they are positive. We achieve that with:

where f is a compatibility function that is large if the

labels yC fit well to the features xC. • This is called the log-linear model. • The function f can be, e.g. a local classifier

Machine Learning for Dr. Rudolph Triebel Computer Vision 28 Computer Vision Group CRF Training and Inference Training: • Using pseudo-likelihood, training is efficient. We have to minimize:

Log-pseudo-likelihood Gaussian prior • This is a convex function that can be minimized using gradient descent Inference: • Only approximatively, e.g. using loopy belief propagation

Machine Learning for Dr. Rudolph Triebel Computer Vision 29 Computer Vision Group Summary • Undirected Graphical Models represent conditional independence more intuitively using graph separation • Their factorization is done based on potential functions The normalizer is called the partition function, which in general is intractable to compute • Inference in graphical models can be done efficiently using the sum-product algorithm (message passing). • Another inference algorithm is loopy belief propagation, which is approximate, but tractable • Conditional Random Fields are a special kind of MRFs and can be used for classification Machine Learning for Dr. Rudolph Triebel Computer Vision 30 Computer Vision Group Computer Vision Group Prof. Daniel Cremers

5. Hidden Markov Models Graphical Representation (Rep.)

We can describe the overall process using a Dynamic Bayes Network:

• This incorporates the following Markov assumptions: (measurement) (state)

Machine Learning for Dr. Rudolph Triebel Computer Vision 32 Computer Vision Group Graphical Representation

We can describe the overall process using a of latent variables:

Notation Discrete differs from Bishop! Variables

• This incorporates the following Markov assumptions: (measurement) (state)

Machine Learning for Dr. Rudolph Triebel Computer Vision 33 Computer Vision Group Example “Occasionally dishonest casino”:

• observations: faces of a die zt 1, 2,...,6 2 { } • hidden states: two different dice, one fair, one loaded

Rolls: 664153216162115234653214356634261655234232315142464156663246 Die: LLLLLLLLLLLLLLFFFFFFLLLLLLLLLLLLLLFFFFFFFFFFFFFFFFFFLLLLLLLL

Machine Learning for Dr. Rudolph Triebel Computer Vision 34 Computer Vision Group Formulation as HMM 1.Discrete random variables

• Observation variables: {zn}, n = 1..N

• State variables (unobservable): {xn}, n = 1..N

• Number of states K: xnє{1..K} Model Parameters θ 2.Transition model p(xi |xi-1)

• Markov assumption (xi only depends on xi-1) • Represented as a K×K transition matrix A

• Initial probability: p(x0) repr. as π1, π2, π3

3.Observation model p(zi|xi) with parameters φ • Observation only depends on the current state • Example: output of a “local” place classifier Machine Learning for Dr. Rudolph Triebel Computer Vision 35 Computer Vision Group The Trellis Representation

time A A k=1 11 11

k=2

k=3 A33 A33 n-2 n-1 n

Machine Learning for Dr. Rudolph Triebel Computer Vision 36 Computer Vision Group Application Example (1)

• Given an observation sequence z1,z2,z3… • Assume that the model parameters θ =(A, π, φ) are known • What is the probability that the given observation sequence is actually observed under this model, i.e. p(Z| θ)? • If we are given several different models, we can choose the one with highest probability • Expressed as a supervised learning problem, this can be interpreted as the inference step (classification step)

Machine Learning for Dr. Rudolph Triebel Computer Vision 37 Computer Vision Group Application Example (2)

• Given an observation sequence z1,z2,z3… • Assume that the model parameters θ =(A, π, φ) are known

• What is the state sequence x1,x2,x3… that explains best the given observation sequence? • In the case of place recognition: which is the sequence of truly visited places that explains best the sequence of obtained place labels (classifications)?

Machine Learning for Dr. Rudolph Triebel Computer Vision 38 Computer Vision Group Application Example (3)

• Given an observation sequence z1,z2,z3… • What are the optimal model parameters θ =(A, π, φ)? • This can be interpreted as the training step • It is in general the most difficult problem

Machine Learning for Dr. Rudolph Triebel Computer Vision 39 Computer Vision Group Summary: 3 Operations on HMMs

1. Compute data likelihood p(Z|θ) from a known model • Can be computed with the forward-backward algorithm

2. Compute optimal state sequence with a known model • Can be computed with the Viterbi-Algorithm

3. Learn model parameters for an observation sequence • Can be computed using Expectation-Maximization (or Baum-Welch)

Machine Learning for Dr. Rudolph Triebel Computer Vision 40 Computer Vision Group 1. Computing the Data Likelihood

• Assume: given a state sequence x1,x2,x3… Two possible operations:

• Filtering: computes p ( x t z 1: t ) , i.e. state | probability only based on previous observations

• Smoothing: computes p ( x t z 1: T ) , state | probability based on all observations (including those from the future) filtered smoothed 1 1

0.5 0.5 p(loaded) p(loaded) Filtering Smoothing

0 0 0 50 100 150 200 250 300 0 50 100 150 200 250 300 roll number roll number

Machine Learning for Dr. Rudolph Triebel Computer Vision 41 Computer Vision Group The Forward Algorithm • First, we compute the prediction from the last time step:

p(xt = j z1:t 1)= p(xt = j xt 1 = i)p(xt 1 = i z1:t 1) | | | i X • Then, we do the update using Bayes rule:

↵t(j):=p(xt = j z1:t)=p(xt = j zt, z1:t 1) | | 1 = p(zt xt = j, z1:t 1)p(xt = j z1:t 1) Zt | | • This is exactly the same as the Bayes filter from the first lecture!

Machine Learning for Dr. Rudolph Triebel Computer Vision 42 Computer Vision Group The Forward-Backward Algorithm

• As before we set ↵ t(j):=p(xt = j z1:t) | • We also define t(j):=p(zt+1:T xt = j) |

Machine Learning for Dr. Rudolph Triebel Computer Vision 43 Computer Vision Group The Forward-Backward Algorithm

• As before we set ↵ t(j):=p(xt = j z1:t) | • We also define t(j):=p(zt+1:T xt = j) | • This can be recursively computed (backwards): t 1(i)=p(zt:T xt 1 = i) |

= p(xt = j, zt, zt+1:T xt 1 = i) | j X = p(zt+1:T xt = j, xt 1 = i, zt)p(xt = j, zt xt 1 = i) | | j X = p(zt+1:T xt = j)p(zt xt = j, xt 1 = i)p(xt = j xt 1 = i) | | | j X = t(j)p(zt xt = j)p(xt = j xt 1 = i) | | j X Machine Learning for Dr. Rudolph Triebel Computer Vision 44 Computer Vision Group The Forward-Backward Algorithm

• As before we set ↵ t(j):=p(xt = j z1:t) | • We also define t(j):=p(zt+1:T xt = j) | • This can be recursively computed (backwards): t 1(i)=p(zt:T xt 1 = i) |

= t(j)p(zt xt = j)p(xt = j xt 1 = i) | | j X • This is exactly the same as the message-passing algorithm (sum-product)!

• forward messages ↵ t (vector of length K) • backward messages t (vector of length K)

Machine Learning for Dr. Rudolph Triebel Computer Vision 45 Computer Vision Group 2. Computing the Most Likely States

• Goal: find a state sequence x1,x2,x3… that maximizes the probability p(X,Z|θ)

• Define t(j) := max p(x1:t 1,xt = j z1:t) x1,...,xt 1 | This is the probability of state j by taking the most probable path.

Machine Learning for Dr. Rudolph Triebel Computer Vision 46 Computer Vision Group 2. Computing the Most Likely States

• Goal: find a state sequence x1,x2,x3… that maximizes the probability p(X,Z|θ)

• Define t(j) := max p(x1:t 1,xt = j z1:t) x1,...,xt 1 | This can be computed recursively:

t(j) := max t 1(i)p(xt xt 1)p(zt xt) i | | we also have to compute the argmax:

at(j) := arg max t 1(i)p(xt xt 1)p(zt xt) i | |

Machine Learning for Dr. Rudolph Triebel Computer Vision 47 Computer Vision Group The Viterbi algorithm • Initialize:

• δ(x0)= p(x0) p(z0 | x0)

• ψ(x0)= 0 • Compute recursively for n=1…N:

• δ(xn)= p(zn|xn) max [δ(xn-1) p(xn|xn-1)] xn-1

• a(xn)= argmax [δ(xn-1) p(xn|xn-1)] xn-1

• On termination:

• p(Z,X|θ) = max δ(xN) xN * • xN = argmax δ(xN) xN • Backtracking:

* • xn = a(xn+1) Machine Learning for Dr. Rudolph Triebel Computer Vision 48 Computer Vision Group 3. Learning the Model Parameters

• Given an observation sequence z1,z2,z3… • Find optimal model parameters θ • We need to maximize the likelihood p(Z|θ) • Can not be solved in closed form • Iterative algorithm: Expectation Maximization (EM) or for the case of HMMs: Baum-Welch algorithm

Machine Learning for Dr. Rudolph Triebel Computer Vision 49 Computer Vision Group Expectation Maximisation

• Objective: Find the model parameters knowing the observations: π,A,φ • Result: •Train the HMM to recognize sequences of input •Train the HMM to generate sequences of input • Technique: Expectation Maximisation •E: Find the best state sequence given the parameters •M: Find the parameters using the state sequence •Maximisation of the log-likelihood:

Machine Learning for Dr. Rudolph Triebel Computer Vision 50 Computer Vision Group The Baum-Welsh algorithm

• E-Step (assuming we know π,A,φ, i.e. θold) • Define the posterior probability of being in state i at step k:

• Define γ(xn)= p(xn|Z)

Machine Learning for Dr. Rudolph Triebel Computer Vision 51 Computer Vision Group The Baum-Welsh algorithm

• E-Step (assuming we know π,A,φ, i.e. θold) • Define the posterior probability of being in state i at step k:

• Define γ(xn)= p(xn|Z)

• It follows that γ(xn)= α(xn) β(xn) / p(Z)

Machine Learning for Dr. Rudolph Triebel Computer Vision 52 Computer Vision Group The Baum-Welsh algorithm • E-Step (assuming we know π,A,φ, i.e. θold) • Define the posterior probability of being in state i at step k:

• Define γ(xn)= p(xn|Z)

• It follows that γ(xn)= α(xn) β(xn) / p(Z)

• Define ξ(xn-1 ,xn)= p(xn-1 ,xn|Z) • It follows that

ξ(xn-1 ,xn)= α(xn-1)p(zn|xn)p(xn|xn-1)β(xn) / p(Z) • We need to compute: Expected complete data old old Q(θ,θ )= Σ p(X|Z, θ )log p(Z,X|θ) log-likelihood X Machine Learning for Dr. Rudolph Triebel Computer Vision 53 Computer Vision Group The Baum-Welsh algorithm

• Maximizing Q also maximizes the likelihood: p(Z|θ) ≥ p(Z|θold) • M-Step: x (x)x1k • ⇡k = (x)x jP=1 x 1j here, we need forwardP P and backward step! T ⇠(xt 1,j,xtk) • t=2 Ajk = K T ⇠(xt 1,j,xtl) lP=1 t=2 • With these new values,P P Q is recomputed • This is done until the likelihood does not increase anymore (convergence)

Machine Learning for Dr. Rudolph Triebel Computer Vision 54 Computer Vision Group The Baum-Welsh algorithm - summary • Start with an initial estimate of θ=(π,A,φ) e.g. uniformly and k-means for φ • Compute Q(θ,θold) (E-Step) • Maximize Q (M-step) • Iterate E and M until convergence • In each iteration one full application of the forward-backward algorithm is performed • Result gives a local optimum • For other local optima, the algorithm needs to be started again with new initialization

Machine Learning for Dr. Rudolph Triebel Computer Vision 55 Computer Vision Group The Scaling problem

• Probability of sequences

p(x ...) << 1 i | i Y <1 •Probabilities are very small •The product of the terms soon is very small • Usually: converting to log-space works • But: we have sums of products! • Solution: Rescale/Normalise the probability during the computation, e.g.:

^ α(xn)= α(xn) / p(z1,z2,…,zn)

Machine Learning for Dr. Rudolph Triebel Computer Vision 56 Computer Vision Group Summary • HMMs are a way to model sequential data • They assume discrete states • Three possible operations can be performed with HMMs: •Data likelihood, given a model and an observation •Most likely state sequence, given a model and an observation •Optimal Model parameters, given an observation • Appropriate scaling solves numerical problems • HMMs are widely used, e.g. in

Machine Learning for Dr. Rudolph Triebel Computer Vision 57 Computer Vision Group