Loopy Belief Propagation for Approximate Inference: an Empirical Study

Loopy Belief Propagation for Approximate Inference: an Empirical Study

467 Loopy Belief Propagation for Approximate Inference: An Empirical Study Kevin P. Murphy and Yair Weiss and Michael I. Jordan Computer Science Division University of California Berkeley, CA 94705 { murphyk,yweiss,jordan }@cs.berkeley.edu Abstract formance of "loopy belief propagation". This refers to using the well-known Pearl polytree algorithm [12] on a Recently, researchers have demonstrated that Bayesian network with loops (undirected cycles). The "loopy belief propagation" - the use of algorithm is an exact inference algorithm for singly­ Pearl's polytree algorithm in a Bayesian connected networks - the beliefs converge to the cor­ network with loops - can perform well rect marginals in a number of iterations equal to the in the context of error-correcting codes. diameter of the graph.1 However, as Pearl noted, the The most dramatic instance of this is the same algorithm will not give the correct beliefs for mul­ near Shannon-limit performance of "Turbo tiply connected networks: Codes" - codes whose decoding algorithm is equivalent to loopy belief propagation in a When loops are present, the network is no chain-structured Bayesian network. longer singly connected and local propaga­ In this paper we ask: is there something spe­ tion schemes will invariably run into trouble ... If we ignore the existence of loops and cial about the error-correcting code context, or does loopy propagation work as an ap­ permit the nodes to continue communicat­ proximate inference scheme in a more gen­ ing with each other as if the network were eral setting? We compare the marginals com­ singly connected, messages may circulate in­ puted using loopy propagation to the exact definitely around the loops and the process ones in four Bayesian network architectures, may not converge to a stable equilibrium ... including two real-world networks: ALARM Such oscillations do not normally occur in and QMR. We find that the loopy beliefs of­ probabilistic networks ...which tend to bring ten converge and when they do, they give a all messages to some stable equilibrium as good approximation to the correct marginals. time goes on. However, this asymptotic equi­ However, on the QMR network, the loopy be­ librium is not coherent, in the sense that it liefs oscillated and had no obvious relation­ does not represent the posterior probabilities ship to the correct posteriors. We present of all nodes of the network [12, p.l95] some initial investigations into the cause of Despite these reservations, Pearl advocated the use of these oscillations, and show that some sim­ belief ple methods of preventing them lead to the propagation in loopy networks as an approxima­ tion scheme (J. Pearl, personal communication) and wrong results. exercise 4.7 in [12] investigates the quality of the ap­ proximation when it is applied to a particular loopy belief network. 1 Introduction Several groups have recently reported excellent exper­ The task of calculating posterior marginals on nodes imental results by using this approximation scheme­ in an arbitrary Bayesian network is known to be NP­ by running algorithms equivalent to Pearl's algorithm hard [5]. This is true even for the seemingly easier on networks with loops. Perhaps the most dramatic task of calculating approximate posteriors [6]. Never­ instance of this performance is in an error correcting theless, due to the obvious practical importance of this code scheme known as "Turbo Codes" [4]. These codes task, there has been considerable interest in assessing have been described as "the most exciting and poten­ the quality of different approximation schemes, in an tially important development in coding theory in many attempt to delimit the types of networks and parame­ ter regimes for which each scheme works best. 1 This assumes parallel updating of all nodes. The algo­ rithm can also be implemented in a centralized fashion in In this paper we investigate the approximation per- which case it converges in two iterations [13). 468 Murphy, Weiss, and Jordan years" [11] and have recently been shown [9, 10] to uti­ Is there something special about the error-correcting lize an algorithm equivalent to belief propagation in code setting, or does loopy propagation work as an a network with loops. Although there is widespread approximation scheme for a wider range of networks? agreement in the coding community that these codes "represent a genuine, and perhaps historic, break­ through" [11], a theoretical understanding of their per­ 2 The algorithm formance has yet to be achieved. Yet McEliece et. a! conjectured that the performance of loopy belief prop­ For completeness, we briefly summarize Pearl's belief agation on the Turbo code structure was a special case propagation algorithm. Each node X computes a be­ of a more general phenomenon: lief BEL(:x)= P(X = :xiE), where E denotes the ob­ served evidence, by combining messages from its chil­ We believe there are general undiscovered dren ..\y;(:x) and messages from its parents 1rx(uk). theorems about the performance of belief (Following Peot and Shachter [13], we incorporate ev­ propagation on loopy DAGs. These theo­ idence by letting a node send a message to itself, rems, which may have nothing directly to ..\x(:x).) do with coding or decoding will show that BEL(:x) = a..\(:x)1r(x) (1) in some sense belief propagation "converges where: with high probability to a near-optimum ,x(tl(x)= ..\x(x)IT ..\�}(x) (2) value" of the desired belief on a class of loopy j DAGs [10]. and: Progress in the analysis of loopy belief propagation (3) 7r('l(x) = LP(X = xiU = u) IT1r�) (uk) has been made for the case of networks with a single u k loop [18, 19, 2, 1]. For the sum-product (or "belief update") version it can be shown that: The message X passes to its parent U; is given by: • Unless all the conditional probabilities are deter­ ministic, belief propagation will converge. • There is an analytic expression relating the cor­ (4) rect marginals to the loopy marginals. The ap­ and the message X sends to its child Yj is given by: proximation error is related to the convergence rate of the messages- the faster the convergence 7r�;+l)(:x)= a?C('l(x)..\x(x) IT.>.W( x) (5) the more exact the approximation. k;Cj • If the hidden nodes are binary, then thresholding For noisy-or links between parents and children, there the loopy beliefs is guaranteed to give the most exists an analytic expression for 1r(x) and Ax( u;) that probable assignment, even though the numerical avoids the exhaustive enumeration over parent config­ value of the beliefs may be incorrect. This result urations [12]. only holds for nodes in the loop. We made a slight modification to the update rules in In the max-product (or "belief revision") version, that we normalized both ..\ and 1r messages at each Weiss [19] showed that ( 1) belief propagation may con­ iteration. As Pearl [12] pointed out, normalizing the verge to a stable value or oscillate in a limit cycle and messages makes no difference to the final beliefs but (2) if it converges then it is guaranteed to give the cor­ avoids numerical underflow. rect assignment of values to the hidden nodes. This Nodes were updated in parallel: at each iteration all result is independent of the arity of the nodes and nodes calculated their outgoing messages based on the whether the nodes are inside or outside the loop. incoming messages of their neighbors from the pre­ For the case of networks with multiple loops, Richard­ vious iteration. The messages were said to converge son [14] has analyzed the special case of Turbo codes. if none of the beliefs in successive iterations changed He has shown that fixed points of the sum-product ver­ by more than a small threshold (10-4). All messages sion always exist, and has given sufficient conditions were initialized to a vector of ones; random initializa­ under which they will be unique and stable (although tion yielded similar results, since the initial conditions verifying these conditions may be difficultfor large net­ rapidly get "washed out". works). For comparison, we also implemented likelihood To summarize, what is currently known about loopy weighting [17], which is a simple form of importance propagation is that (1) it works very well in an error­ sampling. Like any sampling algorithm, the errors can correcting code setting and (2) there are conditions for be driven towards zero by running the algorithm for a single-loop network for which it can be guaranteed long enough; in this paper, we usually used 200 sam­ to work well. In this paper we investigate loopy prop­ ples, so that the total amount of computation time was agation empirically under a wider range of conditions. roughly comparable (to within an order of magnitude) Loopy Belief Propagation 469 to loopy propagation. We did not implement some of 3.3 The ALARM network the more sophisticated versions of likelihood weight­ ing, such as Markov blanket scoring (16], since our goal Figure 3 shows the structure of the ALARM network in this paper was to evaluate loopy propagation rather - a Bayesian network for monitoring patients in in­ than exhaustively compare the performance of alter­ tensive care. This network was used by (3] to compare native algorithms. (For a more careful evaluation of various inference algorithms. The arity of the nodes likelihood weighted sampling in the case of the QMR ranges from two to four and all conditional distribu­ network, see (8].) tions are represented by tables. The structure and the CPTs were downloaded from Nir Friedman's Bayesian network repository at: www. cs. huj i. ac. il/"nir. 3 The networks 3.4 We used two synthetic networks, PYRAMID and The QMR-DT network toyQMR, and two real world networks, ALARM and The QMR-DT is a bipartite network whose structure is QMR.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    10 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us