Properties, Learning Algorithms, and Applications of Chain Graphs and Bayesian Hypergraphs

Total Page:16

File Type:pdf, Size:1020Kb

Properties, Learning Algorithms, and Applications of Chain Graphs and Bayesian Hypergraphs University of South Carolina Scholar Commons Theses and Dissertations Fall 2019 Properties, Learning Algorithms, and Applications of Chain Graphs and Bayesian Hypergraphs Mohammad Ali Javidian Follow this and additional works at: https://scholarcommons.sc.edu/etd Part of the Computer Sciences Commons, and the Engineering Commons Recommended Citation Javidian, M. A.(2019). Properties, Learning Algorithms, and Applications of Chain Graphs and Bayesian Hypergraphs. (Doctoral dissertation). Retrieved from https://scholarcommons.sc.edu/etd/5558 This Open Access Dissertation is brought to you by Scholar Commons. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of Scholar Commons. For more information, please contact [email protected]. Properties,Learning Algorithms, and applications of Chain Graphs and Bayesian Hypergraphs by Mohammad Ali Javidian Bachelor of Science Shahid Bahonar University of Kerman 2003 Master of Science Sharif University of Technology 2013 Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in Computer Science and Engineering College of Engineering and Computing University of South Carolina 2019 Accepted by: Marco Valtorta, Major Professor Pooyan Jamshidi, Committee Member Linyuan Lu, Committee Member Chin-Tser Huang, Committee Member John Rose, Committee Member Cheryl L. Addy, Vice Provost and Dean of the Graduate School c Copyright by Mohammad Ali Javidian, 2019 All Rights Reserved. ii Dedication To Marco Valtorta, my dearest professor. iii Abstract Probabilistic graphical models (PGMs) use graphs, either undirected, directed, or mixed, to represent possible dependencies among the variables of a multivariate probability distri- bution. PGMs, such as Bayesian networks and Markov networks, are now widely accepted as a powerful and mature framework for reasoning and decision making under uncertainty in knowledge-based systems. With the increase of their popularity, the range of graphical models being investigated and used has also expanded. Several types of graphs with dif- ferent conditional independence interpretations - also known as Markov properties - have been proposed and used in graphical models. The graphical structure of a Bayesian network has the form of a directed acyclic graph (DAG), which has the advantage of supporting an interpretation of the graph in terms of cause-effect relationships. However, a limitation is that only asymmetric relationships, such as cause and effect relationships, can be modeled between variables in a DAG. Chain graphs, which admit both directed and undirected edges, can be used to overcome this limitation. Today there exist three main different interpretations of chain graphs in the lit- erature. These are the Lauritzen-Wermuth-Frydenberg, the Andersson-Madigan-Perlman, and the multivariate regression interpretations. In this thesis, we study these interpreta- tions based on their separation criteria and the intuition behind their edges. Since structure learning is a critical component in constructing an intelligent system based on a chain graph model, we propose new feasible and efficient structure learning algorithms to learn chain graphs from data under the faithfulness assumption. The proliferation of different PGMs that allow factorizations of different kinds leads us to consider a more general graphical structure in this thesis, namely directed acyclic iv hypergraphs. Directed acyclic hypergraphs are the graphical structure of a new proba- bilistic graphical model that we call Bayesian hypergraphs. Since there are many more hypergraphs than DAGs, undirected graphs, chain graphs, and, indeed, other graph-based networks, Bayesian hypergraphs can model much finer factorizations and thus are more computationally efficient. Bayesian hypergraphs also allow a modeler to represent causal patterns of interaction such as Noisy-OR graphically (without additional annotations). We introduce global, local and pairwise Markov properties of Bayesian hypergraphs and prove under which conditions they are equivalent. We also extend the causal interpretation of LWF chain graphs to Bayesian hypergraphs and provide corresponding formulas and a graphical criterion for intervention. The framework of graphical models, which provides algorithms for discovering and analyzing structure in complex distributions to describe them succinctly and extract un- structured information, allows them to be constructed and utilized effectively. Two of the most important applications of graphical models are causal inference and information ex- traction. To address these abilities of graphical models, we conduct a causal analysis, comparing the performance behavior of highly-configurable systems across environmen- tal conditions (changing workload, hardware, and software versions), to explore when and how causal knowledge can be commonly exploited for performance analysis. v Table of Contents Dedication .................................... iii Abstract ..................................... iv List of Tables .................................. ix List of Figures .................................. xi Chapter 1 Introduction ............................. 1 Chapter 2 LWF Chain Graphs ......................... 8 2.1 Basic Definitions and Concepts . 9 2.2 On the Properties of LWF Chain Graphs . 11 2.3 Finding Minimal Separators in LWF Chain Graphs . 13 2.4 Efficient Learning of LWF Chain Graphs under the Faithfulness Assumption 21 Chapter 3 MVR Chain Graphs ......................... 42 3.1 Basic Definitions and Concepts . 43 3.2 On the Properties of MVR Chain Graphs . 48 3.3 Finding Minimal Separators in MVR Chain Graphs . 65 3.4 Order-Independent Structure Learning of Multivariate Regression Chain Graphs . 74 vi 3.5 A Decomposition-Based Algorithm for Structure Learning of MVR Chain Graphs . 93 3.6 Discussion and Conclusion . 117 Chapter 4 AMP Chain Graphs . 120 4.1 Basic Definitions and Concepts . 121 4.2 On the Properties of AMP Chain Graphs . 124 4.3 Finding Minimal Separators in AMP Chain Graphs . 126 4.4 Order-Independent Structure Learning of AMP Chain Graphs . 134 4.5 A Decomposition-Based Algorithm for Learning the Structure of AMP CGs 144 4.6 Discussion and Conclusion . 161 Chapter 5 Bayesian Hypergraphs . 166 5.1 Basic Definitions and Concepts . 168 5.2 Markov properties for undirected graphs . 177 5.3 Bayesian Hypergraphs . 179 5.4 Intervention in Bayesian hypergraphs . 193 Chapter 6 Causal Transfer Learning . 205 6.1 Introduction . 205 6.2 Causal Graphs . 208 6.3 Research Questions and Methodology . 210 6.4 Identification of Causal Effects (RQ1) . 212 6.5 Transportability of Causal and Statistical Relations Across Environ- ments (RQ2) . 215 vii 6.6 Generalizing Statistical Findings Across Sampling Conditions (RQ3) . 217 6.7 Threats to Validity . 218 Chapter 7 Concluding Remarks . 220 Bibliography ................................... 225 Appendix AProofs of Correctness of the Theorems and Algorithms in Sec- tion 2.4 . 239 Appendix BProof of Theorem 3.10 . 243 Appendix CProofs of Theoretical Results in Section 3.5 . 247 Appendix DProofs of Correctness of Algorithms 11 and 12 . 253 Appendix EProofs of Correctness of the Algorithms in Sections 4.4 and 4.5 . 255 viii List of Tables Table 2.1 The trace table of Algorithm 4 for i = 3 and order1(V) = (d; e; a; c; b). 29 Table 2.2 The trace table of Algorithm 4 for i = 3 and order2(V) = (d; c; e; a; b). 29 Table 2.3 Results for discrete samples from the ASIA, INSURANCE, ALARM, and HAILFINDER networks respectively. Each row corresponds to the significance level: α = 0:05=0:01=0:005 respectively. 41 Table 3.1 The trace table of Algorithm 9 for i = 3 and order1(V) = (d; e; a; c; b). 79 Table 3.2 The trace table of Algorithm 9 for i = 3 and order2(V) = (d; c; e; a; b). 79 Table 3.3 Order-dependence issues and corresponding modifications of the PC- like algorithm that remove the problem. “Yes" indicates that the cor- responding aspect of the graph is estimated order-independently in the sample version. 88 Table 3.4 Results for discrete samples from the ASIA network. Each row cor- responds to the significance level: α = 0:05=0:01=0:005 respectively. 115 Table 3.5 Results for discrete samples from the INSURANCE network. Each row corresponds to the significance level: α = 0:05=0:01=0:005 re- spectively. 116 Table 3.6 Results for discrete samples from the ALARM network. Each row corresponds to the significance level: α = 0:05=0:01=0:005 respectively. 116 Table 3.7 Results for discrete samples from the HAILFINDER network. Each row corresponds to the significance level: α = 0:05=0:01=0:005 re- spectively. 117 Table 4.1 The trace table of Algorithm 17 for i = 1 and order1(V) = (d; c; b; a; e). For simplicity, we define ADJ (u):= [ad (u) ad (ad (u))] u; v . 138 H H [ H H n f g Table 4.2 The trace table of Algorithm 17 for i = 1 and order2(V) = (d; e; a; c; b). For simplicity, we define ADJ (u):= [ad (u) ad (ad (u))] u; v . 138 H H [ H H n f g ix Table 4.3 The trace table of Algorithm 18 for i = 1, order1(V) = (d; c; b; a; e), and order2(V) = (d; e; a; c; b). For simplicity, we define ADJH(u):= [ad (u) ad (ad (u))] u; v . 142 H [ H H n f g Table 4.4 Order-dependence issues and corresponding modifications of the PC- like algorithm that remove the problem. “Yes" indicates that the cor- responding aspect of the graph is estimated order-independently in the sample version. 144 Table 4.5 Results for discrete samples from the ASIA, INSURANCE, ALARM, and HAILFINDER networks respectively. Each row corresponds to the significance level: α = 0:05=0:01=0:005 respectively. 162 Table 5.1 The conditional probability distribution P(C A; B) for a model with: j (a) noisy functional dependence (Noisy-OR), and therefore P(C = n A = y; B = y) = P(C = n A = y)P(C = n B = y); (b) non-noisy j j j functional dependence. 167 Table 5.2 Factorizations and corresponding BH representations . 185 Table 6.1 Overview of the real-world subject systems . 213 Table 7.1 Properties of chain graphs under different interpretations .
Recommended publications
  • Applications of DFS
    (b) acyclic (tree) iff a DFS yeilds no Back edges 2.A directed graph is acyclic iff a DFS yields no back edges, i.e., DAG (directed acyclic graph) , no back edges 3. Topological sort of a DAG { next 4. Connected components of a undirected graph (see Homework 6) 5. Strongly connected components of a drected graph (see Sec.22.5 of [CLRS,3rd ed.]) Applications of DFS 1. For a undirected graph, (a) a DFS produces only Tree and Back edges 1 / 7 2.A directed graph is acyclic iff a DFS yields no back edges, i.e., DAG (directed acyclic graph) , no back edges 3. Topological sort of a DAG { next 4. Connected components of a undirected graph (see Homework 6) 5. Strongly connected components of a drected graph (see Sec.22.5 of [CLRS,3rd ed.]) Applications of DFS 1. For a undirected graph, (a) a DFS produces only Tree and Back edges (b) acyclic (tree) iff a DFS yeilds no Back edges 1 / 7 3. Topological sort of a DAG { next 4. Connected components of a undirected graph (see Homework 6) 5. Strongly connected components of a drected graph (see Sec.22.5 of [CLRS,3rd ed.]) Applications of DFS 1. For a undirected graph, (a) a DFS produces only Tree and Back edges (b) acyclic (tree) iff a DFS yeilds no Back edges 2.A directed graph is acyclic iff a DFS yields no back edges, i.e., DAG (directed acyclic graph) , no back edges 1 / 7 4. Connected components of a undirected graph (see Homework 6) 5.
    [Show full text]
  • 9 the Graph Data Model
    CHAPTER 9 ✦ ✦ ✦ ✦ The Graph Data Model A graph is, in a sense, nothing more than a binary relation. However, it has a powerful visualization as a set of points (called nodes) connected by lines (called edges) or by arrows (called arcs). In this regard, the graph is a generalization of the tree data model that we studied in Chapter 5. Like trees, graphs come in several forms: directed/undirected, and labeled/unlabeled. Also like trees, graphs are useful in a wide spectrum of problems such as com- puting distances, finding circularities in relationships, and determining connectiv- ities. We have already seen graphs used to represent the structure of programs in Chapter 2. Graphs were used in Chapter 7 to represent binary relations and to illustrate certain properties of relations, like commutativity. We shall see graphs used to represent automata in Chapter 10 and to represent electronic circuits in Chapter 13. Several other important applications of graphs are discussed in this chapter. ✦ ✦ ✦ ✦ 9.1 What This Chapter Is About The main topics of this chapter are ✦ The definitions concerning directed and undirected graphs (Sections 9.2 and 9.10). ✦ The two principal data structures for representing graphs: adjacency lists and adjacency matrices (Section 9.3). ✦ An algorithm and data structure for finding the connected components of an undirected graph (Section 9.4). ✦ A technique for finding minimal spanning trees (Section 9.5). ✦ A useful technique for exploring graphs, called “depth-first search” (Section 9.6). 451 452 THE GRAPH DATA MODEL ✦ Applications of depth-first search to test whether a directed graph has a cycle, to find a topological order for acyclic graphs, and to determine whether there is a path from one node to another (Section 9.7).
    [Show full text]
  • A Low Complexity Topological Sorting Algorithm for Directed Acyclic Graph
    International Journal of Machine Learning and Computing, Vol. 4, No. 2, April 2014 A Low Complexity Topological Sorting Algorithm for Directed Acyclic Graph Renkun Liu then return error (graph has at Abstract—In a Directed Acyclic Graph (DAG), vertex A and least one cycle) vertex B are connected by a directed edge AB which shows that else return L (a topologically A comes before B in the ordering. In this way, we can find a sorted order) sorting algorithm totally different from Kahn or DFS algorithms, the directed edge already tells us the order of the Because we need to check every vertex and every edge for nodes, so that it can simply be sorted by re-ordering the nodes the “start nodes”, then sorting will check everything over according to the edges from a Direction Matrix. No vertex is again, so the complexity is O(E+V). specifically chosen, which makes the complexity to be O*E. An alternative algorithm for topological sorting is based on Then we can get an algorithm that has much lower complexity than Kahn and DFS. At last, the implement of the algorithm by Depth-First Search (DFS) [2]. For this algorithm, edges point matlab script will be shown in the appendix part. in the opposite direction as the previous algorithm. The algorithm loops through each node of the graph, in an Index Terms—DAG, algorithm, complexity, matlab. arbitrary order, initiating a depth-first search that terminates when it hits any node that has already been visited since the beginning of the topological sort: I.
    [Show full text]
  • CS302 Final Exam, December 5, 2016 - James S
    CS302 Final Exam, December 5, 2016 - James S. Plank Question 1 For each of the following algorithms/activities, tell me its running time with big-O notation. Use the answer sheet, and simply circle the correct running time. If n is unspecified, assume the following: If a vector or string is involved, assume that n is the number of elements. If a graph is involved, assume that n is the number of nodes. If the number of edges is not specified, then assume that the graph has O(n2) edges. A: Sorting a vector of uniformly distributed random numbers with bucket sort. B: In a graph with exactly one cycle, determining if a given node is on the cycle, or not on the cycle. C: Determining the connected components of an undirected graph. D: Sorting a vector of uniformly distributed random numbers with insertion sort. E: Finding a minimum spanning tree of a graph using Prim's algorithm. F: Sorting a vector of uniformly distributed random numbers with quicksort (average case). G: Calculating Fib(n) using dynamic programming. H: Performing a topological sort on a directed acyclic graph. I: Finding a minimum spanning tree of a graph using Kruskal's algorithm. J: Finding the minimum cut of a graph, after you have found the network flow. K: Finding the first augmenting path in the Edmonds Karp implementation of network flow. L: Processing the residual graph in the Ford Fulkerson algorithm, once you have found an augmenting path. Question 2 Please answer the following statements as true or false. A: Kruskal's algorithm requires a starting node.
    [Show full text]
  • Merkle Trees and Directed Acyclic Graphs (DAG) As We've Discussed, the Decentralized Web Depends on Linked Data Structures
    Merkle trees and Directed Acyclic Graphs (DAG) As we've discussed, the decentralized web depends on linked data structures. Let's explore what those look like. Merkle trees A Merkle tree (or simple "hash tree") is a data structure in which every node is hashed. +--------+ | | +---------+ root +---------+ | | | | | +----+---+ | | | | +----v-----+ +-----v----+ +-----v----+ | | | | | | | node A | | node B | | node C | | | | | | | +----------+ +-----+----+ +-----+----+ | | +-----v----+ +-----v----+ | | | | | node D | | node E +-------+ | | | | | +----------+ +-----+----+ | | | +-----v----+ +----v-----+ | | | | | node F | | node G | | | | | +----------+ +----------+ In a Merkle tree, nodes point to other nodes by their content addresses (hashes). (Remember, when we run data through a cryptographic hash, we get back a "hash" or "content address" that we can think of as a link, so a Merkle tree is a collection of linked nodes.) As previously discussed, all content addresses are unique to the data they represent. In the graph above, node E contains a reference to the hash for node F and node G. This means that the content address (hash) of node E is unique to a node containing those addresses. Getting lost? Let's imagine this as a set of directories, or file folders. If we run directory E through our hashing algorithm while it contains subdirectories F and G, the content-derived hash we get back will include references to those two directories. If we remove directory G, it's like Grace removing that whisker from her kitten photo. Directory E doesn't have the same contents anymore, so it gets a new hash. As the tree above is built, the final content address (hash) of the root node is unique to a tree that contains every node all the way down this tree.
    [Show full text]
  • Graph Algorithms G
    Bipartiteness Graph G = (V,E) is bipartite iff it can be partitioned into two sets of nodes A and B such that each edge has one end in A and the other Introduction to Algorithms end in B Alternatively: Graph Algorithms • Graph G = (V,E) is bipartite iff all its cycles have even length • Graph G = (V,E) is bipartite iff nodes can be coloured using two colours CSE 680 Question: given a graph G, how to test if the graph is bipartite? Prof. Roger Crawfis Note: graphs without cycles (trees) are bipartite bipartite: non bipartite Testing bipartiteness Toppgological Sort Method: use BFS search tree Recall: BFS is a rooted spanning tree. Algorithm: Want to “sort” or linearize a directed acyygp()clic graph (DAG). • Run BFS search and colour all nodes in odd layers red, others blue • Go through all edges in adjacency list and check if each of them has two different colours at its ends - if so then G is bipartite, otherwise it is not A B D We use the following alternative definitions in the analysis: • Graph G = (V,E) is bipartite iff all its cycles have even length, or • Graph G = (V, E) is bipartite iff it has no odd cycle C E A B C D E bipartit e non bipartite Toppgological Sort Example z PerformedonaPerformed on a DAG. A B D z Linear ordering of the vertices of G such that if 1/ (u, v) ∈ E, then u appears before v. Topological-Sort (G) 1. call DFS(G) to compute finishing times f [v] for all v ∈ V 2.
    [Show full text]
  • 6.042J Chapter 6: Directed Graphs
    “mcs-ftl” — 2010/9/8 — 0:40 — page 189 — #195 6 Directed Graphs 6.1 Definitions So far, we have been working with graphs with undirected edges. A directed edge is an edge where the endpoints are distinguished—one is the head and one is the tail. In particular, a directed edge is specified as an ordered pair of vertices u, v and is denoted by .u; v/ or u v. In this case, u is the tail of the edge and v is the ! head. For example, see Figure 6.1. A graph with directed edges is called a directed graph or digraph. Definition 6.1.1. A directed graph G .V; E/ consists of a nonempty set of D nodes V and a set of directed edges E. Each edge e of E is specified by an ordered pair of vertices u; v V . A directed graph is simple if it has no loops (that is, edges 2 of the form u u) and no multiple edges. ! Since we will focus on the case of simple directed graphs in this chapter, we will generally omit the word simple when referring to them. Note that such a graph can contain an edge u v as well as the edge v u since these are different edges ! ! (for example, they have a different tail). Directed graphs arise in applications where the relationship represented by an edge is 1-way or asymmetric. Examples include: a 1-way street, one person likes another but the feeling is not necessarily reciprocated, a communication channel such as a cable modem that has more capacity for downloading than uploading, one entity is larger than another, and one job needs to be completed before another job can begin.
    [Show full text]
  • Graph Algorithms
    Graph Algorithms Graphs and Graph Repre- sentations Graph Graph Algorithms Traversals COMP4128 Programming Challenges Directed Graphs and Cycles Strongly Connected Components Example: School of Computer Science and Engineering 2SAT UNSW Australia Minimum Spanning Trees Table of Contents 2 Graph Algorithms 1 Graphs and Graph Representations Graphs and Graph Repre- sentations 2 Graph Graph Traversals Traversals Directed Graphs and 3 Directed Graphs and Cycles Cycles Strongly Connected 4 Strongly Connected Components Components Example: 2SAT 5 Example: 2SAT Minimum Spanning Trees 6 Minimum Spanning Trees Graphs 3 Graph Algorithms Graphs and Graph Repre- sentations A graph is a collection of vertices and edges connecting Graph pairs of vertices. Traversals Directed Generally, graphs can be thought of as abstract Graphs and Cycles representations of objects and connections between those Strongly objects, e.g. intersections and roads, people and Connected Components friendships, variables and equations. Example: Many unexpected problems can be solved with graph 2SAT techniques. Minimum Spanning Trees Graphs 4 Graph Algorithms Graphs and Graph Repre- sentations Graph Many different types of graphs Traversals Directed graphs Directed Graphs and Acyclic graphs, including trees Cycles Weighted graphs Strongly Connected Flow graphs Components Other labels for the vertices and edges Example: 2SAT Minimum Spanning Trees Graph Representations 5 Graph Algorithms Graphs and Graph Repre- sentations Graph You’ll usually want to either use an adjacency list or an Traversals adjacency matrix to store your graph. Directed Graphs and An adjacency matrix is just a table (usually implemented Cycles as an array) where the jth entry in the ith row represents Strongly Connected the edge from i to j, or lack thereof.
    [Show full text]
  • The Shortest Path Problem with Edge Information Reuse Is NP-Complete
    The Shortest Path Problem with Edge Information Reuse is NP-Complete Jesper Larsson Tr¨aff TU Wien Faculty of Informatics, Institute of Information Systems Research Group Parallel Computing Favoritenstrasse 16/184-5, 1040 Vienna, Austria email: [email protected] 15.9.2015, Revised 6.6.2016 Abstract We show that the following variation of the single-source shortest path problem is NP- complete. Let a weighted, directed, acyclic graph G = (V,E,w) with source and sink vertices s and t be given. Let in addition a mapping f on E be given that associates information with the edges (e.g., a pointer), such that f(e)= f(e′) means that edges e and e′ carry the same information; for such edges it is required that w(e)= w(e′). The length of a simple st path U is the sum of the weights of the edges on U but edges with f(e)= f(e′) are counted only once. The problem is to determine a shortest such st path. We call this problem the edge information reuse shortest path problem. It is NP-complete by reduction from 3SAT. 1 The Edge Information Reuse Shortest Path Problem A weighted, directed, acyclic graph G = (V, E, w) with source and sink vertices s,t ∈ V is given. Edges represent some possible substructures and a path from s to t determines how substructures are put together to form a desired superstructure. Edge weights reflect the cost of the associated substructures (e.g., memory consumption). An ordinary shortest path determines an ordered, tree-like representation of the superstructure of least cost: a single root with children given by the edges of the path.
    [Show full text]
  • 5 Directed Acyclic Graphs
    5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate such knowledge into a graphical model for the variables. Let XV = (Xv)v∈V be a vector of real-valued random variables with probability distribution P and density p. Then the density p can always be decomposed into a product of conditional densities, d Q p(x) = p(xd|x1, . , xd−1)p(x1, . , xd−1) = ... = p(xv|x1, . , xv−1). (5.1) v=1 Note that this can be achieved for any ordering of the variables. Now suppose that the conditional density of some variable Xv does not depend on all its predecessors, namely X1,...,Xv−1, but only on a subset XUv , that is, Xv is conditionally inde- pendent of its predecessors given XUv . Substituting p(xv|xUv ) for p(xv|x1, . , xv−1) in the product (5.1), we obtain d Q p(x) = p(xv|xUv ). (5.2) v=1 This recursive dependence structure can be represented by a directed graph G by drawing an arrow from each vertex in Uv to v. As an immediate consequence of the recursive factorization, the resulting graph is acyclic, that is, it does not contain any loops. On the other hand, P factorizes with respect to the undirected graph Gm which is given by the class of complete subsets D = {{v} ∪ Uv|v ∈ V }. This graph can be obtained from the directed graph G by completing all sets {v} ∪ Uv and then converting all directed edges into undirected ones.
    [Show full text]
  • Single-Source Shortest Paths (Ch
    Single-Source Shortest Paths (Ch. 24) Single-Source Shortest Paths The single-source shortest path problem (SSSP) Weights in SSSP algorithms include distances, times, hops, input: a directed graph G = (V, E) with edge weights, and a cost, etc. Used in lots of routing applications. specific source node s. goal: find a minimum weight (shortest) path from s to every other Note: BFS finds the shortest paths for the special case when all node in V (sometimes we only want the cost of these paths) edge weights are 1. Running time = O(V + E) It turns out that, in the worst case, finding the shortest path between s and a single The result of a SSSP algorithm can be viewed as a tree node t is no easier than finding the shortest paths between s and all other nodes rooted at s, containing a shortest (wrt weight) path from s à g = ¥ s to all other nodes. s à b = 2 2 3 s à c = 5 s à h = 7 a b c 2 3 5 6 a b 6 c s à d = 6 S 5 3 d 1 e S 3 s à e = 5 2 d 1 e 2 s à f = 3 f 4 g 2 h f 4 g 2 h 2 Negative-weight cycles Single-Source Shortest Paths Some graphs may have negative-weight cycles and these are Question: Can a shortest path contain a cycle? a problem for SSSP algorithms. What is the shortest path from a to d? No. b 9 · path a-c-e-d = -12+3-2 = -11 2 8 6 Suppose we have a shortest path p = áv0, v1, ..., vkñ and S -12 3 c = áv , v , ..., v ñ is a positive-weight cycle on p so that v = v · path a-c-e-d-a-c-e-d = a c e i i+1 j i j -12+3-2+(10-12+3-2)= -13 6 4 and w(c) > 0.
    [Show full text]
  • Algorithms (Winter 2021) Ungraded Problem Set 5 Due: Not for Submission Topic: Graphs: Basics + DFS
    CS 31: Algorithms (Winter 2021) Ungraded Problem Set 5 Due: Not for submission Topic: Graphs: Basics + DFS Problem 1 (Some operations with graphs). For the parts below, describe your algorithm summary in English and also provide a clear pseudocode. You may assume the graph G = (V; E) is input as adjacency lists. a. In this part the graph G is directed. You need to return the reverse graph Grev where all edges have been reversed. Grev should also be represented as adjacency lists. What is the runtime of your algo- rithm? K b. Recall the definition of a walk from vertex u to vertex v and a path from vertex u to v. Design an algorithm which takes input a walk from u to v, and returns a path from u to v. K Problem 2 (DFS firsts and lasts). For each of the following assertions, prove or give a counterexample. The graph G, unless mentioned, is directed. When you give a counter-example, you should explain why it is a counter-example. (a) If G contains a path from u to v and if first[u] < first[v] in a depth-first search of G, then last[v] < last[u]. K (b) If (u=v1; v2; : : : ; vk =v) is a path from u to v in G such that last[vi] < last[v] for all 1 ≤ i < k, then first[v] < first[vi] for all 1 ≤ i < k. K (c) If we run depth-first search on G in an arbitrary order to obtain first’s and last’s, and let σ be the increasing order of last[v]’s.
    [Show full text]