6.006 Course Notes

Total Page:16

File Type:pdf, Size:1020Kb

6.006 Course Notes 6.006 Course Notes Wanlin Li Fall 2018 1 September 6 • Computation • Definition. Problem: matching inputs to correct outputs • Definition. Algorithm: procedure/function matching each input to single out- put, correct if matches problem • Want program to be able to handle infinite space of inputs, arbitrarily large size of input • Want finite and efficient program, need recursive code • Efficiency determined by asymptotic growth Θ() • O(f(n)) is all functions below order of growth of f(n) • Ω(f(n)) is all functions above order of growth of f(n) • Θ(f(n)) is tight bound • Definition. Log-linear: Θ(n log n) Θ(nc) • Definition. Exponential: b • Algorithm generally considered efficient if runs in polynomial time • Model of computation: computer has memory (RAM - random access memory) and CPU for performing computation; what operations are allowed in constant time • CPU has registers to access and process memory address, needs log n bits to handle size n input • Memory chunked into “words” of size at least log n 1 6.006 Wanlin Li 32 10 • E.g. 32 bit machine can handle 2 ∼ 4 GB, 64 bit machine can handle 10 GB • Model of computaiton in word-RAM • Ways to solve algorithms problem: design new algorithm or reduce to already known algorithm (particularly search in data structure, sort, shortest path) • Classification of algorithms based on graph on function calls • Class examples and design paradigms: brute force (completely disconnected), decrease and conquer (path), divide and conquer (tree with branches), dynamic programming (directed acyclic graph), greedy (choose which paths on directed acyclic graph to take) • Computation doesn’t have to be a tree, just has to be directed acyclic graph 2 September 11 • Example. Check if array contains item 1 def contains(A,v): 2 fori in range(len(A)): 3 if A[i] == v: 4 return True 5 return False • Other option to call function recursively using additional parameter, decrease and conquer; contains(A, v, i = 0) • Divide and conquer would check if middle element is desired element, otherwise return if element is in either half of array; still O(n) • Maximum subarray sum problem: what is the largest sum of any nonempty (contiguous) subarray? • Example. Brute force 1 def MSS(A): 2 m = A[0] 3 forj in range(1,len(A)+1): 4 fori in range(0,j): 5 m = max(m, SS(A,i,j)) 6 returnm where SS is helper function for summing subarray; O(n3) because there are O(n2) nodes and each requires linear time • Divide and conquer approach: MSS fully in left half of array, fully in right half of array, or contains middle two elements 2 6.006 Wanlin Li 1 def MSS(A, i=0, j=None): 2 ifj is None: 3 j = len(A) 4 if j-i == 1: 5 return A[i] 6 c = (i+j)//2 7 return max(MSS(A,i,c), MSS(A,c,j), MSS_EA(A,i,c) + MSS_SA(A,c,j)) 8 def MSS_SA(A,i,j): 9 s = m = A[i] 10 fork in range(1, j-i): 11 s += A[i+k] 12 m = max(s,m) 13 returnm 14 def MSS_EA(A,i,j): 15 s = m = A[j-1] 16 fork in range(1,j-i): 17 s += A[j-1-k] 18 m = max(s,m) 19 returnm n which is O(n log n); O(n) from subarray in both halves of array, O 2 from each half of array ! recursion tree has log n levels, each requiring O(n) work, so total runtime is O(n log n) • As dynamic programming, 1 def MSS(A): 2 m = mss_ea = A[0] 3 fori in range(1,len(A)): 4 mss_ea = A[i] + max(0, mss_ea) 5 m = max(m,mss_ea) 6 returnm 3 September 13 • Sorting problem: input is array, output is permutation of the array in sorted order • In place sorting: B = A and O(1) extra space (no second array is created, A is modified); list.sort is in place, sorted is not in place sorting • Permutation sort: check if each permutation B of A is sorted, takes Θ(n · n!) time (n! permutations and Θ(n) time to check) • Insertion sort: decrease and conquer, in-place, recursively sort all but one item; sort first n − 1 elements, then insert A[n − 1] in the correct place; shifting ele- ments of array takes linear time so overall time is O(n2) • Example. Bottom-up implementation 3 6.006 Wanlin Li 1 fori in range(1,n): 2 while A[i] < A[i-1]: 3 A[i], A[i-1] = A[i-1], A[i] 4 i -= 1 5 returnA • T (n): running time of algorithm on input of size n but nature of input also mat- ters • T (n) is the worst case running time on input of size n; worst case bound gives guarantee on running time 2 2 • Insertion sort is O(n ) due to two nested loops of ≤ n iterations, Ω(n ) in reverse sorted case (just need a bad case for lower bound), so T (n) = Θ(n2) • Divide and conquer algorithm: if size n input divided into a subproblems of size n n b ;T (n) = aT b + f(n) where f(n) is the time needed to divide and combine • n n Merge sort: recursively sort A[: 2 ] and A[ 2 :]; then merge left and right halves into output by comparing current elements of halves • Example. 1 def merge_sort(A, i=0, j=None): 2 ifj is None: j = len(A) 3 if j - i == 1: return [A[i]] 4 m = (i + j)//2 5 L = merge_sort(A[:m]) 6 R = merge_sort(A[m:]) 7 return list(merge(L,R)) 8 def merge(L,R): 9 l = r = 0 10 whilel< len(L) orr< len(R): 11 ifl< len(L) and (r >= len(R) or L[l] <= R[r]): 12 yield L[l] 13 l += 1 14 else: 15 yield R[r] 16 r += 1 • n T (n) = 2T 2 + Θ(n) commonly called merge sort recurrence 4 September 18 • Sequence interface and data structures (linked lists, dynamic arrays; amortiza- tion), set interface • Definition. Interface (API/ADT): specification, what data can be maintained, supported operations and meaning, problem 4 6.006 Wanlin Li • Definition. Data structure: implementation, how data gets stored, algorithm implementing operations, solution • Python list uses dynamic array • Static sequence: problem of maintaining sequence of objects/items x0; x1; : : : xn−1; subject to len(), sequence iteration, at(i), set-at(i,x) • Ideal solution to static sequence is static array, e.g. 1 def allocate(n):#pretend static array 2 return [None] * n 3 class StaticArray (ArraySequence): 4 def __init__(self,n): 5 self.n = n 6 self.array = allocate(n) 7 def len(self): return self.n 8 def at(self, i): return self.array[i] 9 def set_at(self, i, x): self.array[i] = x • Special cases of left() (at(0)) and right() (at(n − 1)) • Static sequence is Θ(1) for everything except iteration • Dynamic sequence: also includes insert-at(i,x) forcing all the subsequent items to shift back one; delete-at(i), insert/delete at left/right sides • Special cases: stack (insert/delete right), queue (insert right or delete left), deque (insert/delete left/right) • In static array, attempting to do dynamic process requires Θ(n) time because it allocates a new array and shifts the items • Memory allocation: RAM can be viewed as large array, can allocate and initialize array of n words in Θ(n) time; ensures time is Ω(space) • Ways to overcome problem of copying over information in static array: linked lists or dynamic arrays • Linked lists allocate O(1) size arrays instead of contiguous block, use pointers; allocate structure of size 3 for each item, contains previous pointer, item, next pointer; constant time for each insertion and deletion but indexing takes Θ(n) time • Dynamic arrays: still must allocate array of size n words in Θ(n) time, but in- stead of resizing array to exactly size n maintain an invariant of size Θ(n) ≥ n; when inserting and size = n; want to increase size to 2n; when deleting and size n is 4 ; halve the size (here n is size of array) 5 6.006 Wanlin Li • n Python list doesn’t necessarily use 4 and 2n; any constant coefficients a; b work as long as a < 1 < b • Worst-case insert right is still Θ(n); but very few inserts actually require the expensive call; resize only when reach n is a power of 2; total cost for n = 2k insert rights from empty sequence is Θ(1+2+···+2k) = Θ(n); amortized time is constant per operation • In algorithms, care about worst case; in data structures, care more about amor- tization • Definition. Amortized analysis: only meaningful for data structures; operation has amortized cost T (n) if when k operations are performed, they cost at most kT (n) time • Example. Insert right requires constant amortized time • Insertion and deletion in the middle is still expensive • Sequence cares about index of item, set cares about key of item • Static set: find-key(k), basic iteration (can come out in any order) • Dynamic set: includes insert(item), delete(key); static set and dynamic set are basically Python set/dictionary, solved by hashing • Ordered set: find-next(key), find-previous(key), special cases of finding min/max • Dynamic ordered set: delete min/max 5 September 20 • Sequence is like a checkout line (where objects are), set is like a raffle (what the object is) • Priority queue: optimized for finding most important item; can find length, can insert at any time, can find max or find and remove max • Priority queues good for sorting; given array A; insert all elements into priority queue and repeatedly remove max or remove min • Multi-set API can have repeated elements (generally want this behavior in pri- ority queue) • Priority queue: elements stored in array without caring about order (can be dynamic array), insert at end (Q.append) in O(1) (possibly amortized), swap max with end of array and pop last element i.e.
Recommended publications
  • Vertex Ordering Problems in Directed Graph Streams∗
    Vertex Ordering Problems in Directed Graph Streams∗ Amit Chakrabartiy Prantar Ghoshz Andrew McGregorx Sofya Vorotnikova{ Abstract constant-pass algorithms for directed reachability [12]. We consider directed graph algorithms in a streaming setting, This is rather unfortunate given that many of the mas- focusing on problems concerning orderings of the vertices. sive graphs often mentioned in the context of motivating This includes such fundamental problems as topological work on graph streaming are directed, e.g., hyperlinks, sorting and acyclicity testing. We also study the related citations, and Twitter \follows" all correspond to di- problems of finding a minimum feedback arc set (edges rected edges. whose removal yields an acyclic graph), and finding a sink In this paper we consider the complexity of a variety vertex. We are interested in both adversarially-ordered and of fundamental problems related to vertex ordering in randomly-ordered streams. For arbitrary input graphs with directed graphs. For example, one basic problem that 1 edges ordered adversarially, we show that most of these motivated much of this work is as follows: given a problems have high space complexity, precluding sublinear- stream consisting of edges of an acyclic graph in an space solutions. Some lower bounds also apply when the arbitrary order, how much memory is required to return stream is randomly ordered: e.g., in our most technical result a topological ordering of the graph? In the offline we show that testing acyclicity in the p-pass random-order setting, this can be computed in O(m + n) time using model requires roughly n1+1=p space.
    [Show full text]
  • Algorithms (Decrease-And-Conquer)
    Algorithms (Decrease-and-Conquer) Pramod Ganapathi Department of Computer Science State University of New York at Stony Brook August 22, 2021 1 Contents Concepts 1. Decrease by constant Topological sorting 2. Decrease by constant factor Lighter ball Josephus problem 3. Variable size decrease Selection problem ............................................................... Problems Stooge sort 2 Contributors Ayush Sharma 3 Decrease-and-conquer Problem(n) [Step 1. Decrease] Subproblem(n0) [Step 2. Conquer] Subsolution [Step 3. Combine] Solution 4 Types of decrease-and-conquer Decrease by constant. n0 = n − c for some constant c Decrease by constant factor. n0 = n/c for some constant c Variable size decrease. n0 = n − c for some variable c 5 Decrease by constant Size of instance is reduced by the same constant in each iteration of the algorithm Decrease by 1 is common Examples: Array sum Array search Find maximum/minimum element Integer product Exponentiation Topological sorting 6 Decrease by constant factor Size of instance is reduced by the same constant factor in each iteration of the algorithm Decrease by factor 2 is common Examples: Binary search Search/insert/delete in balanced search tree Fake coin problem Josephus problem 7 Variable size decrease Size of instance is reduced by a variable in each iteration of the algorithm Examples: Selection problem Quicksort Search/insert/delete in binary search tree Interpolation search 8 Decrease by Constant HOME 9 Topological sorting Problem Topological sorting of vertices of a directed acyclic graph is an ordering of the vertices v1, v2, . , vn in such a way that there is an edge directed towards vertex vj from vertex vi, then vi comes before vj.
    [Show full text]
  • Topological Sort (An Application of DFS)
    Topological Sort (an application of DFS) CSC263 Tutorial 9 Topological sort • We have a set of tasks and a set of dependencies (precedence constraints) of form “task A must be done before task B” • Topological sort : An ordering of the tasks that conforms with the given dependencies • Goal : Find a topological sort of the tasks or decide that there is no such ordering Examples • Scheduling : When scheduling task graphs in distributed systems, usually we first need to sort the tasks topologically ...and then assign them to resources (the most efficient scheduling is an NP-complete problem) • Or during compilation to order modules/libraries d c a g f b e Examples • Resolving dependencies : apt-get uses topological sorting to obtain the admissible sequence in which a set of Debian packages can be installed/removed Topological sort more formally • Suppose that in a directed graph G = (V, E) vertices V represent tasks, and each edge ( u, v)∊E means that task u must be done before task v • What is an ordering of vertices 1, ..., | V| such that for every edge ( u, v), u appears before v in the ordering? • Such an ordering is called a topological sort of G • Note: there can be multiple topological sorts of G Topological sort more formally • Is it possible to execute all the tasks in G in an order that respects all the precedence requirements given by the graph edges? • The answer is " yes " if and only if the directed graph G has no cycle ! (otherwise we have a deadlock ) • Such a G is called a Directed Acyclic Graph, or just a DAG Algorithm for
    [Show full text]
  • Verified Textbook Algorithms
    Verified Textbook Algorithms A Biased Survey Tobias Nipkow[0000−0003−0730−515X] and Manuel Eberl[0000−0002−4263−6571] and Maximilian P. L. Haslbeck[0000−0003−4306−869X] Technische Universität München Abstract. This article surveys the state of the art of verifying standard textbook algorithms. We focus largely on the classic text by Cormen et al. Both correctness and running time complexity are considered. 1 Introduction Correctness proofs of algorithms are one of the main motivations for computer- based theorem proving. This survey focuses on the verification (which for us always means machine-checked) of textbook algorithms. Their often tricky na- ture means that for the most part they are verified with interactive theorem provers (ITPs). We explicitly cover running time analyses of algorithms, but for reasons of space only those analyses employing ITPs. The rich area of automatic resource analysis (e.g. the work by Jan Hoffmann et al. [103,104]) is out of scope. The following theorem provers appear in our survey and are listed here in alphabetic order: ACL2 [111], Agda [31], Coq [25], HOL4 [181], Isabelle/HOL [151,150], KeY [7], KIV [63], Minlog [21], Mizar [17], Nqthm [34], PVS [157], Why3 [75] (which is primarily automatic). We always indicate which ITP was used in a particular verification, unless it was Isabelle/HOL (which we abbreviate to Isabelle from now on), which remains implicit. Some references to Isabelle formalizations lead into the Archive of Formal Proofs (AFP) [1], an online library of Isabelle proofs. There are a number of algorithm verification frameworks built on top of individual theorem provers.
    [Show full text]
  • Visvesvaraya Technological University a Project Report
    ` VISVESVARAYA TECHNOLOGICAL UNIVERSITY “Jnana Sangama”, Belagavi – 590 018 A PROJECT REPORT ON “PREDICTIVE SCHEDULING OF SORTING ALGORITHMS” Submitted in partial fulfillment for the award of the degree of BACHELOR OF ENGINEERING IN COMPUTER SCIENCE AND ENGINEERING BY RANJIT KUMAR SHA (1NH13CS092) SANDIP SHAH (1NH13CS101) SAURABH RAI (1NH13CS104) GAURAV KUMAR (1NH13CS718) Under the guidance of Ms. Sridevi (Senior Assistant Professor, Dept. of CSE, NHCE) DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING NEW HORIZON COLLEGE OF ENGINEERING (ISO-9001:2000 certified, Accredited by NAAC ‘A’, Permanently affiliated to VTU) Outer Ring Road, Panathur Post, Near Marathalli, Bangalore – 560103 ` NEW HORIZON COLLEGE OF ENGINEERING (ISO-9001:2000 certified, Accredited by NAAC ‘A’ Permanently affiliated to VTU) Outer Ring Road, Panathur Post, Near Marathalli, Bangalore-560 103 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CERTIFICATE Certified that the project work entitled “PREDICTIVE SCHEDULING OF SORTING ALGORITHMS” carried out by RANJIT KUMAR SHA (1NH13CS092), SANDIP SHAH (1NH13CS101), SAURABH RAI (1NH13CS104) and GAURAV KUMAR (1NH13CS718) bonafide students of NEW HORIZON COLLEGE OF ENGINEERING in partial fulfillment for the award of Bachelor of Engineering in Computer Science and Engineering of the Visvesvaraya Technological University, Belagavi during the year 2016-2017. It is certified that all corrections/suggestions indicated for Internal Assessment have been incorporated in the report deposited in the department library. The project report has been approved as it satisfies the academic requirements in respect of Project work prescribed for the Degree. Name & Signature of Guide Name Signature of HOD Signature of Principal (Ms. Sridevi) (Dr. Prashanth C.S.R.) (Dr. Manjunatha) External Viva Name of Examiner Signature with date 1.
    [Show full text]
  • CS 61B Final Exam Guerrilla Section Spring 2018 May 5, 2018
    CS 61B Final Exam Guerrilla Section Spring 2018 May 5, 2018 Instructions Form a small group. Start on the first problem. Check off with a helper or discuss your solution process with another group once everyone understands how to solve the first problem and then repeat for the second problem . You may not move to the next problem until you check off or discuss with another group and everyone understands why the solution is what it is. You may use any course resources at your disposal: the purpose of this review session is to have everyone learning together as a group. 1 The Tortoise or the Hare 1.1 Given an undirected graph G = (V; E) and an edge e = (s; t) in G, describe an O(jV j + jEj) time algorithm to determine whether G has a cycle containing e. Remove e from the graph. Then, run DFS or BFS starting at s to find t. If a path already exists from s to t, then adding e would create a cycle. 1.2 Given a connected, undirected, and weighted graph, describe an algorithm to con- struct a set with as few edges as possible such that if those edges were removed, there would be no cycles in the remaining graph. Additionally, choose edges such that the sum of the weights of the edges you remove is minimized. This algorithm must be as fast as possible. 1. Negate all edges. 2. Form an MST via Kruskal's or Prim's algorithm. 3. Return the set of all edges not in the MST (undo negation).
    [Show full text]
  • Arxiv:1105.2397V1 [Cs.DS] 12 May 2011 [email protected]
    A Incremental Cycle Detection, Topological Ordering, and Strong Component Maintenance BERNHARD HAEUPLER, Massachusetts Institute of Technology TELIKEPALLI KAVITHA, Tata Institute of Fundamental Research ROGERS MATHEW, Indian Institute of Science SIDDHARTHA SEN, Princeton University ROBERT E. TARJAN, Princeton University & HP Laboratories We present two on-line algorithms for maintaining a topological order of a directed n-vertex acyclic graph as arcs are added, and detecting a cycle when one is created. Our first algorithm handles m arc additions in O(m3=2) time. For sparse graphs (m=n = O(1)), this bound improves the best previous bound by a logarithmic factor, and is tight to within a constant factor among algorithms satisfying a natural locality property. Our second algorithm handles an arbitrary sequence of arc additions in O(n5=2) time. For sufficiently dense graphs, this bound improves the best previous boundp by a polynomial factor. Our bound may be far from tight: we show that the algorithm can take Ω(n22 2 lg n) time by relating its performance to a generalization of the k-levels problem of combinatorial geometry. A completely different algorithm running in Θ(n2 log n) time was given recently by Bender, Fineman, and Gilbert. We extend both of our algorithms to the maintenance of strong components, without affecting the asymptotic time bounds. Categories and Subject Descriptors: F.2.2 [Analysis of Algorithms and Problem Complexity]: Nonnumerical Algorithms and Problems—Computations on discrete structures; G.2.2 [Discrete Mathematics]: Graph Theory—Graph algorithms; E.1 [Data]: Data Structures—Graphs and networks General Terms: Algorithms, Theory Additional Key Words and Phrases: Dynamic algorithms, directed graphs, topological order, cycle detection, strong compo- nents, halving intersection, arrangement ACM Reference Format: Haeupler, B., Kavitha, T., Mathew, R., Sen, S., and Tarjan, R.
    [Show full text]
  • A Low Complexity Topological Sorting Algorithm for Directed Acyclic Graph
    International Journal of Machine Learning and Computing, Vol. 4, No. 2, April 2014 A Low Complexity Topological Sorting Algorithm for Directed Acyclic Graph Renkun Liu then return error (graph has at Abstract—In a Directed Acyclic Graph (DAG), vertex A and least one cycle) vertex B are connected by a directed edge AB which shows that else return L (a topologically A comes before B in the ordering. In this way, we can find a sorted order) sorting algorithm totally different from Kahn or DFS algorithms, the directed edge already tells us the order of the Because we need to check every vertex and every edge for nodes, so that it can simply be sorted by re-ordering the nodes the “start nodes”, then sorting will check everything over according to the edges from a Direction Matrix. No vertex is again, so the complexity is O(E+V). specifically chosen, which makes the complexity to be O*E. An alternative algorithm for topological sorting is based on Then we can get an algorithm that has much lower complexity than Kahn and DFS. At last, the implement of the algorithm by Depth-First Search (DFS) [2]. For this algorithm, edges point matlab script will be shown in the appendix part. in the opposite direction as the previous algorithm. The algorithm loops through each node of the graph, in an Index Terms—DAG, algorithm, complexity, matlab. arbitrary order, initiating a depth-first search that terminates when it hits any node that has already been visited since the beginning of the topological sort: I.
    [Show full text]
  • Introduction to Graphs ・V = Nodes
    Undirected graphs Notation. G = (V, E) Introduction to Graphs ・V = nodes. ・E = edges between pairs of nodes. ・Captures pairwise relationship between objects. Tyler Moore ・Graph size parameters: n = | V |, m = | E |. CS 2123, The University of Tulsa V = { 1, 2, 3, 4, 5, 6, 7, 8 } Some slides created by or adapted from Dr. Kevin Wayne. For more information see E = { 1-2, 1-3, 2-3, 2-4, 2-5, 3-5, 3-7, 3-8, 4-5, 5-6, 7-8 } http://www.cs.princeton.edu/~wayne/kleinberg-tardos m = 11, n = 8 3 2 / 36 One week of Enron emails The evolution of FCC lobbying coalitions 4 “The Evolution of FCC Lobbying Coalitions” by Pierre de Vries in JoSS Visualization Symposium 2010 5 3 / 36 4 / 36 The Spread of Obesity in a Large Social Network Over 32 Years educational level; the ego’s obesity status at the ing both their weights. We estimated these mod- previous time point (t); and most pertinent, the els in varied ego–alter pair types. alter’s obesity status at times t and t + 1.25 We To evaluate the possibility that omitted vari- used generalized estimating equations to account ables or unobserved events might explain the as- for multiple observations of the same ego across sociations, we examined how the type or direc- examinations and across ego–alter pairs.26 We tion of the social relationship between the ego assumed an independent working correlation and the alter affected the association between the structure for the clusters.26,27 ego’s obesity and the alter’s obesity.
    [Show full text]
  • CS 270 Algorithms
    CS 270 Algorithms Oliver Week 11 Kullmann Introduction Algorithms and their Revision analysis Graphs Data 1 Introduction structures Conclusion 2 Algorithms and their analysis 3 Graphs 4 Data structures 5 Conclusion CS 270 General remarks Algorithms Oliver Kullmann You need to be able to perform the following: Introduction Algorithms and their Reproducing all definitions. analysis Performing all computations on (arbitrary) examples. Graphs Data Explaining all algorithms in (written) words (possibly using structures some mathematical formalism). Conclusion Stating the time complexity in term of Θ. Check all this by writing it all down! Actually understanding the definitions and algorithms helps a lot (but sometimes “understanding” can become a trap). CS 270 How to prepare Algorithms Oliver Kullmann All information is on the course homepage Introduction Algorithms and their analysis Graphs http://cs.swan.ac.uk/~csoliver/Algorithms201314/index.html Data structures Conclusion Go through all lectures (perhaps you download the slides again — certain additions have been made). Go through the courseworks and their solutions. Go through the lab sessions (run it!). Use the learning goals! CS 270 On the exam Algorithms Oliver Kullmann Introduction Algorithms and their analysis We use the “two-out-of-three” format. Graphs Data Give us a chance to give you marks — write down structures something! Conclusion Actually, better write down a lot — in the second or third round, after answering all questions. Examples are typically helpful. CS 270 The general structure of the module Algorithms Oliver Kullmann Introduction Algorithms and their analysis Graphs Data 1 Weeks 1-3: Algorithms and their analysis structures 2 Weeks 4-5: Graphs — BFS and DFS.
    [Show full text]
  • Topologically Sorting a Directed Acyclic Graph (CLRS 22.4)
    Topologically Sorting a Directed Acyclic Graph (CLRS 22.4) 1 The problem • A topological sorting of a directed acyclic graph G = (V; E) is a linear ordering of vertices V such that (u; v) 2 E ) u appear before v in ordering. Intuitively, we think of an edge (a; b) as meaning that a has to come before b|thus an edge defines a precedence relation. A topological order is an order of the vertices that satisfies all the edges. • Example: Dressing (arrow implies \must come before") Socks Watch Underwear Shoes Shirt Pants Tie Belt Jacket We want to compute order in which to get dressed. One possibility: Socks Underwear Pants Shoes WatchShirt Belt Tie Jacket 1 2 3 4 5 6 7 8 9 The given order is one possible topological order. 1 2 Does a tpopological order always exist? • If the graph has a cycle, a topological order cannot exist. Imagine the simplest cycle, consist- ing of two edges: (a; b) and (b; a). A topological ordering , if it existed, would have to satisfy that a must come before b and b must come before a. This is not possible. • Now consider we have a graph without cycles; this is usually referred to as a DAG (directed acyclic graph). Does any DAG have a topological order? The answer is YES. We'll prove this below indirectly by showing that the toposort algorithm always gives a valid ordering when run on any DAG. 3 Topological sort via DFS It turns out that we can get topological order based on DFS.
    [Show full text]
  • Graphs 1 What Is a Graph? in a Layman’S Terms: a Graph Is a Set of Vertices Or Nodes “ Which Are Connected by Edges
    Graphs 1 What is a Graph? In a layman’s terms: A graph is a set of vertices or nodes “ which are connected by edges. Tree v Graph? “ Tree v Graph? “ “ In Short: Tree is a restricted form of a graph Kinds of Graphs ● Undirected or Directed ● Cyclic or Acyclic ● Weighted or Unweighted ● Simple ● And more! Some additional vocabulary: Adjacent/Neighbors, Degree, Loop. Graph Representation Adjacency Matrix Adjacency List V x V boolean (or numerical) Collection of lists (X) where X[i} matrix where an entry X[a][b] is contains the vertices which are true (or 1) if there exists an edge the neighbors of vertex i. from vertex a to b. SPACE O(V+E)) SPACE O(V^2) BFS / DFS BFS/DFS What is BFS? What is DFS? BFS - Boolean visited” array -> why? - Expand the root and visit all of its children before going to childrens children - So what data structures would be best for this? DFS - Boolean -> why? - Recursion - Go deep into a node - So which is better BFS or DFS? - Lets try implementing before we move on! Tries ● A type of tree - used for characters usually ● Allows for specific quick lookups ● Think of this anytime you have lookup problems (prefix based search, store dictionary, etc) ● Let’s see an example. Insert/Search/Delete ● How would you do each of these? 2 Let’s trie some problems! Q1: Implementation ● Implement either BFS or DFS Q2: Detect a Graph Cycle ● Undirected or Directed ● Cyclic or Acyclic ● Weighted or Unweighted ● Simple ● And more! Some additional vocabulary: Adjacent/Neighbors, Degree, Loop.
    [Show full text]