A Survey of Data Mining of Graphs Using Spectral Graph Theory
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Three Conjectures in Extremal Spectral Graph Theory
Three conjectures in extremal spectral graph theory Michael Tait and Josh Tobin June 6, 2016 Abstract We prove three conjectures regarding the maximization of spectral invariants over certain families of graphs. Our most difficult result is that the join of P2 and Pn−2 is the unique graph of maximum spectral radius over all planar graphs. This was conjectured by Boots and Royle in 1991 and independently by Cao and Vince in 1993. Similarly, we prove a conjecture of Cvetkovi´cand Rowlinson from 1990 stating that the unique outerplanar graph of maximum spectral radius is the join of a vertex and Pn−1. Finally, we prove a conjecture of Aouchiche et al from 2008 stating that a pineapple graph is the unique connected graph maximizing the spectral radius minus the average degree. To prove our theorems, we use the leading eigenvector of a purported extremal graph to deduce structural properties about that graph. Using this setup, we give short proofs of several old results: Mantel's Theorem, Stanley's edge bound and extensions, the K}ovari-S´os-Tur´anTheorem applied to ex (n; K2;t), and a partial solution to an old problem of Erd}oson making a triangle-free graph bipartite. 1 Introduction Questions in extremal graph theory ask to maximize or minimize a graph invariant over a fixed family of graphs. Perhaps the most well-studied problems in this area are Tur´an-type problems, which ask to maximize the number of edges in a graph which does not contain fixed forbidden subgraphs. Over a century old, a quintessential example of this kind of result is Mantel's theorem, which states that Kdn=2e;bn=2c is the unique graph maximizing the number of edges over all triangle-free graphs. -
Manifold Learning
MANIFOLD LEARNING Kavé Salamatian- LISTIC Learning Machine Learning: develop algorithms to automatically extract ”patterns” or ”regularities” from data (generalization) Typical tasks Clustering: Find groups of similar points Dimensionality reduction: Project points in a lower dimensional space while preserving structure Semi-supervised: Given labelled and unlabelled points, build a labelling function Supervised: Given labelled points, build a labelling function All these tasks are not well-defined Clustering Identify the two groups Dimensionality Reduction 4096 Objects in R ? But only 3 parameters: 2 angles and 1 for illumination. They span a 3- dimensional sub-manifold of R4096! Semi-Supervised Learning Assign labels to black points Supervised learning Build a function that predicts the label of all points in the space Formal definitions Clustering: Given build a function Dimensionality reduction: Given build a function Semi-supervised: Given and with m << n, build a function Supervised: Given , build Geometric learning Consider and exploit the geometric structure of the data Clustering: two ”close” points should be in the same cluster Dimensionality reduction: ”closeness” should be preserved Semi-supervised/supervised: two ”close” points should have the same label Examples: k-means clustering, k-nearest neighbors. Manifold ? A manifold is a topological space that is locally Euclidian Around every point, there is a neighborhood that is topologically isomorph to unit ball Regularization Adopt a functional viewpoint -
Contemporary Spectral Graph Theory
CONTEMPORARY SPECTRAL GRAPH THEORY GUEST EDITORS Maurizio Brunetti, University of Naples Federico II, Italy, [email protected] Raffaella Mulas, The Alan Turing Institute, UK, [email protected] Lucas Rusnak, Texas State University, USA, [email protected] Zoran Stanić, University of Belgrade, Serbia, [email protected] DESCRIPTION Spectral graph theory is a striking theory that studies the properties of a graph in relationship to the characteristic polynomial, eigenvalues and eigenvectors of matrices associated with the graph, such as the adjacency matrix, the Laplacian matrix, the signless Laplacian matrix or the distance matrix. This theory emerged in the 1950s and 1960s. Over the decades it has considered not only simple undirected graphs but also their generalizations such as multigraphs, directed graphs, weighted graphs or hypergraphs. In the recent years, spectra of signed graphs and gain graphs have attracted a great deal of attention. Besides graph theoretic research, another major source is the research in a wide branch of applications in chemistry, physics, biology, electrical engineering, social sciences, computer sciences, information theory and other disciplines. This thematic special issue is devoted to the publication of original contributions focusing on the spectra of graphs or their generalizations, along with the most recent applications. Contributions to the Special Issue may address (but are not limited) to the following aspects: f relations between spectra of graphs (or their generalizations) and their structural properties in the most general shape, f bounds for particular eigenvalues, f spectra of particular types of graph, f relations with block designs, f particular eigenvalues of graphs, f cospectrality of graphs, f graph products and their eigenvalues, f graphs with a comparatively small number of eigenvalues, f graphs with particular spectral properties, f applications, f characteristic polynomials. -
Learning on Hypergraphs: Spectral Theory and Clustering
Learning on Hypergraphs: Spectral Theory and Clustering Pan Li, Olgica Milenkovic Coordinated Science Laboratory University of Illinois at Urbana-Champaign March 12, 2019 Learning on Graphs Graphs are indispensable mathematical data models capturing pairwise interactions: k-nn network social network publication network Important learning on graphs problems: clustering (community detection), semi-supervised/active learning, representation learning (graph embedding) etc. Beyond Pairwise Relations A graph models pairwise relations. Recent work has shown that high-order relations can be significantly more informative: Examples include: Understanding the organization of networks (Benson, Gleich and Leskovec'16) Determining the topological connectivity between data points (Zhou, Huang, Sch}olkopf'07). Graphs with high-order relations can be modeled as hypergraphs (formally defined later). Meta-graphs, meta-paths in heterogeneous information networks. Algorithmic methods for analyzing high-order relations and learning problems are still under development. Beyond Pairwise Relations Functional units in social and biological networks. High-order network motifs: Motif (Benson’16) Microfauna Pelagic fishes Crabs & Benthic fishes Macroinvertebrates Algorithmic methods for analyzing high-order relations and learning problems are still under development. Beyond Pairwise Relations Functional units in social and biological networks. Meta-graphs, meta-paths in heterogeneous information networks. (Zhou, Yu, Han'11) Beyond Pairwise Relations Functional units in social and biological networks. Meta-graphs, meta-paths in heterogeneous information networks. Algorithmic methods for analyzing high-order relations and learning problems are still under development. Review of Graph Clustering: Notation and Terminology Graph Clustering Task: Cluster the vertices that are \densely" connected by edges. Graph Partitioning and Conductance A (weighted) graph G = (V ; E; w): for e 2 E, we is the weight. -
Spectral Geometry for Structural Pattern Recognition
Spectral Geometry for Structural Pattern Recognition HEWAYDA EL GHAWALBY Ph.D. Thesis This thesis is submitted in partial fulfilment of the requirements for the degree of Doctor of Philosophy. Department of Computer Science United Kingdom May 2011 Abstract Graphs are used pervasively in computer science as representations of data with a network or relational structure, where the graph structure provides a flexible representation such that there is no fixed dimensionality for objects. However, the analysis of data in this form has proved an elusive problem; for instance, it suffers from the robustness to structural noise. One way to circumvent this problem is to embed the nodes of a graph in a vector space and to study the properties of the point distribution that results from the embedding. This is a problem that arises in a number of areas including manifold learning theory and graph-drawing. In this thesis, our first contribution is to investigate the heat kernel embed- ding as a route to computing geometric characterisations of graphs. The reason for turning to the heat kernel is that it encapsulates information concerning the distribution of path lengths and hence node affinities on the graph. The heat ker- nel of the graph is found by exponentiating the Laplacian eigensystem over time. The matrix of embedding co-ordinates for the nodes of the graph is obtained by performing a Young-Householder decomposition on the heat kernel. Once the embedding of its nodes is to hand we proceed to characterise a graph in a geometric manner. With the embeddings to hand, we establish a graph character- ization based on differential geometry by computing sets of curvatures associated ii Abstract iii with the graph nodes, edges and triangular faces. -
CS 598: Spectral Graph Theory. Lecture 10
CS 598: Spectral Graph Theory. Lecture 13 Expander Codes Alexandra Kolla Today Graph approximations, part 2. Bipartite expanders as approximations of the bipartite complete graph Quasi-random properties of bipartite expanders, expander mixing lemma Building expander codes Encoding, minimum distance, decoding Expander Graph Refresher We defined expander graphs to be d- regular graphs whose adjacency matrix eigenvalues satisfy |훼푖| ≤ 휖푑 for i>1, and some small 휖. We saw that such graphs are very good approximations of the complete graph. Expanders as Approximations of the Complete Graph Refresher Let G be a d-regular graph whose adjacency eigenvalues satisfy |훼푖| ≤ 휖푑. As its Laplacian eigenvalues satisfy 휆푖 = 푑 − 훼푖, all non-zero eigenvalues are between (1 − ϵ)d and 1 + ϵ 푑. 푇 푇 Let H=(d/n)Kn, 푠표 푥 퐿퐻푥 = 푑푥 푥 We showed that G is an ϵ − approximation of H 푇 푇 푇 1 − ϵ 푥 퐿퐻푥 ≼ 푥 퐿퐺푥 ≼ 1 + ϵ 푥 퐿퐻푥 Bipartite Expander Graphs Define them as d-regular good approximations of (some multiple of) the complete bipartite 푑 graph H= 퐾 : 푛 푛,푛 푇 푇 푇 1 − ϵ 푥 퐿퐻푥 ≼ 푥 퐿퐺푥 ≼ 1 + ϵ 푥 퐿퐻푥 ⇒ 1 − ϵ 퐻 ≼ 퐺 ≼ 1 + ϵ 퐻 G 퐾푛,푛 U V U V The Spectrum of Bipartite Expanders 푑 Eigenvalues of Laplacian of H= 퐾 are 푛 푛,푛 휆1 = 0, 휆2푛 = 2푑, 휆푖 = 푑, 0 < 푖 < 2푛 푇 푇 푇 1 − ϵ 푥 퐿퐻푥 ≼ 푥 퐿퐺푥 ≼ 1 + ϵ 푥 퐿퐻푥 Says we want d-regular graph G that has evals 휆1 = 0, 휆2푛 = 2푑, 휆푖 ∈ ( 1 − 휖 푑, 1 + 휖 푑), 0 < 푖 < 2푛 Construction of Bipartite Expanders Definition (Double Cover). -
Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts
University of Pennsylvania ScholarlyCommons Technical Reports (CIS) Department of Computer & Information Science 1-1-2013 Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts Jean H. Gallier University of Pennsylvania, [email protected] Follow this and additional works at: https://repository.upenn.edu/cis_reports Part of the Computer Engineering Commons, and the Computer Sciences Commons Recommended Citation Jean H. Gallier, "Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts", . January 2013. University of Pennsylvania Department of Computer and Information Science Technical Report No. MS-CIS-13-09. This paper is posted at ScholarlyCommons. https://repository.upenn.edu/cis_reports/986 For more information, please contact [email protected]. Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts Abstract These are notes on the method of normalized graph cuts and its applications to graph clustering. I provide a fairly thorough treatment of this deeply original method due to Shi and Malik, including complete proofs. I include the necessary background on graphs and graph Laplacians. I then explain in detail how the eigenvectors of the graph Laplacian can be used to draw a graph. This is an attractive application of graph Laplacians. The main thrust of this paper is the method of normalized cuts. I give a detailed account for K = 2 clusters, and also for K > 2 clusters, based on the work of Yu and Shi. Three points that do not appear to have been clearly articulated before are elaborated: 1. The solutions of the main optimization problem should be viewed as tuples in the K-fold cartesian product of projective space RP^{N-1}. -
Chapter 2 a Short Review of Matrix Algebra
Chapter 2 A short review of matrix algebra 2.1 Vectors and vector spaces Definition 2.1.1. A vector a of dimension n is a collection of n elements typically written as ⎛ ⎞ ⎜ a1 ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ a2 ⎟ a = ⎜ ⎟ =(ai)n. ⎜ . ⎟ ⎝ . ⎠ an Vectors of length 2 (two-dimensional vectors) can be thought of points in 33 BIOS 2083 Linear Models Abdus S. Wahed the plane (See figures). Chapter 2 34 BIOS 2083 Linear Models Abdus S. Wahed Figure 2.1: Vectors in two and three dimensional spaces (-1.5,2) (1, 1) (1, -2) x1 (2.5, 1.5, 0.95) x2 (0, 1.5, 0.95) x3 Chapter 2 35 BIOS 2083 Linear Models Abdus S. Wahed • A vector with all elements equal to zero is known as a zero vector and is denoted by 0. • A vector whose elements are stacked vertically is known as column vector whereas a vector whose elements are stacked horizontally will be referred to as row vector. (Unless otherwise mentioned, all vectors will be referred to as column vectors). • A row vector representation of a column vector is known as its trans- T pose. We will use⎛ the⎞ notation ‘ ’or‘ ’ to indicate a transpose. For ⎜ a1 ⎟ ⎜ ⎟ ⎜ a2 ⎟ ⎜ ⎟ T instance, if a = ⎜ ⎟ and b =(a1 a2 ... an), then we write b = a ⎜ . ⎟ ⎝ . ⎠ an or a = bT . • Vectors of same dimension are conformable to algebraic operations such as additions and subtractions. Sum of two or more vectors of dimension n results in another n-dimensional vector with elements as the sum of the corresponding elements of summand vectors. -
Spectral Graph Theory – from Practice to Theory
Spectral Graph Theory – From Practice to Theory Alexander Farrugia [email protected] Abstract Graph theory is the area of mathematics that studies networks, or graphs. It arose from the need to analyse many diverse network-like structures like road networks, molecules, the Internet, social networks and electrical networks. In spectral graph theory, which is a branch of graph theory, matrices are constructed from such graphs and analysed from the point of view of their so-called eigenvalues and eigenvectors. The first practical need for studying graph eigenvalues was in quantum chemistry in the thirties, forties and fifties, specifically to describe the Hückel molecular orbital theory for unsaturated conjugated hydrocarbons. This study led to the field which nowadays is called chemical graph theory. A few years later, during the late fifties and sixties, graph eigenvalues also proved to be important in physics, particularly in the solution of the membrane vibration problem via the discrete approximation of the membrane as a graph. This paper delves into the journey of how the practical needs of quantum chemistry and vibrating membranes compelled the creation of the more abstract spectral graph theory. Important, yet basic, mathematical results stemming from spectral graph theory shall be mentioned in this paper. Later, areas of study that make full use of these mathematical results, thus benefitting greatly from spectral graph theory, shall be described. These fields of study include the P versus NP problem in the field of computational complexity, Internet search, network centrality measures and control theory. Keywords: spectral graph theory, eigenvalue, eigenvector, graph. Introduction: Spectral Graph Theory Graph theory is the area of mathematics that deals with networks. -
Facts from Linear Algebra
Appendix A Facts from Linear Algebra Abstract We introduce the notation of vector and matrices (cf. Section A.1), and recall the solvability of linear systems (cf. Section A.2). Section A.3 introduces the spectrum σ(A), matrix polynomials P (A) and their spectra, the spectral radius ρ(A), and its properties. Block structures are introduced in Section A.4. Subjects of Section A.5 are orthogonal and orthonormal vectors, orthogonalisation, the QR method, and orthogonal projections. Section A.6 is devoted to the Schur normal form (§A.6.1) and the Jordan normal form (§A.6.2). Diagonalisability is discussed in §A.6.3. Finally, in §A.6.4, the singular value decomposition is explained. A.1 Notation for Vectors and Matrices We recall that the field K denotes either R or C. Given a finite index set I, the linear I space of all vectors x =(xi)i∈I with xi ∈ K is denoted by K . The corresponding square matrices form the space KI×I . KI×J with another index set J describes rectangular matrices mapping KJ into KI . The linear subspace of a vector space V spanned by the vectors {xα ∈V : α ∈ I} is denoted and defined by α α span{x : α ∈ I} := aαx : aα ∈ K . α∈I I×I T Let A =(aαβ)α,β∈I ∈ K . Then A =(aβα)α,β∈I denotes the transposed H matrix, while A =(aβα)α,β∈I is the adjoint (or Hermitian transposed) matrix. T H Note that A = A holds if K = R . Since (x1,x2,...) indicates a row vector, T (x1,x2,...) is used for a column vector. -
On the Distance Spectra of Graphs
On the distance spectra of graphs Ghodratollah Aalipour∗ Aida Abiad† Zhanar Berikkyzy‡ Jay Cummings§ Jessica De Silva¶ Wei Gaok Kristin Heysse‡ Leslie Hogben‡∗∗ Franklin H.J. Kenter†† Jephian C.-H. Lin‡ Michael Tait§ October 5, 2015 Abstract The distance matrix of a graph G is the matrix containing the pairwise distances between vertices. The distance eigenvalues of G are the eigenvalues of its distance matrix and they form the distance spectrum of G. We determine the distance spectra of halved cubes, double odd graphs, and Doob graphs, completing the determination of distance spectra of distance regular graphs having exactly one positive distance eigenvalue. We characterize strongly regular graphs having more positive than negative distance eigenvalues. We give examples of graphs with few distinct distance eigenvalues but lacking regularity properties. We also determine the determinant and inertia of the distance matrices of lollipop and barbell graphs. Keywords. distance matrix, eigenvalue, distance regular graph, Kneser graph, double odd graph, halved cube, Doob graph, lollipop graph, barbell graph, distance spectrum, strongly regular graph, optimistic graph, determinant, inertia, graph AMS subject classifications. 05C12, 05C31, 05C50, 15A15, 15A18, 15B48, 15B57 1 Introduction The distance matrix (G) = [d ] of a graph G is the matrix indexed by the vertices of G where d = d(v , v ) D ij ij i j is the distance between the vertices vi and vj , i.e., the length of a shortest path between vi and vj . Distance matrices were introduced in the study of a data communication problem in [16]. This problem involves finding appropriate addresses so that a message can move efficiently through a series of loops from its origin to its destination, choosing the best route at each switching point. -
Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts
Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA e-mail: [email protected] c Jean Gallier November 12, 2013 arXiv:1311.2492v1 [cs.CV] 11 Nov 2013 2 Contents 1 Introduction 5 2 Graphs and Graph Laplacians; Basic Facts 13 2.1 Directed Graphs, Undirected Graphs, Weighted Graphs . 13 2.2 Laplacian Matrices of Graphs . 18 3 Spectral Graph Drawing 25 3.1 Graph Drawing and Energy Minimization . 25 3.2 Examples of Graph Drawings . 28 4 Graph Clustering 33 4.1 Graph Clustering Using Normalized Cuts . 33 4.2 Special Case: 2-Way Clustering Using Normalized Cuts . 34 4.3 K-Way Clustering Using Normalized Cuts . 41 4.4 K-Way Clustering; Using The Dependencies Among X1;:::;XK ....... 53 4.5 Discrete Solution Close to a Continuous Approximation . 56 A The Rayleigh Ratio and the Courant-Fischer Theorem 61 B Riemannian Metrics on Quotient Manifolds 67 Bibliography 73 3 4 CONTENTS Chapter 1 Introduction In the Fall of 2012, my friend Kurt Reillag suggested that I should be ashamed about knowing so little about graph Laplacians and normalized graph cuts. These notes are the result of my efforts to rectify this situation. I begin with a review of basic notions of graph theory. Even though the graph Laplacian is fundamentally associated with an undirected graph, I review the definition of both directed and undirected graphs. For both directed and undirected graphs, I define the degree matrix D, the incidence matrix De, and the adjacency matrix A.