Minimizing Localized Ratio Cut Objectives in Hypergraphs

Total Page:16

File Type:pdf, Size:1020Kb

Minimizing Localized Ratio Cut Objectives in Hypergraphs Research Track Paper KDD '20, August 23–27, 2020, Virtual Event, USA Minimizing Localized Ratio Cut Objectives in Hypergraphs Nate Veldt Austin R. Benson Jon Kleinberg Cornell University Cornell University Cornell University Center for Applied Mathematics Department of Computer Science Department of Computer Science [email protected] [email protected] [email protected] ABSTRACT to identify communities [13], predict class labels in machine learn- Hypergraphs are a useful abstraction for modeling multiway rela- ing applications [7], and segment images [18]. Standard models tionships in data, and hypergraph clustering is the task of detecting for such clusters are ratio cut objectives, which measure the ratio groups of closely related nodes in such data. Graph clustering has between the number of edges leaving a cluster (the cut) and some been studied extensively, and there are numerous methods for de- notion of the cluster’s size (e.g., the number of edges or nodes in the tecting small, localized clusters without having to explore an entire cluster); common ratio cut objectives include conductance, sparsest input graph. However, there are only a few specialized approaches cut, and normalized cut. Ratio cut objectives are intimately related for localized clustering in hypergraphs. Here we present a frame- to spectral clustering techniques, with the latter providing approxi- work for local hypergraph clustering based on minimizing localized mation guarantees for ratio cut objectives (such as conductance) ratio cut objectives. Our framework takes an input set of reference via so-called Cheeger inequalities [11]. In some cases, these ratio nodes in a hypergraph and solves a sequence of hypergraph mini- cut objectives are optimized over an entire graph to solve a global mum s-t cut problems in order to identify a nearby well-connected clustering or classification task [18]. In other situations, the goal is cluster of nodes that overlaps substantially with the input set. to find sets of nodes that have a small ratio cut and are localized to Our methods extend graph-based techniques but are significantly a certain region of a large graph [3, 4, 27, 34]. more general and have new output quality guarantees. First, our Recently, there has been a surge of hypergraph methods for ma- methods can minimize new generalized notions of hypergraph chine learning and data mining [1, 2, 5, 10, 23, 25, 37, 39], as hyper- cuts, which depend on specific configurations of nodes within each graphs can better model multiway relationships in data. Common hyperedge, rather than just on the number of cut hyperedges. Sec- examples of multiway relationships include academic researchers ond, our framework has several attractive theoretical properties in co-authoring papers, retail items co-purchased by shoppers, and terms of output cluster quality. Most importantly, our algorithm is sets of products or services reviewed by the same person. Due to strongly-local, meaning that its runtime depends only on the size broader modeling ability, there are many hypergraph generaliza- of the input set, and does not need to explore the entire hypergraph tions of graph-cut objectives, including hypergraph variants of ratio to find good local clusters. We use our methodology to effectively cut objectives like conductance and normalized cut [6, 8, 9, 23, 39]. identify clusters in hypergraphs of real-world data with millions of Nevertheless, there are numerous challenges in extending graph- nodes, millions of hyperedges, and large average hyperedge size cut techniques to the hypergraph setting, and current methods with runtimes ranging between a few seconds and a few minutes. for hypergraph-based learning are much less developed than their graph-based counterparts. One major challenge in generalizing ACM Reference Format: graph cut methods is that the concept of a cut hyperedge — how Nate Veldt, Austin R. Benson, and Jon Kleinberg. 2020. Minimizing Localized to define it and how to penalize it — is more nuanced thanthe Ratio Cut Objectives in Hypergraphs. In Proceedings of the 26th ACM SIGKDD concept of a cut edge. While there is only one way to separate the Conference on Knowledge Discovery and Data Mining (KDD ’20), August 23–27, 2020, Virtual Event, CA, USA. ACM, New York, NY, USA, 11 pages. endpoints of an edge into two clusters, there are several ways to https://doi.org/10.1145/3394486.3403222 split up a set of three or more nodes in a hypergedge. Many objective functions model cuts with an all-or-nothing penalty function, which assigns the same penalty to any way of splitting up the nodes of the 1 INTRODUCTION hyperedge (and a penalty of zero if all nodes in the hyperedge are Graphs are a common mathematical abstraction for modeling pair- placed together) [16, 17, 21] (Fig. 1a). However, a common practical wise interactions between objects in a dataset. A standard task in heuristic is a clique expansion, which replaces each hyperedge with graph-based data analysis is to identify well-connected clusters of a weighted clique in a graph [6, 16, 23, 25, 39, 40] (Fig. 1b). The nodes, which share more edges with each other than the rest of advantage is that graph methods can be directly applied, but this the graph [30]. For example, detecting clusters in a graph is used heuristic actually penalizes cut hyperedges differently than the all- or-nothing model. Another downside is that for hypergraph with Permission to make digital or hard copies of all or part of this work for personal or large hyperedges, clique expansion produces a very dense graph. classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation Solving an all-or-nothing cut or applying clique expansion are on the first page. Copyrights for components of this work owned by others than the only two specific models for higher-order relationships. And ifone author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or uses a ratio cut model for clusters in a hypergraph, how to penalize republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. a cut hyperedge may depend on the application. Along these lines, KDD ’20, August 23–27, 2020, Virtual Event, CA, USA inhomogeneous hypergraphs model every possible way to split up a © 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM. hyperedge to assign different penalties [23]; however, these more ACM ISBN 978-1-4503-7998-4/20/08...$15.00 https://doi.org/10.1145/3394486.3403222 sophisticated models are still approximated with weighted clique 1708 Research Track Paper KDD '20, August 23–27, 2020, Virtual Event, USA A major feature of our methods is that they run in strongly-local penalty = w (A) penalty 0,w <latexit sha1_base64="UEJT5+8D0X9OI8ullh/5OE+XvqY=">AAACVHicbVFNSwMxEM2uX7V+VT16CRZBQcquCnoRrF48VrAqdMuSTWdtaDa7JLNq2fZH6kHwl3jxYFp70NaBwOO9mczLS5RJYdDzPhx3bn5hcam0XF5ZXVvfqGxu3Zk01xyaPJWpfoiYASkUNFGghIdMA0siCfdR72qk3z+BNiJVt9jPoJ2wRyViwRlaKqz09gKEFywyUExif0jP6aA+CHgnRTqAIGK8ZyQzXVoflKdbA6FoUHiH9DkEGgzLMzeNiSgunoch7NcPwkrVq3njorPAn4AqmVQjrLwFnZTnCSjk1oRp+V6G7YJpFFyCXZgbyKxD9ggtCxVLwLSLcShDumeZDo1TbY9COmZ/TxQsMaafRLYzYdg109qI/E9r5RiftQuhshxB8Z9FcS4ppnSUMO0IDRxl3wLGtbBeKe8yzTjafyjbEPzpJ8+Cu6Oaf1w7ujmpXlxO4iiRHbJL9olPTskFuSYN0iScvJJPhziO8+58uXPuwk+r60xmtsmfcte/AYg6s0k=</latexit> e time <latexit sha1_base64="WR9A8BSWHwq2U8N0UqbYyS5Y5J4=">AAACMnicbVDLSgMxFM34tr6qLt0ES8GFlJkq6EaoutGdgn1Ap5RMemtDM5khuaOWab/JjV8iuNCFIm79CNPahVYPBA7n3MvJPUEshUHXfXampmdm5+YXFjNLyyura9n1jYqJEs2hzCMZ6VrADEihoIwCJdRiDSwMJFSD7unQr96ANiJSV9iLoRGyayXagjO0UjN7nvcR7jCNQTGJvQE9ov3jvs9bEdI++AHjXSOZ6dDjfmZi0heK+qm7S2+bQP1BM5tzC+4I9C/xxiRHxrhoZh/9VsSTEBRyG2HqnhtjI2UaBZcwyPiJgdjms2uoW6pYCKaRjk4e0LxVWrQdafsU0pH6cyNloTG9MLCTIcOOmfSG4n9ePcH2YSMVKk4QFP8OaieSYkSH/dGW0MBR9ixhXAv7V8o7TDOOtuWMLcGbPPkvqRQL3l6heLmfK52M61ggW2Sb7BCPHJASOSMXpEw4uSdP5JW8OQ/Oi/PufHyPTjnjnU3yC87nF+dBqoA=</latexit> penalty = c( A e A ) , meaning that the runtime is dependent only on the size of e <latexit sha1_base64="IMax4W/PS4BbwAgH7hb17MtAuNU=">AAACm3icbVFdaxNBFJ1dv2r8aNQXQYTBUEjBht0oKEKlsQ8W8aGKaQuZGGYnd7NDZ2eWmbvGsNk/5U/xzX/jJF1EEy8MHM459879SAolHUbRryC8dv3GzVs7t1t37t67v9t+8PDMmdIKGAqjjL1IuAMlNQxRooKLwgLPEwXnyeXxSj//BtZJo7/gooBxzmdaplJw9NSk/YMhfMeqAM0VLmp6SEV3OVgyMTVIl8ASLi6d4i6jg+V+a2/DzaSmrIqe0/kEKKu39EO6JpK0mtcT6A72GWs8DUXf0P5Xn2rlLENurZlTlnPMkqT6/KccV+rA2ANtMJN6Vk/anagXrYNug7gBHdLE6aT9k02NKHPQKPwobhRHBY4rblEKBXWLlQ4KPyefwchDzXNw42q925rueWZKU2P900jX7N8ZFc+dW+SJd64ad5vaivyfNioxfT2upC5KBC2uPkpLRdHQ1aHoVFoQqBYecGGl75WKjFsu0J+z5ZcQb468Dc76vfhFr//pZefoXbOOHfKEPCNdEpNX5IickFMyJCJ4HLwN3gcn4dPwOPwQfryyhkGT84j8E+HwN9Y2y+U=</latexit> 2 { } | |·| \ | A<latexit sha1_base64="Y1jZPZQ/1y5QPUmPr0XFTLTeAiM=">AAAChHicbVFNb9NAEF0bKMV8NMCRy4iookhVZKcgEFJRAxeOBZG2UjZE6804WXW9tnbHhMjxL+FfcePfsEl9gISRVnp67432zUxaauUojn8H4a3bd/bu7t+L7j94+Oig8/jJhSsqK3EoC13Yq1Q41MrgkBRpvCotijzVeJlef1zrl9/ROlWYr7QscZyLmVGZkoI8Nen8POSEP6gu0QhNywZOYTVYcTktCFbIUyGvnRZuDoNVtG3lygCv42NYTBB4s6OfwoZIs3rRTPBo8JLz1tNS8A7633yrVbM5CWuLBfBc0DxN6y9NNJh0unEv3hTsgqQFXdbW+aTzi08LWeVoSPrQbpTEJY1rYUlJjU3EK4eln0jMcOShETm6cb1ZYgOHnplCVlj/DMGG/bujFrlzyzz1znVGt62tyf9po4qyt+NambIiNPLmo6zSQAWsLwJTZVGSXnogpFU+K8i5sEKSv1vkl5Bsj7wLLvq95KTX//yqe/ahXcc+e8aesyOWsDfsjH1i52zIZBAEL4I4SMK98Dg8CV/fWMOg7XnK/qnw/R9t7MII</latexit> the input set, rather than the entire hypergraph. Therefore, we can A<latexit sha1_base64="Y1jZPZQ/1y5QPUmPr0XFTLTeAiM=">AAAChHicbVFNb9NAEF0bKMV8NMCRy4iookhVZKcgEFJRAxeOBZG2UjZE6804WXW9tnbHhMjxL+FfcePfsEl9gISRVnp67432zUxaauUojn8H4a3bd/bu7t+L7j94+Oig8/jJhSsqK3EoC13Yq1Q41MrgkBRpvCotijzVeJlef1zrl9/ROlWYr7QscZyLmVGZkoI8Nen8POSEP6gu0QhNywZOYTVYcTktCFbIUyGvnRZuDoNVtG3lygCv42NYTBB4s6OfwoZIs3rRTPBo8JLz1tNS8A7633yrVbM5CWuLBfBc0DxN6y9NNJh0unEv3hTsgqQFXdbW+aTzi08LWeVoSPrQbpTEJY1rYUlJjU3EK4eln0jMcOShETm6cb1ZYgOHnplCVlj/DMGG/bujFrlzyzz1znVGt62tyf9po4qyt+NambIiNPLmo6zSQAWsLwJTZVGSXnogpFU+K8i5sEKSv1vkl5Bsj7wLLvq95KTX//yqe/ahXcc+e8aesyOWsDfsjH1i52zIZBAEL4I4SMK98Dg8CV/fWMOg7XnK/qnw/R9t7MII</latexit>
Recommended publications
  • Learning on Hypergraphs: Spectral Theory and Clustering
    Learning on Hypergraphs: Spectral Theory and Clustering Pan Li, Olgica Milenkovic Coordinated Science Laboratory University of Illinois at Urbana-Champaign March 12, 2019 Learning on Graphs Graphs are indispensable mathematical data models capturing pairwise interactions: k-nn network social network publication network Important learning on graphs problems: clustering (community detection), semi-supervised/active learning, representation learning (graph embedding) etc. Beyond Pairwise Relations A graph models pairwise relations. Recent work has shown that high-order relations can be significantly more informative: Examples include: Understanding the organization of networks (Benson, Gleich and Leskovec'16) Determining the topological connectivity between data points (Zhou, Huang, Sch}olkopf'07). Graphs with high-order relations can be modeled as hypergraphs (formally defined later). Meta-graphs, meta-paths in heterogeneous information networks. Algorithmic methods for analyzing high-order relations and learning problems are still under development. Beyond Pairwise Relations Functional units in social and biological networks. High-order network motifs: Motif (Benson’16) Microfauna Pelagic fishes Crabs & Benthic fishes Macroinvertebrates Algorithmic methods for analyzing high-order relations and learning problems are still under development. Beyond Pairwise Relations Functional units in social and biological networks. Meta-graphs, meta-paths in heterogeneous information networks. (Zhou, Yu, Han'11) Beyond Pairwise Relations Functional units in social and biological networks. Meta-graphs, meta-paths in heterogeneous information networks. Algorithmic methods for analyzing high-order relations and learning problems are still under development. Review of Graph Clustering: Notation and Terminology Graph Clustering Task: Cluster the vertices that are \densely" connected by edges. Graph Partitioning and Conductance A (weighted) graph G = (V ; E; w): for e 2 E, we is the weight.
    [Show full text]
  • Multiplex Conductance and Gossip Based Information Spreading in Multiplex Networks
    1 Multiplex Conductance and Gossip Based Information Spreading in Multiplex Networks Yufan Huang, Student Member, IEEE, and Huaiyu Dai, Fellow, IEEE Abstract—In this network era, not only people are connected, different networks are also coupled through various interconnections. This kind of network of networks, or multilayer networks, has attracted research interest recently, and many beneficial features have been discovered. However, quantitative study of information spreading in such networks is essentially lacking. Despite some existing results in single networks, the layer heterogeneity and complicated interconnections among the layers make the study of information spreading in this type of networks challenging. In this work, we study the information spreading time in multiplex networks, adopting the gossip (random-walk) based information spreading model. A new metric called multiplex conductance is defined based on the multiplex network structure and used to quantify the information spreading time in a general multiplex network in the idealized setting. Multiplex conductance is then evaluated for some interesting multiplex networks to facilitate understanding in this new area. Finally, the tradeoff between the information spreading efficiency improvement and the layer cost is examined to explain the user’s social behavior and motivate effective multiplex network designs. Index Terms—Information spreading, multiplex networks, gossip algorithm, multiplex conductance F 1 INTRODUCTION N the election year, one of the most important tasks Arguably, this somewhat simplified version of multilayer I for presidential candidates is to disseminate their words networks already captures many interesting multi-scale and and opinions to voters in a fast and efficient manner. multi-component features, and serves as a good starting The underlying research problem on information spreading point for our intended study.
    [Show full text]
  • Rumour Spreading and Graph Conductance∗
    Rumour spreading and graph conductance∗ Flavio Chierichetti, Silvio Lattanzi, Alessandro Panconesi fchierichetti,lattanzi,[email protected] Dipartimento di Informatica Sapienza Universit`adi Roma October 12, 2009 Abstract if its completion time is poly-logarithmic in the size of We show that if a connected graph with n nodes has the network regardless of the source, and that it is slow conductance φ then rumour spreading, also known as otherwise. randomized broadcast, successfully broadcasts a mes- Randomized broadcast has been intensely investi- sage within O(log4 n/φ6) many steps, with high proba- gated (see the related-work section). Our long term bility, using the PUSH-PULL strategy. An interesting goal is to characterize a set of necessary and/or suffi- feature of our approach is that it draws a connection be- cient conditions for rumour spreading to be fast in a tween rumour spreading and the spectral sparsification given network. In this work, we provide a very general procedure of Spielman and Teng [23]. sufficient condition{ high conductance. Our main moti- vation comes from the study of social networks. Loosely 1 Introduction stated, we are looking after a theorem of the form \Ru- mour spreading is fast in social networks". Our result is Rumour spreading, also known as randomized broadcast a good step in this direction because there are reasons or randomized gossip (all terms that will be used as to believe that social networks have high conductance. synonyms throughout the paper), refers to the following This is certainly the case for preferential attachment distributed algorithm. Starting with one source node models such as that of [18].
    [Show full text]
  • Absorbing State 34, 36 Adaptive Boolean Networks 78 Adaptive Chemical Network 73 Adaptive Coupled Map Lattices 66 Adaptive Epide
    239 Index a b absorbing state 34, 36 b-value 107, 109, 112, 119 adaptive Boolean networks 78 Bak–Sneppen model 74, 75 adaptive chemical network 73 benchmark 216, 217, 234 adaptive coupled map lattices 66 Bethe–Peierls approach 225 adaptive epidemiological network betweenness of a node 68 95 bill-tracking website 6 adaptive network of coupled biodiversity 46 oscillators 83 biological evolution 73 adaptive networks 63–66, 70–72, bipartite networks 211 74, 83, 90, 98, 100, 102 bipartite structure 216 adaptive rewiring 91, 92 bipartition 218, 220 adaptive SIS model 96 bipartitioning problem 219 adjacency matrix 9, 66, 206, 207, birth-death processes 31, 37 209–211, 217, 223 bisection problem 232 aftershock 108, 113, 121, bistability 27 125–129, 131, 134, 135, 137 bivariate analysis 164 aftershock magnitudes 127 bivariate measures 162 aftershock sequence 121, 123 bivariate nonlinear approaches aftershock sequences 136 163 agent-based models 73 bivariate time-series analysis agent-based simulation 172, 173, 178 frameworks 3 block-structure 201, 203, 217, air transportation network 2, 10 222, 234, 235 anti-epileptic drugs 159, 163 Boolean networks 65, 76, 79 assortative 67 bounded rationality 25 attractor 164 box-counting technique 112, 113 autoregressive modeling 161, brain 161 172, 175 Brownian motion 6, 7 avalanche size 139 burst-firing 169, 170 average shortest path length 167 bursting 166, 167, 170 Reviews of Nonlinear Dynamics and Complexity. Edited by Heinz Georg Schuster Copyright c 2009 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim ISBN:
    [Show full text]
  • Random Boolean Networks As a Toy Model for the Brain
    UNIVERSITY OF GENEVA SCIENCE FACULTY VRIJE UNIVERSITEIT OF AMSTERDAM PHYSICS SECTION Random Boolean Networks as a toy model for the brain MASTER THESIS presented at the science faculty of the University of Geneva for obtaining the Master in Theoretical Physics by Chlo´eB´eguin Supervisor (VU): Pr. Greg J Stephens Co-Supervisor (UNIGE): Pr. J´er^ome Kasparian July 2017 Contents Introduction1 1 Biology, physics and the brain4 1.1 Biological description of the brain................4 1.2 Criticality in the brain......................8 1.2.1 Physics reminder..................... 10 1.2.2 Experimental evidences.................. 15 2 Models of neural networks 20 2.1 Classes of models......................... 21 2.1.1 Spiking models...................... 21 2.1.2 Rate-based models.................... 23 2.1.3 Attractor networks.................... 24 2.1.4 Links between the classes of models........... 25 2.2 Random Boolean Networks.................... 28 2.2.1 General definition..................... 28 2.2.2 Kauffman network.................... 30 2.2.3 Hopfield network..................... 31 2.2.4 Towards a RBN for the brain.............. 32 2.2.5 The model......................... 33 3 Characterisation of RBNs 34 3.1 Attractors............................. 34 3.2 Damage spreading........................ 36 3.3 Canonical specific heat...................... 37 4 Results 40 4.1 One population with Gaussian weights............. 40 4.2 Dale's principle and balance of inhibition - excitation..... 46 4.3 Lognormal distribution of the weights.............. 51 4.4 Discussion............................. 55 i 5 Conclusion 58 Bibliography 60 Acknowledgements 66 A Python Code 67 A.1 Dynamics............................. 67 A.2 Attractor search.......................... 69 A.3 Hamming Distance........................ 73 A.4 Canonical specific heat.....................
    [Show full text]
  • Boolean Analysis of MOS Circuits
    Bo olean Analysis of MOS Circuits Randal E Bryant Computer Science Department CarnegieMellon University Pittsburgh PA February Abstract The switchlevel mo del represents a digital metaloxide semiconductor MOS cir cuit as a network of charge storage no des connected by resistive transistor switches The functionality of such a network can b e expressed as a series of systems of Bo olean equations Solving these equations symb olically yields a set of Bo olean formulas that describ e the mapping from input and current state to the new network state This analysis supp orts the same class of networks as the switchlevel simulator MOSSIM I I and provides the same functionality including the handling of bidirectional eects and indeterminate X logic values In the worst case the analysis of an n no de network can yield a set of formulas containing a total of O n op erations However all but a limited set of dense passtransistor networks give formulas with O n total op era tions The analysis can serve as the basis of ecient programs for a varietyoflogic design tasks including logic simulation on b oth conventional and sp ecial purp ose computers fault simulation test generation and symb olic verication Keywords and phrases switchlevel networks symb olic analysis logic simulation fault simulation simulation accelerators Intro duction The switchlevel mo del has proved successful as an abstract representation of digital metaloxide semiconductor MOS circuits for a variety of applications This mo del repre sents a circuit in terms of its exact transistor
    [Show full text]
  • Percolation Theory Are Well-Suited To
    Models of Disordered Media and Predictions of Associated Hydraulic Conductivity A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science By L AARON BLANK B.S., Wright State University, 2004 2006 Wright State University WRIGHT STATE UNIVERSITY SCHOOL OF GRADUATE STUDIES Novermber 6, 2006 I HEREBY RECOMMEND THAT THE THESIS PREPARED UNDER MY SUPERVISION BY L Blank ENTITLED Models of Disordered Media and Predictions of Associated Hydraulic Conductivity BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Master of Science. _______________________ Allen Hunt, Ph.D. Thesis Advisor _______________________ Lok Lew Yan Voon, Ph.D. Department Chair _______________________ Joseph F. Thomas, Jr., Ph.D. Dean of the School of Graduate Studies Committee on Final Examination ____________________ Allen Hunt, Ph.D. ____________________ Brent D. Foy, Ph.D. ____________________ Gust Bambakidis, Ph.D. ____________________ Thomas Skinner, Ph.D. Abstract In the late 20th century there was a spill of Technetium in eastern Washington State at the US Department of Energy Hanford site. Resulting contamination of water supplies would raise serious health issues for local residents. Therefore, the ability to predict how these contaminants move through the soil is of great interest. The main contribution to contaminant transport arises from being carried along by flowing water. An important control on the movement of the water through the medium is the hydraulic conductivity, K, which defines the ease of water flow for a given pressure difference (analogous to the electrical conductivity). The overall goal of research in this area is to develop a technique which accurately predicts the hydraulic conductivity as well as its distribution, both in the horizontal and the vertical directions, for media representative of the Hanford subsurface.
    [Show full text]
  • Submodular Hypergraphs: P-Laplacians, Cheeger Inequalities and Spectral Clustering
    Submodular Hypergraphs: p-Laplacians, Cheeger Inequalities and Spectral Clustering Pan Li 1 Olgica Milenkovic 1 Abstract approximations have been based on the assumption that each hyperedge cut has the same weight, in which case the We introduce submodular hypergraphs, a family underlying hypergraph is termed homogeneous. of hypergraphs that have different submodular weights associated with different cuts of hyper- However, in image segmentation, MAP inference on edges. Submodular hypergraphs arise in cluster- Markov random fields (Arora et al., 2012; Shanu et al., ing applications in which higher-order structures 2016), network motif studies (Li & Milenkovic, 2017; Ben- carry relevant information. For such hypergraphs, son et al., 2016; Tsourakakis et al., 2017) and rank learn- we define the notion of p-Laplacians and derive ing (Li & Milenkovic, 2017), higher order relations between corresponding nodal domain theorems and k-way vertices captured by hypergraphs are typically associated Cheeger inequalities. We conclude with the de- with different cut weights. In (Li & Milenkovic, 2017), Li scription of algorithms for computing the spectra and Milenkovic generalized the notion of hyperedge cut of 1- and 2-Laplacians that constitute the basis of weights by assuming that different hyperedge cuts have new spectral hypergraph clustering methods. different weights, and that consequently, each hyperedge is associated with a vector of weights rather than a single scalar weight. If the weights of the hyperedge cuts are sub- 1. Introduction modular, then one can use a graph with nonnegative edge Spectral clustering algorithms are designed to solve a relax- weights to efficiently approximate the hypergraph, provided ation of the graph cut problem based on graph Laplacians that the largest size of a hyperedge is a relatively small con- that capture pairwise dependencies between vertices, and stant.
    [Show full text]
  • Artificial Neural Networks Based on Memristive Devices
    SCIENCE CHINA Information Sciences . REVIEW . June 2018, Vol. 61 060423:1–060423:14 https://doi.org/10.1007/s11432-018-9425-1 Special Focus on Memristive Devices and Neuromorphic Computing Artificial neural networks based on memristive devices Vignesh RAVICHANDRAN, Can LI, Ali BANAGOZAR, J. Joshua YANG & Qiangfei XIA* Department of Electrical and Computer Engineering, University of Massachusetts, Amherst MA 01003, USA Received 22 February 2018/Accepted 27 March 2018/Published online 15 May 2018 Abstract The advent of memristive devices and the continuing research and development in the field of neuromorphic computing show great potential as an alternative to traditional von Neumann computing and bring us ever closer to realizing a true “thinking machine”. Novel neural network architectures and algo- rithms inspired by the brain are becoming more and more attractive to keep up with computing needs, relying on intrinsic parallelism and reduced power consumption to outperform more conventional comput- ing methods. This article provides an overview of various neural networks with an emphasis on networks based on memristive emerging devices, with the advantages of memristor neural networks compared with pure complementary metal oxide semiconductor (CMOS) implementations. A general description of neural networks is presented, followed by a survey of prominent CMOS networks, and finally networks implemented using emerging memristive devices are discussed, along with the motivation for developing memristor based networks and the computational potential
    [Show full text]
  • Controllable Fabrication of Percolative Metal Nanoparticle Arrays Applied for Quantum Conductance-Based Strain Sensors
    materials Article Controllable Fabrication of Percolative Metal Nanoparticle Arrays Applied for Quantum Conductance-Based Strain Sensors Zhengyang Du, Ji’an Chen, Chang Liu, Chen Jin and Min Han * National Laboratory of Solid State Microstructures, College of Engineering and Applied Sciences and Collaborative Innovation Centre of Advanced Microstructures, Nanjing University, Nanjing 210093, China; [email protected] (Z.D.); [email protected] (J.C.); [email protected] (C.L.); [email protected] (C.J.) * Correspondence: [email protected]; Tel.: +86-25-83686248 Received: 13 October 2020; Accepted: 28 October 2020; Published: 29 October 2020 Abstract: We use gas phase deposition of well-defined nanoparticles (NPs) to fabricate closely-spaced Pd NP arrays on flexible membranes prepatterned with interdigital electrodes (IDEs). The evolution of the morphology and electron conductance of the NP arrays during deposition is analyzed. The growth of two-dimensional percolation clusters of interconnected NPs, which correlate with the percolation pathway for electron conduction in the NP deposits, is demonstrated. The percolative nature of the NP arrays permits us to finely control the percolation geometries and conductance of the NP film by controlling the NP deposition time so as to realize a precise and reproducible fabrication of sensing materials. Electron transport measurements reveal that the electrical conductance of the NP films is dominated by electron tunneling or hopping across the NP percolating networks. Based on the percolative and quantum tunneling nature, the closely-spaced Pd NP films on PET membranes are used as flexible strain sensors. The sensor demonstrates an excellent response ability to distinguish tiny deformations down to 5 10 4 strain and a high sensitivity with a large gauge factor of 200 up to × − 4% applied strain.
    [Show full text]
  • Approximating Graph Conductance: from Global to Local
    APPROXIMATING GRAPH CONDUCTANCE: FROM GLOBAL TO LOCAL YUEHENG ZHANG Abstract. In this expository paper, we introduce spectral algorithms and random walk based algorithms on graphs through the problem of finding sparse cuts in both global and local settings. First, we consider the global problem of finding a set with minimum conductance, and show how the Laplacian and normalized Laplacian matrices arise naturally when we relax this problem to an eigenvalue computation problem. Then we bound the relaxed quantity, completing a spectral approximation algorithm and a proof of Cheeger's in- equality. Next, we discuss limitations of this global algorithm and motivate the problem of finding local conductance, for which we introduce random walks on graphs. We discuss how to gain information about conductance from the con- vergence of random walks, and motivate convergence bounds related to local conductance. Finally, we briefly introduce algorithms utilizing such bounds. Throughout, we place an emphasis on providing motivation and explanations from an algorithmic point of view. Contents 1. Introduction 1 1.1. Notation and general definitions 3 1.2. Conductance of cuts and of a graph 4 2. Generalized Rayleigh quotient and variational characterization of eigenvalues 5 3. Spectral approximation of the conductance of a graph 7 3.1. Relaxing the problem of finding the conductance of a graph 7 3.1.1. Cut capacity and the Laplacian 7 3.1.2. Normalized Laplacian and the easy side of Cheeger 9 3.2. Lower bounding the relaxation and the hard side of Cheeger 11 4. Local low-conductance sets and random walks on a graph 15 4.1.
    [Show full text]
  • Graph Clustering Outline
    Graph Clustering Outline • Min s-t cut problem • Min cut problem • Multiway cut • Minimum k-cut • Other normalized cuts and spectral graph partitionings Min s-t cut • Weighted graph G(V,E) • An s-t cut C = (S,T) of a graph G = (V, E) is a cut partition of V into S and T such that s∈S and t∈T • Cost of a cut: Cost(C) = Σe(u,v) uЄS, v ЄT w(e) • Problem: Given G, s and t find the minimum cost s-t cut Max flow problem • Flow network – Abstraction for material flowing through the edges – G = (V,E) directed graph with no parallel edges – Two distinguished nodes: s = source, t= sink – c(e) = capacity of edge e Cuts • An s-t cut is a partition (S,T) of V with sЄS and tЄT • capacity of a cut (S,T) is cap(S,T) = Σe out of Sc(e) • Find s-t cut with the minimum capacity: this problem can be solved optimally in polynomial time by using flow techniques Flows • An s-t flow is a function that satisfies – For each eЄE 0≤f(e) ≤c(e) [capacity] – For each vЄV-{s,t}: Σe in to vf(e) = Σe out of vf(e) [conservation] • The value of a flow f is: v(f) = Σe out of s f(e) Max flow problem • Find s-t flow of maximum value Flows and cuts • Flow value lemma: Let f be any flow and let (S,T) be any s-t cut. Then, the net flow sent across the cut is equal to the amount leaving s Σe out of S f(e) – Σe in to S f(e) = v(f) Flows and cuts • Weak duality: Let f be any flow and let (S,T) be any s-t cut.
    [Show full text]