Percolation and Random Graphs
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Optimal Subgraph Structures in Scale-Free Configuration
Submitted to the Annals of Applied Probability OPTIMAL SUBGRAPH STRUCTURES IN SCALE-FREE CONFIGURATION MODELS , By Remco van der Hofstad∗ §, Johan S. H. van , , Leeuwaarden∗ ¶ and Clara Stegehuis∗ k Eindhoven University of Technology§, Tilburg University¶ and Twente Universityk Subgraphs reveal information about the geometry and function- alities of complex networks. For scale-free networks with unbounded degree fluctuations, we obtain the asymptotics of the number of times a small connected graph occurs as a subgraph or as an induced sub- graph. We obtain these results by analyzing the configuration model with degree exponent τ (2, 3) and introducing a novel class of op- ∈ timization problems. For any given subgraph, the unique optimizer describes the degrees of the vertices that together span the subgraph. We find that subgraphs typically occur between vertices with specific degree ranges. In this way, we can count and characterize all sub- graphs. We refrain from double counting in the case of multi-edges, essentially counting the subgraphs in the erased configuration model. 1. Introduction Scale-free networks often have degree distributions that follow power laws with exponent τ (2, 3) [1, 11, 22, 34]. Many net- ∈ works have been reported to satisfy these conditions, including metabolic networks, the internet and social networks. Scale-free networks come with the presence of hubs, i.e., vertices of extremely high degrees. Another property of real-world scale-free networks is that the clustering coefficient (the probability that two uniformly chosen neighbors of a vertex are neighbors themselves) decreases with the vertex degree [4, 10, 24, 32, 34], again following a power law. -
Detecting Statistically Significant Communities
1 Detecting Statistically Significant Communities Zengyou He, Hao Liang, Zheng Chen, Can Zhao, Yan Liu Abstract—Community detection is a key data analysis problem across different fields. During the past decades, numerous algorithms have been proposed to address this issue. However, most work on community detection does not address the issue of statistical significance. Although some research efforts have been made towards mining statistically significant communities, deriving an analytical solution of p-value for one community under the configuration model is still a challenging mission that remains unsolved. The configuration model is a widely used random graph model in community detection, in which the degree of each node is preserved in the generated random networks. To partially fulfill this void, we present a tight upper bound on the p-value of a single community under the configuration model, which can be used for quantifying the statistical significance of each community analytically. Meanwhile, we present a local search method to detect statistically significant communities in an iterative manner. Experimental results demonstrate that our method is comparable with the competing methods on detecting statistically significant communities. Index Terms—Community Detection, Random Graphs, Configuration Model, Statistical Significance. F 1 INTRODUCTION ETWORKS are widely used for modeling the structure function that is able to always achieve the best performance N of complex systems in many fields, such as biology, in all possible scenarios. engineering, and social science. Within the networks, some However, most of these objective functions (metrics) do vertices with similar properties will form communities that not address the issue of statistical significance of commu- have more internal connections and less external links. -
Modern Discrete Probability I
Preliminaries Some fundamental models A few more useful facts about... Modern Discrete Probability I - Introduction Stochastic processes on graphs: models and questions Sebastien´ Roch UW–Madison Mathematics September 6, 2017 Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... 1 Preliminaries Review of graph theory Review of Markov chain theory 2 Some fundamental models Random walks on graphs Percolation Some random graph models Markov random fields Interacting particles on finite graphs 3 A few more useful facts about... ...graphs ...Markov chains ...other things Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... Graphs Definition (Undirected graph) An undirected graph (or graph for short) is a pair G = (V ; E) where V is the set of vertices (or nodes, sites) and E ⊆ ffu; vg : u; v 2 V g; is the set of edges (or bonds). The V is either finite or countably infinite. Edges of the form fug are called loops. We do not allow E to be a multiset. We occasionally write V (G) and E(G) for the vertices and edges of G. Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... An example: the Petersen graph Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about.. -
Probabilistic and Geometric Methods in Last Passage Percolation
Probabilistic and geometric methods in last passage percolation by Milind Hegde A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Mathematics in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Alan Hammond, Co‐chair Professor Shirshendu Ganguly, Co‐chair Professor Fraydoun Rezakhanlou Professor Alistair Sinclair Spring 2021 Probabilistic and geometric methods in last passage percolation Copyright 2021 by Milind Hegde 1 Abstract Probabilistic and geometric methods in last passage percolation by Milind Hegde Doctor of Philosophy in Mathematics University of California, Berkeley Professor Alan Hammond, Co‐chair Professor Shirshendu Ganguly, Co‐chair Last passage percolation (LPP) refers to a broad class of models thought to lie within the Kardar‐ Parisi‐Zhang universality class of one‐dimensional stochastic growth models. In LPP models, there is a planar random noise environment through which directed paths travel; paths are as‐ signed a weight based on their journey through the environment, usually by, in some sense, integrating the noise over the path. For given points y and x, the weight from y to x is defined by maximizing the weight over all paths from y to x. A path which achieves the maximum weight is called a geodesic. A few last passage percolation models are exactly solvable, i.e., possess what is called integrable structure. This gives rise to valuable information, such as explicit probabilistic resampling prop‐ erties, distributional convergence information, or one point tail bounds of the weight profile as the starting point y is fixed and the ending point x varies. -
Lecture 13: Financial Disasters and Econophysics
Lecture 13: Financial Disasters and Econophysics Big problem: Power laws in economy and finance vs Great Moderation: (Source: http://www.mckinsey.com/business-functions/strategy-and-corporate-finance/our-insights/power-curves-what- natural-and-economic-disasters-have-in-common Analysis of big data, discontinuous change especially of financial sector, where efficient market theory missed the boat has drawn attention of specialists from physics and mathematics. Wall Street“quant”models may have helped the market implode; and collapse spawned econophysics work on finance instability. NATURE PHYSICS March 2013 Volume 9, No 3 pp119-197 : “The 2008 financial crisis has highlighted major limitations in the modelling of financial and economic systems. However, an emerging field of research at the frontiers of both physics and economics aims to provide a more fundamental understanding of economic networks, as well as practical insights for policymakers. In this Nature Physics Focus, physicists and economists consider the state-of-the-art in the application of network science to finance.” The financial crisis has made us aware that financial markets are very complex networks that, in many cases, we do not really understand and that can easily go out of control. This idea, which would have been shocking only 5 years ago, results from a number of precise reasons. What does physics bring to social science problems? 1- Heterogeneous agents – strange since physics draws strength from electron is electron is electron but STATISTICAL MECHANICS --- minority game, finance artificial agents, 2- Facility with huge data sets – data-mining for regularities in time series with open eyes. 3- Network analysis 4- Percolation/other models of “phase transition”, which directs attention at boundary conditions AN INTRODUCTION TO ECONOPHYSICS Correlations and Complexity in Finance ROSARIO N. -
Processes on Complex Networks. Percolation
Chapter 5 Processes on complex networks. Percolation 77 Up till now we discussed the structure of the complex networks. The actual reason to study this structure is to understand how this structure influences the behavior of random processes on networks. I will talk about two such processes. The first one is the percolation process. The second one is the spread of epidemics. There are a lot of open problems in this area, the main of which can be innocently formulated as: How the network topology influences the dynamics of random processes on this network. We are still quite far from a definite answer to this question. 5.1 Percolation 5.1.1 Introduction to percolation Percolation is one of the simplest processes that exhibit the critical phenomena or phase transition. This means that there is a parameter in the system, whose small change yields a large change in the system behavior. To define the percolation process, consider a graph, that has a large connected component. In the classical settings, percolation was actually studied on infinite graphs, whose vertices constitute the set Zd, and edges connect each vertex with nearest neighbors, but we consider general random graphs. We have parameter ϕ, which is the probability that any edge present in the underlying graph is open or closed (an event with probability 1 − ϕ) independently of the other edges. Actually, if we talk about edges being open or closed, this means that we discuss bond percolation. It is also possible to talk about the vertices being open or closed, and this is called site percolation. -
Percolation Thresholds for Robust Network Connectivity
Percolation Thresholds for Robust Network Connectivity Arman Mohseni-Kabir1, Mihir Pant2, Don Towsley3, Saikat Guha4, Ananthram Swami5 1. Physics Department, University of Massachusetts Amherst, [email protected] 2. Massachusetts Institute of Technology 3. College of Information and Computer Sciences, University of Massachusetts Amherst 4. College of Optical Sciences, University of Arizona 5. Army Research Laboratory Abstract Communication networks, power grids, and transportation networks are all examples of networks whose performance depends on reliable connectivity of their underlying network components even in the presence of usual network dynamics due to mobility, node or edge failures, and varying traffic loads. Percolation theory quantifies the threshold value of a local control parameter such as a node occupation (resp., deletion) probability or an edge activation (resp., removal) probability above (resp., below) which there exists a giant connected component (GCC), a connected component comprising of a number of occupied nodes and active edges whose size is proportional to the size of the network itself. Any pair of occupied nodes in the GCC is connected via at least one path comprised of active edges and occupied nodes. The mere existence of the GCC itself does not guarantee that the long-range connectivity would be robust, e.g., to random link or node failures due to network dynamics. In this paper, we explore new percolation thresholds that guarantee not only spanning network connectivity, but also robustness. We define and analyze four measures of robust network connectivity, explore their interrelationships, and numerically evaluate the respective robust percolation thresholds arXiv:2006.14496v1 [physics.soc-ph] 25 Jun 2020 for the 2D square lattice. -
Understanding Complex Systems: a Communication Networks Perspective
1 Understanding Complex Systems: A Communication Networks Perspective Pavlos Antoniou and Andreas Pitsillides Networks Research Laboratory, Computer Science Department, University of Cyprus 75 Kallipoleos Street, P.O. Box 20537, 1678 Nicosia, Cyprus Telephone: +357-22-892687, Fax: +357-22-892701 Email: [email protected], [email protected] Technical Report TR-07-01 Department of Computer Science University of Cyprus February 2007 Abstract Recent approaches on the study of networks have exploded over almost all the sciences across the academic spectrum. Over the last few years, the analysis and modeling of networks as well as networked dynamical systems have attracted considerable interdisciplinary interest. These efforts were driven by the fact that systems as diverse as genetic networks or the Internet can be best described as complex networks. On the contrary, although the unprecedented evolution of technology, basic issues and fundamental principles related to the structural and evolutionary properties of networks still remain unaddressed and need to be unraveled since they affect the function of a network. Therefore, the characterization of the wiring diagram and the understanding on how an enormous network of interacting dynamical elements is able to behave collectively, given their individual non linear dynamics are of prime importance. In this study we explore simple models of complex networks from real communication networks perspective, focusing on their structural and evolutionary properties. The limitations and vulnerabilities of real communication networks drive the necessity to develop new theoretical frameworks to help explain the complex and unpredictable behaviors of those networks based on the aforementioned principles, and design alternative network methods and techniques which may be provably effective, robust and resilient to accidental failures and coordinated attacks. -
Critical Percolation As a Framework to Analyze the Training of Deep Networks
Published as a conference paper at ICLR 2018 CRITICAL PERCOLATION AS A FRAMEWORK TO ANALYZE THE TRAINING OF DEEP NETWORKS Zohar Ringel Rodrigo de Bem∗ Racah Institute of Physics Department of Engineering Science The Hebrew University of Jerusalem University of Oxford [email protected]. [email protected] ABSTRACT In this paper we approach two relevant deep learning topics: i) tackling of graph structured input data and ii) a better understanding and analysis of deep networks and related learning algorithms. With this in mind we focus on the topological classification of reachability in a particular subset of planar graphs (Mazes). Doing so, we are able to model the topology of data while staying in Euclidean space, thus allowing its processing with standard CNN architectures. We suggest a suitable architecture for this problem and show that it can express a perfect solution to the classification task. The shape of the cost function around this solution is also derived and, remarkably, does not depend on the size of the maze in the large maze limit. Responsible for this behavior are rare events in the dataset which strongly regulate the shape of the cost function near this global minimum. We further identify an obstacle to learning in the form of poorly performing local minima in which the network chooses to ignore some of the inputs. We further support our claims with training experiments and numerical analysis of the cost function on networks with up to 128 layers. 1 INTRODUCTION Deep convolutional networks have achieved great success in the last years by presenting human and super-human performance on many machine learning problems such as image classification, speech recognition and natural language processing (LeCun et al. -
Markov Chain Monte Carlo Estimation of Exponential Random Graph Models
Markov Chain Monte Carlo Estimation of Exponential Random Graph Models Tom A.B. Snijders ICS, Department of Statistics and Measurement Theory University of Groningen ∗ April 19, 2002 ∗Author's address: Tom A.B. Snijders, Grote Kruisstraat 2/1, 9712 TS Groningen, The Netherlands, email <[email protected]>. I am grateful to Paul Snijders for programming the JAVA applet used in this article. In the revision of this article, I profited from discussions with Pip Pattison and Garry Robins, and from comments made by a referee. This paper is formatted in landscape to improve on-screen readability. It is read best by opening Acrobat Reader in a full screen window. Note that in Acrobat Reader, the entire screen can be used for viewing by pressing Ctrl-L; the usual screen is returned when pressing Esc; it is possible to zoom in or zoom out by pressing Ctrl-- or Ctrl-=, respectively. The sign in the upper right corners links to the page viewed previously. ( Tom A.B. Snijders 2 MCMC estimation for exponential random graphs ( Abstract bility to move from one region to another. In such situations, convergence to the target distribution is extremely slow. To This paper is about estimating the parameters of the exponential be useful, MCMC algorithms must be able to make transitions random graph model, also known as the p∗ model, using frequen- from a given graph to a very different graph. It is proposed to tist Markov chain Monte Carlo (MCMC) methods. The exponen- include transitions to the graph complement as updating steps tial random graph model is simulated using Gibbs or Metropolis- to improve the speed of convergence to the target distribution. -
Exploring Healing Strategies for Random Boolean Networks
Exploring Healing Strategies for Random Boolean Networks Christian Darabos 1, Alex Healing 2, Tim Johann 3, Amitabh Trehan 4, and Am´elieV´eron 5 1 Information Systems Department, University of Lausanne, Switzerland 2 Pervasive ICT Research Centre, British Telecommunications, UK 3 EML Research gGmbH, Heidelberg, Germany 4 Department of Computer Science, University of New Mexico, Albuquerque, USA 5 Division of Bioinformatics, Institute for Evolution and Biodiversity, The Westphalian Wilhelms University of Muenster, Germany. [email protected], [email protected], [email protected], [email protected], [email protected] Abstract. Real-world systems are often exposed to failures where those studied theoretically are not: neuron cells in the brain can die or fail, re- sources in a peer-to-peer network can break down or become corrupt, species can disappear from and environment, and so forth. In all cases, for the system to keep running as it did before the failure occurred, or to survive at all, some kind of healing strategy may need to be applied. As an example of such a system subjected to failure and subsequently heal, we study Random Boolean Networks. Deletion of a node in the network was considered a failure if it affected the functional output and we in- vestigated healing strategies that would allow the system to heal either fully or partially. More precisely, our main strategy involves allowing the nodes directly affected by the node deletion (failure) to iteratively rewire in order to achieve healing. We found that such a simple method was ef- fective when dealing with small networks and single-point failures. -
Percolation Threshold Results on Erdos-Rényi Graphs
arXiv: arXiv:0000.0000 Percolation Threshold Results on Erd}os-R´enyi Graphs: an Empirical Process Approach Michael J. Kane Abstract: Keywords and phrases: threshold, directed percolation, stochastic ap- proximation, empirical processes. 1. Introduction Random graphs and discrete random processes provide a general approach to discovering properties and characteristics of random graphs and randomized al- gorithms. The approach generally works by defining an algorithm on a random graph or a randomized algorithm. Then, expected changes for each step of the process are used to propose a limiting differential equation and a large deviation theorem is used to show that the process and the differential equation are close in some sense. In this way a connection is established between the resulting process's stochastic behavior and the dynamics of a deterministic, asymptotic approximation using a differential equation. This approach is generally referred to as stochastic approximation and provides a powerful tool for understanding the asymptotic behavior of a large class of processes defined on random graphs. However, little work has been done in the area of random graph research to investigate the weak limit behavior of these processes before the asymptotic be- havior overwhelms the random component of the process. This context is par- ticularly relevant to researchers studying news propagation in social networks, sensor networks, and epidemiological outbreaks. In each of these applications, investigators may deal graphs containing tens to hundreds of vertices and be interested not only in expected behavior over time but also error estimates. This paper investigates the connectivity of graphs, with emphasis on Erd}os- R´enyi graphs, near the percolation threshold when the number of vertices is not asymptotically large.