Discovering Latent Network Structure in Point Process Data

Total Page:16

File Type:pdf, Size:1020Kb

Discovering Latent Network Structure in Point Process Data Discovering Latent Network Structure in Point Process Data Scott W. Linderman [email protected] Harvard University, Cambridge, MA 02138 USA Ryan P. Adams [email protected] Harvard University, Cambridge, MA 02138 USA Abstract are considered known and the data are the entries in the as- sociated adjacency matrix. A rich literature has arisen in Networks play a central role in modern data anal- recent years for applying statistical machine learning mod- ysis, enabling us to reason about systems by els to this type of problem, e.g., Liben-Nowell & Kleinberg studying the relationships between their parts. (2007); Hoff (2008); Goldenberg et al. (2010). Most often in network analysis, the edges are given. However, in many systems it is diffi- In this paper we are concerned with implicit networks that cult or impossible to measure the network di- cannot be observed directly, but about which we wish to rectly. Examples of latent networks include eco- perform analysis. In an implicit network, the vertices nomic interactions linking financial instruments or edges of the graph may not be directly observed, but and patterns of reciprocity in gang violence. In the graph structure may be inferred from noisy emissions. these cases, we are limited to noisy observations These noisy observations are assumed to have been gen- of events associated with each node. To enable erated according to underlying dynamics that respect the analysis of these implicit networks, we develop latent network structure. a probabilistic model that combines mutually- For example, trades on financial stock markets are executed exciting point processes with random graph mod- thousands of times per second. Trades of one stock are els. We show how the Poisson superposition likely to cause subsequent activity on stocks in related in- principle enables an elegant auxiliary variable dustries. How can we infer such interactions and disen- formulation and a fully-Bayesian, parallel infer- tangle them from market-wide fluctuations? Discovering ence algorithm. We evaluate this new model em- latent structure underlying financial markets not only re- pirically on several datasets. veals interpretable patterns of interaction, but also provides insight into the stability of the market. In Section 4 we will analyze the stability of mutually-excitatory systems, and in 1. Introduction Section 6 we will explore how stock similarity may be in- Many types of modern data are characterized via relation- ferred from trading activity. ships on a network. Social network analysis is the most As another example, both the edges and vertices may be commonly considered example, where the properties of in- latent. In Section 7, we examine patterns of violence in dividuals (vertices) can be inferred from “friendship” type Chicago, which can often be attributed to social structures connections (edges). Such analyses are also critical to un- in the form of gangs. We would expect that attacks from derstanding regulatory biological pathways, trade relation- one gang onto another might induce cascades of violence, ships between nations, and propagation of disease. The but the vertices (gang identity of both perpetrator and vic- tasks associated with such data may be unsupervised (e.g., tim) are unobserved. As with the financial data, it should identifying low-dimensional representations of edges or be possible to exploit dynamics to infer these social struc- vertices) or supervised (e.g., predicting unobserved links tures. In this case spatial information is available as well, in the graph). Traditionally, network analysis has focused which can help inform latent vertex identities. on explicit network problems in which the graph itself is considered to be the observed data. That is, the vertices In both of these examples, the noisy emissions have the form of events in time, or “spikes,” and our intuition is Proceedings of the 31 st International Conference on Machine that a spike at a vertex will induce activity at adjacent ver- Learning, Beijing, China, 2014. JMLR: W&CP volume 32. Copy- tices. In this paper, we formalize this idea into a probabilis- right 2014 by the author(s). Discovering Latent Network Structure in Point Process Data tic model based on mutually-interacting point processes. 1 4 Specifically, we combine the Hawkes process (Hawkes, I 1971) with recently developed exchangeable random graph priors. This combination allows us to reason about latent 3 5a 5c networks in terms of the way that they regulate interaction in the Hawkes process. Inference in the resulting model can II be done with Markov chain Monte Carlo, and an elegant 2 5b data augmentation scheme results in efficient parallelism. III 2. Preliminaries 2.1. Poisson Processes Figure 1: Illustration of a Hawkes process. Events induce impulse responses on connected processes and spawn “child” events. See Point processes are fundamental statistical objects that the main text for a complete description. f gN ⊂ S S 0 yield random finite sets of events sn n=1 , where to the intensity of other processes k . Causality and locality is a compact subset of RD, for example, space or of influence are enforced by requiring hk;k0 (∆t) to be zero time. The Poisson process is the canonical exam- for ∆t2 = [0; ∆tmax]. ple. It is governed by a nonnegative “rate” or “inten- sity” function, λ(s): S! R+. The number of events By the superposition theorem for Poisson processes, these 0 in a subsetR S ⊂ S follows a Poisson distribution with additive components can be considered independent pro- mean 0 λ(s)ds. Moreover, the number of events in dis- cesses, each giving rise to their own events. We augment S 2 f − g joint subsets are independent. our data with a latent random variable zn 0; : : : ; n 1 to indicate the cause of the n-th event (0 if the event is due f gN ∼ PP We use the notation sn n=1 (λ(s)) to indicate that to the background rate and 1 : : : n − 1 if it was caused by f gN a set of events sn n=1 is drawn from a Poisson process a preceding event). The augmented Hawkes likelihood is with rate λ(s). The likelihood is given by then the product of likelihoods of each Poisson process: { } Z YN N N f g j − f g j f g ff 0 gg p( sn n=1 λ(s)) = exp λ(s)ds λ(sn): (1) p( (sn; cn; zn) n=1 λ0;k(t) ; hk;k (∆t) ) = S n=1 YK p(fs : c = k ^ z = 0g j λ (t)) × In this work we will make use of a special property of Pois- n n n 0;k k=1 son processes, the Poisson superposition theorem, which YN YK states that fs g ∼ PP(λ (s) + ::: + λ (s)) can be de- n 1 K p(fs 0 : c 0 = k ^ z 0 = ng j h (t − s )); composed into K independent Poisson processes. Let- n n n cn;k n n=1 k=1 ting zn denote the origin of the n-th event, we perform the decomposition by independently sampling each zn where the densities in the product are given by Equation 1. from Pr(z = k) / λ (s ), for k 2 f1 :::Kg (Daley & n k n Figure 1 illustrates a causal cascades of events for a simple Vere-Jones, 1988). network of three processes (I-III). The first event is caused by the background rate (z1 = 0), and it induces impulse 2.2. Hawkes Processes responses on processes II and III. Event 2 is spawned by Though Poisson processes have many nice properties, they the impulse on the third process (z2 = 1), and feeds back cannot capture interactions between events. For this we onto processes I and II. In some cases a single parent event turn to a more general model known as Hawkes pro- induces multiple children, e.g., event 4 spawns events 5a- cesses (Hawkes, 1971). A Hawkes process consists of K c. In this simple example, processes excite one another, point processes and gives rise to sets of marked events but do not excite themselves. Next we will introduce more f gN 2 f g sophisticated models for such interaction networks. sn; cn n=1, where cn 1;:::;K specifies the process on which the n-th event occurred. For now, we assume the events are points in time, i.e., sn 2 [0;T ]. Each of 2.3. Random Graph Models the K processes is a conditionally Poisson process with Graphs of K nodes correspond to K × K matrices. a rate λ (t j fs : s < tg) that depends on the history of k n n Unweighted graphs are binary adjacency matrices A events up to time t. where Ak;k0 = 1 indicates a directed edge from node k to Hawkes processes have additive interactions. Each process node k0. Weighted directed graphs can be represented by has a “background rate” λ0;k(t), and each event sn on pro- a real matrix W whose entries indicate the weights of the cess k adds a nonnegative impulse response hk;k0 (t − sn) edges. Random graph models reflect the probability of dif- Discovering Latent Network Structure in Point Process Data ferent network structures through distributions over these plete graph recovers the standard Hawkes process. Mak- matrices. ing g a probability density endows W with units of “ex- pected number of events” and allows us to compare the Recently, many random graph models have been unified relative strength of interactions. The form suggests an under an elegant theoretical framework due to Aldous and intuitive generative model: for each impulse response Hoover (Aldous, 1981; Hoover, 1979). See Lloyd et al. draw m ∼ Poisson(W 0 ) number of induced events and (2012) for an overview.
Recommended publications
  • Cox Process Representation and Inference for Stochastic Reaction-Diffusion Processes
    Edinburgh Research Explorer Cox process representation and inference for stochastic reaction-diffusion processes Citation for published version: Schnoerr, D, Grima, R & Sanguinetti, G 2016, 'Cox process representation and inference for stochastic reaction-diffusion processes', Nature Communications, vol. 7, 11729. https://doi.org/10.1038/ncomms11729 Digital Object Identifier (DOI): 10.1038/ncomms11729 Link: Link to publication record in Edinburgh Research Explorer Document Version: Publisher's PDF, also known as Version of record Published In: Nature Communications General rights Copyright for the publications made accessible via the Edinburgh Research Explorer is retained by the author(s) and / or other copyright owners and it is a condition of accessing these publications that users recognise and abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to ensure that Edinburgh Research Explorer content complies with UK legislation. If you believe that the public display of this file breaches copyright please contact [email protected] providing details, and we will remove access to the work immediately and investigate your claim. Download date: 02. Oct. 2021 ARTICLE Received 16 Dec 2015 | Accepted 26 Apr 2016 | Published 25 May 2016 DOI: 10.1038/ncomms11729 OPEN Cox process representation and inference for stochastic reaction–diffusion processes David Schnoerr1,2,3, Ramon Grima1,3 & Guido Sanguinetti2,3 Complex behaviour in many systems arises from the stochastic interactions of spatially distributed particles or agents. Stochastic reaction–diffusion processes are widely used to model such behaviour in disciplines ranging from biology to the social sciences, yet they are notoriously difficult to simulate and calibrate to observational data.
    [Show full text]
  • Poisson Representations of Branching Markov and Measure-Valued
    The Annals of Probability 2011, Vol. 39, No. 3, 939–984 DOI: 10.1214/10-AOP574 c Institute of Mathematical Statistics, 2011 POISSON REPRESENTATIONS OF BRANCHING MARKOV AND MEASURE-VALUED BRANCHING PROCESSES By Thomas G. Kurtz1 and Eliane R. Rodrigues2 University of Wisconsin, Madison and UNAM Representations of branching Markov processes and their measure- valued limits in terms of countable systems of particles are con- structed for models with spatially varying birth and death rates. Each particle has a location and a “level,” but unlike earlier con- structions, the levels change with time. In fact, death of a particle occurs only when the level of the particle crosses a specified level r, or for the limiting models, hits infinity. For branching Markov pro- cesses, at each time t, conditioned on the state of the process, the levels are independent and uniformly distributed on [0,r]. For the limiting measure-valued process, at each time t, the joint distribu- tion of locations and levels is conditionally Poisson distributed with mean measure K(t) × Λ, where Λ denotes Lebesgue measure, and K is the desired measure-valued process. The representation simplifies or gives alternative proofs for a vari- ety of calculations and results including conditioning on extinction or nonextinction, Harris’s convergence theorem for supercritical branch- ing processes, and diffusion approximations for processes in random environments. 1. Introduction. Measure-valued processes arise naturally as infinite sys- tem limits of empirical measures of finite particle systems. A number of ap- proaches have been developed which preserve distinct particles in the limit and which give a representation of the measure-valued process as a transfor- mation of the limiting infinite particle system.
    [Show full text]
  • Modern Discrete Probability I
    Preliminaries Some fundamental models A few more useful facts about... Modern Discrete Probability I - Introduction Stochastic processes on graphs: models and questions Sebastien´ Roch UW–Madison Mathematics September 6, 2017 Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... 1 Preliminaries Review of graph theory Review of Markov chain theory 2 Some fundamental models Random walks on graphs Percolation Some random graph models Markov random fields Interacting particles on finite graphs 3 A few more useful facts about... ...graphs ...Markov chains ...other things Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... Graphs Definition (Undirected graph) An undirected graph (or graph for short) is a pair G = (V ; E) where V is the set of vertices (or nodes, sites) and E ⊆ ffu; vg : u; v 2 V g; is the set of edges (or bonds). The V is either finite or countably infinite. Edges of the form fug are called loops. We do not allow E to be a multiset. We occasionally write V (G) and E(G) for the vertices and edges of G. Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about... An example: the Petersen graph Sebastien´ Roch, UW–Madison Modern Discrete Probability – Models and Questions Preliminaries Review of graph theory Some fundamental models Review of Markov chain theory A few more useful facts about..
    [Show full text]
  • Local Conditioning in Dawson–Watanabe Superprocesses
    The Annals of Probability 2013, Vol. 41, No. 1, 385–443 DOI: 10.1214/11-AOP702 c Institute of Mathematical Statistics, 2013 LOCAL CONDITIONING IN DAWSON–WATANABE SUPERPROCESSES By Olav Kallenberg Auburn University Consider a locally finite Dawson–Watanabe superprocess ξ =(ξt) in Rd with d ≥ 2. Our main results include some recursive formulas for the moment measures of ξ, with connections to the uniform Brown- ian tree, a Brownian snake representation of Palm measures, continu- ity properties of conditional moment densities, leading by duality to strongly continuous versions of the multivariate Palm distributions, and a local approximation of ξt by a stationary clusterη ˜ with nice continuity and scaling properties. This all leads up to an asymptotic description of the conditional distribution of ξt for a fixed t> 0, given d that ξt charges the ε-neighborhoods of some points x1,...,xn ∈ R . In the limit as ε → 0, the restrictions to those sets are conditionally in- dependent and given by the pseudo-random measures ξ˜ orη ˜, whereas the contribution to the exterior is given by the Palm distribution of ξt at x1,...,xn. Our proofs are based on the Cox cluster representa- tions of the historical process and involve some delicate estimates of moment densities. 1. Introduction. This paper may be regarded as a continuation of [19], where we considered some local properties of a Dawson–Watanabe super- process (henceforth referred to as a DW-process) at a fixed time t> 0. Recall that a DW-process ξ = (ξt) is a vaguely continuous, measure-valued diffu- d ξtf µvt sion process in R with Laplace functionals Eµe− = e− for suitable functions f 0, where v = (vt) is the unique solution to the evolution equa- 1 ≥ 2 tion v˙ = 2 ∆v v with initial condition v0 = f.
    [Show full text]
  • POISSON PROCESSES 1.1. the Rutherford-Chadwick-Ellis
    POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS 1.1. The Rutherford-Chadwick-Ellis Experiment. About 90 years ago Ernest Rutherford and his collaborators at the Cavendish Laboratory in Cambridge conducted a series of pathbreaking experiments on radioactive decay. In one of these, a radioactive substance was observed in N = 2608 time intervals of 7.5 seconds each, and the number of decay particles reaching a counter during each period was recorded. The table below shows the number Nk of these time periods in which exactly k decays were observed for k = 0,1,2,...,9. Also shown is N pk where k pk = (3.87) exp 3.87 =k! {− g The parameter value 3.87 was chosen because it is the mean number of decays/period for Rutherford’s data. k Nk N pk k Nk N pk 0 57 54.4 6 273 253.8 1 203 210.5 7 139 140.3 2 383 407.4 8 45 67.9 3 525 525.5 9 27 29.2 4 532 508.4 10 16 17.1 5 408 393.5 ≥ This is typical of what happens in many situations where counts of occurences of some sort are recorded: the Poisson distribution often provides an accurate – sometimes remarkably ac- curate – fit. Why? 1.2. Poisson Approximation to the Binomial Distribution. The ubiquity of the Poisson distri- bution in nature stems in large part from its connection to the Binomial and Hypergeometric distributions. The Binomial-(N ,p) distribution is the distribution of the number of successes in N independent Bernoulli trials, each with success probability p.
    [Show full text]
  • Cox Process Functional Learning Gérard Biau, Benoît Cadre, Quentin Paris
    Cox process functional learning Gérard Biau, Benoît Cadre, Quentin Paris To cite this version: Gérard Biau, Benoît Cadre, Quentin Paris. Cox process functional learning. Statistical Inference for Stochastic Processes, Springer Verlag, 2015, 18 (3), pp.257-277. 10.1007/s11203-015-9115-z. hal- 00820838 HAL Id: hal-00820838 https://hal.archives-ouvertes.fr/hal-00820838 Submitted on 6 May 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Cox Process Learning G´erard Biau Universit´ePierre et Marie Curie1 & Ecole Normale Sup´erieure2, France [email protected] Benoˆıt Cadre IRMAR, ENS Cachan Bretagne, CNRS, UEB, France3 [email protected] Quentin Paris IRMAR, ENS Cachan Bretagne, CNRS, UEB, France [email protected] Abstract This article addresses the problem of supervised classification of Cox process trajectories, whose random intensity is driven by some exoge- nous random covariable. The classification task is achieved through a regularized convex empirical risk minimization procedure, and a nonasymptotic oracle inequality is derived. We show that the algo- rithm provides a Bayes-risk consistent classifier. Furthermore, it is proved that the classifier converges at a rate which adapts to the un- known regularity of the intensity process.
    [Show full text]
  • Spatio-Temporal Cluster Detection and Local Moran Statistics of Point Processes
    Old Dominion University ODU Digital Commons Mathematics & Statistics Theses & Dissertations Mathematics & Statistics Spring 2019 Spatio-Temporal Cluster Detection and Local Moran Statistics of Point Processes Jennifer L. Matthews Old Dominion University Follow this and additional works at: https://digitalcommons.odu.edu/mathstat_etds Part of the Applied Statistics Commons, and the Biostatistics Commons Recommended Citation Matthews, Jennifer L.. "Spatio-Temporal Cluster Detection and Local Moran Statistics of Point Processes" (2019). Doctor of Philosophy (PhD), Dissertation, Mathematics & Statistics, Old Dominion University, DOI: 10.25777/3mps-rk62 https://digitalcommons.odu.edu/mathstat_etds/46 This Dissertation is brought to you for free and open access by the Mathematics & Statistics at ODU Digital Commons. It has been accepted for inclusion in Mathematics & Statistics Theses & Dissertations by an authorized administrator of ODU Digital Commons. For more information, please contact [email protected]. ABSTRACT Approved for public release; distribution is unlimited SPATIO-TEMPORAL CLUSTER DETECTION AND LOCAL MORAN STATISTICS OF POINT PROCESSES Jennifer L. Matthews Commander, United States Navy Old Dominion University, 2019 Director: Dr. Norou Diawara Moran's index is a statistic that measures spatial dependence, quantifying the degree of dispersion or clustering of point processes and events in some location/area. Recognizing that a single Moran's index may not give a sufficient summary of the spatial autocorrelation measure, a local
    [Show full text]
  • Log Gaussian Cox Processes
    Log Gaussian Cox processes by Jesper Møller, Anne Randi Syversveen, and Rasmus Plenge Waagepetersen. Log Gaussian Cox processes JESPER MØLLER Aalborg University ANNE RANDI SYVERSVEEN The Norwegian University of Science and Technology RASMUS PLENGE WAAGEPETERSEN University of Aarhus ABSTRACT. Planar Cox processes directed by a log Gaussian intensity process are investigated in the univariate and multivariate cases. The appealing properties of such models are demonstrated theoretically as well as through data examples and simulations. In particular, the first, second and third-order properties are studied and utilized in the statistical analysis of clustered point patterns. Also empirical Bayesian inference for the underlying intensity surface is considered. Key words: empirical Bayesian inference; ergodicity; Markov chain Monte Carlo; Metropolis-adjusted Langevin algorithm; multivariate Cox processes; Neyman-Scott processes; pair correlation function; parameter estimation; spatial point processes; third- order properties. AMS 1991 subject classification: Primary 60G55, 62M30. Secondary 60D05. 1 Introduction Cox processes provide useful and frequently applied models for aggregated spatial point patterns where the aggregation is due to a stochastic environmental heterogeneity, see e.g. Diggle (1983), Cressie (1993), Stoyan et al. (1995), and the references therein. A Cox process is ’doubly stochastic’ as it arises as an inhomogeneous Poisson process with a random intensity measure. The random intensity measure is often specified by a random intensity function or as we prefer to call it an intensity process or surface. There may indeed be other sources of aggregation in a spatial point pattern than spatial heterogeneity. Cluster processes is a well-known class of models where clusters are generated by an unseen point process, cf.
    [Show full text]
  • Markov Chain Monte Carlo Estimation of Exponential Random Graph Models
    Markov Chain Monte Carlo Estimation of Exponential Random Graph Models Tom A.B. Snijders ICS, Department of Statistics and Measurement Theory University of Groningen ∗ April 19, 2002 ∗Author's address: Tom A.B. Snijders, Grote Kruisstraat 2/1, 9712 TS Groningen, The Netherlands, email <[email protected]>. I am grateful to Paul Snijders for programming the JAVA applet used in this article. In the revision of this article, I profited from discussions with Pip Pattison and Garry Robins, and from comments made by a referee. This paper is formatted in landscape to improve on-screen readability. It is read best by opening Acrobat Reader in a full screen window. Note that in Acrobat Reader, the entire screen can be used for viewing by pressing Ctrl-L; the usual screen is returned when pressing Esc; it is possible to zoom in or zoom out by pressing Ctrl-- or Ctrl-=, respectively. The sign in the upper right corners links to the page viewed previously. ( Tom A.B. Snijders 2 MCMC estimation for exponential random graphs ( Abstract bility to move from one region to another. In such situations, convergence to the target distribution is extremely slow. To This paper is about estimating the parameters of the exponential be useful, MCMC algorithms must be able to make transitions random graph model, also known as the p∗ model, using frequen- from a given graph to a very different graph. It is proposed to tist Markov chain Monte Carlo (MCMC) methods. The exponen- include transitions to the graph complement as updating steps tial random graph model is simulated using Gibbs or Metropolis- to improve the speed of convergence to the target distribution.
    [Show full text]
  • Exploring Healing Strategies for Random Boolean Networks
    Exploring Healing Strategies for Random Boolean Networks Christian Darabos 1, Alex Healing 2, Tim Johann 3, Amitabh Trehan 4, and Am´elieV´eron 5 1 Information Systems Department, University of Lausanne, Switzerland 2 Pervasive ICT Research Centre, British Telecommunications, UK 3 EML Research gGmbH, Heidelberg, Germany 4 Department of Computer Science, University of New Mexico, Albuquerque, USA 5 Division of Bioinformatics, Institute for Evolution and Biodiversity, The Westphalian Wilhelms University of Muenster, Germany. [email protected], [email protected], [email protected], [email protected], [email protected] Abstract. Real-world systems are often exposed to failures where those studied theoretically are not: neuron cells in the brain can die or fail, re- sources in a peer-to-peer network can break down or become corrupt, species can disappear from and environment, and so forth. In all cases, for the system to keep running as it did before the failure occurred, or to survive at all, some kind of healing strategy may need to be applied. As an example of such a system subjected to failure and subsequently heal, we study Random Boolean Networks. Deletion of a node in the network was considered a failure if it affected the functional output and we in- vestigated healing strategies that would allow the system to heal either fully or partially. More precisely, our main strategy involves allowing the nodes directly affected by the node deletion (failure) to iteratively rewire in order to achieve healing. We found that such a simple method was ef- fective when dealing with small networks and single-point failures.
    [Show full text]
  • Percolation Threshold Results on Erdos-Rényi Graphs
    arXiv: arXiv:0000.0000 Percolation Threshold Results on Erd}os-R´enyi Graphs: an Empirical Process Approach Michael J. Kane Abstract: Keywords and phrases: threshold, directed percolation, stochastic ap- proximation, empirical processes. 1. Introduction Random graphs and discrete random processes provide a general approach to discovering properties and characteristics of random graphs and randomized al- gorithms. The approach generally works by defining an algorithm on a random graph or a randomized algorithm. Then, expected changes for each step of the process are used to propose a limiting differential equation and a large deviation theorem is used to show that the process and the differential equation are close in some sense. In this way a connection is established between the resulting process's stochastic behavior and the dynamics of a deterministic, asymptotic approximation using a differential equation. This approach is generally referred to as stochastic approximation and provides a powerful tool for understanding the asymptotic behavior of a large class of processes defined on random graphs. However, little work has been done in the area of random graph research to investigate the weak limit behavior of these processes before the asymptotic be- havior overwhelms the random component of the process. This context is par- ticularly relevant to researchers studying news propagation in social networks, sensor networks, and epidemiological outbreaks. In each of these applications, investigators may deal graphs containing tens to hundreds of vertices and be interested not only in expected behavior over time but also error estimates. This paper investigates the connectivity of graphs, with emphasis on Erd}os- R´enyi graphs, near the percolation threshold when the number of vertices is not asymptotically large.
    [Show full text]
  • Probability on Graphs Random Processes on Graphs and Lattices
    Probability on Graphs Random Processes on Graphs and Lattices GEOFFREY GRIMMETT Statistical Laboratory University of Cambridge c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Geoffrey Grimmett Statistical Laboratory Centre for Mathematical Sciences University of Cambridge Wilberforce Road Cambridge CB3 0WB United Kingdom 2000 MSC: (Primary) 60K35, 82B20, (Secondary) 05C80, 82B43, 82C22 With 44 Figures c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Contents Preface ix 1 Random walks on graphs 1 1.1 RandomwalksandreversibleMarkovchains 1 1.2 Electrical networks 3 1.3 Flowsandenergy 8 1.4 Recurrenceandresistance 11 1.5 Polya's theorem 14 1.6 Graphtheory 16 1.7 Exercises 18 2 Uniform spanning tree 21 2.1 De®nition 21 2.2 Wilson's algorithm 23 2.3 Weak limits on lattices 28 2.4 Uniform forest 31 2.5 Schramm±LownerevolutionsÈ 32 2.6 Exercises 37 3 Percolation and self-avoiding walk 39 3.1 Percolationandphasetransition 39 3.2 Self-avoiding walks 42 3.3 Coupledpercolation 45 3.4 Orientedpercolation 45 3.5 Exercises 48 4 Association and in¯uence 50 4.1 Holley inequality 50 4.2 FKGinequality 53 4.3 BK inequality 54 4.4 Hoeffdinginequality 56 c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 vi Contents 4.5 In¯uenceforproductmeasures 58 4.6 Proofsofin¯uencetheorems 63 4.7 Russo'sformulaandsharpthresholds 75 4.8 Exercises 78 5 Further percolation 81 5.1 Subcritical phase 81 5.2 Supercritical phase 86 5.3 Uniquenessofthein®nitecluster 92 5.4 Phase transition 95 5.5 Openpathsinannuli 99 5.6 The critical probability in two dimensions 103 5.7 Cardy's formula 110 5.8 The
    [Show full text]