Markov Random Field Models of Multicasting in Tree Networks

Total Page:16

File Type:pdf, Size:1020Kb

Markov Random Field Models of Multicasting in Tree Networks Markov Random Field Models of Multicasting in Tree Networks Author(s): Kavita Ramanan, Anirvan Sengupta, Ilze Ziedins and Partha Mitra Source: Advances in Applied Probability, Vol. 34, No. 1 (Mar., 2002), pp. 58-84 Published by: Applied Probability Trust Stable URL: http://www.jstor.org/stable/1428372 . Accessed: 18/09/2013 14:55 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Applied Probability Trust is collaborating with JSTOR to digitize, preserve and extend access to Advances in Applied Probability. http://www.jstor.org This content downloaded from 128.148.231.12 on Wed, 18 Sep 2013 14:55:03 PM All use subject to JSTOR Terms and Conditions Adv.Appl. Prob. 34, 58-84 (2002) Printed in NorthernIreland @ AppliedProbability Trust 2002 MARKOV RANDOM FIELD MODELS OF MULTICASTING IN TREE NETWORKS KAVITARAMANAN * ** AND ANIRVANSENGUPTA,* Bell Laboratories,Lucent Technologies ILZE ZIEDINS,***University ofAuckland PARTHAMITRA,* Bell Laboratories,Lucent Technologies Abstract In this paper,we analyse a model of a regulartree loss networkthat supportstwo types of calls: unicast calls that requireunit capacity on a single link, and multicastcalls that require unit capacity on every link emanating from a node. We study the behaviour of the distributionof calls in the core of a large networkthat has uniform unicast and multicastarrival rates. At sufficientlyhigh multicastcall arrivalrates the networkexhibits a 'phasetransition', leading to unfairnessdue to spatialvariation in the multicastblocking probabilities. We study the dependenceof the phase transitionon unicast arrivalrates, the coordinationnumber of the network, and the parity of the capacity of edges in the network. Numericalresults suggest that the natureof phase transitionsis qualitatively differentwhen thereare odd and even capacitieson the links. These phenomenaare seen to persist even with the introductionof nonuniformarrival rates and multihopmulticast calls into the network. Finally, we also show the inadequacyof approximationssuch as the Erlangfixed-point approximations when multicastingis present. Keywords:Multicasting; unicasting; broadcasting; Cayley trees; Bethe lattice; loss networks;blocking probabilities;phase transitions;Erlang fixed-point approximations; reduced load approximations;statistical mechanics; spin models; hard core models; symmetrybreaking AMS 2000 SubjectClassification: Primary 60G60; 90B 15 Secondary68M20; 60K35 1. Introduction In this paper, we present and analyse a model of multicast and unicast calls arriving at an idealized loss network which has the form of a symmetric tree. Multicasting arises in both queueing and loss networks. Instead of having a simple end-to- end connection, a transmission is made to a group of individuals from a single site. An example of this in the loss network setting is a conference call. In a queueing setting this arises, for instance, when multiple copies of a message on the Internet are broadcast from one person to a number of sites via a mailing list distribution. In recent years there has been a growing interest in multicasting applications that require reliable data delivery such as software distribution and news broadcasts. This has given rise to several interesting problems. One major problem is the issue of how to construct connections between individuals in a way that makes communication Received 4 December2000; revision received 2 January2002. * Postal address:Bell Laboratories,Lucent Technologies, 600 MountainAvenue, Murray Hill, NJ 07974, USA. ** Email address:[email protected] ***Postal address:Department of Statistics,University of Auckland,Private Bag 92019, Auckland,New Zealand. 58 This content downloaded from 128.148.231.12 on Wed, 18 Sep 2013 14:55:03 PM All use subject to JSTOR Terms and Conditions Markovrandom field modelsof multicastingin treenetworks 59 efficient and reliable and minimizes demands on network resources. Much effort has been expended on this [1], [6], [16]. Before this can be answered, however, we need to be able to analyse networkperformance in the presence of multicastcalls, a task that is considerably more complex than the performanceanalysis of a purely unicast network. In this paper we study performancein the context of a symmetric loss networkin the form of a tree. This is a very simple networkwhich neverthelessexhibits interestingbehaviour, and raises questions aboutthe performanceanalysis of more complicatednetworks using traditionaltechniques and approximations.We expect thatthe insightgleaned here will be useful in moregeneral contexts, especially in view of the fact that many of the qualitativeproperties observed in the case of a symmetricnetwork persist even when asymmetryis introduced. A general loss network without controls can be described as follows. Let J be the finite collection of resources(or links) in the network,and C = {Cj, j E J} where Cj is the capacity of resource j. Let R denote the finite set of possible call (or customer)types in the network. Calls of each type r E R arriveas a Poisson process of rate Vrand have identicallydistributed holding or service times that we assume, without loss of generality,to have mean 1 (see [5]). Each call of type r requirescapacity Ajr from resource j E J for the durationof its holding time. If one or more of the resources(links) j for which Ajr > 0 does not have sufficientfree capacityto carrythe call of type r, then the call is blocked and consideredlost. Otherwisethe call is accepted. All arrivalprocesses and holding times are assumedto be independentof one another.One of the standardmeasures of performancefor a loss system is the probabilitythat a call is blocked or lost (the loss, or blocking probability). Loss networkshave been widely studied, as models of circuit-switchedand ATM networks, and in other contexts. Excellent introductionsto this general model are found in [13] and [17]. We confine our attentionin this paper to a special kind of symmetricloss networkwith a regulartree structure.Let Toobe an infinite tree with q + 1 edges emanatingfrom each node. Let SL be the finite sphericalsubtree of radiusL which consists of a single centralnode and all nodes no more thana distanceL from it, where the distancebetween two nodes is the minimum numberof edges in any pathconnecting the two nodes. There are q + 1 edges emanatingfrom each node in SL except the terminalnodes, which are attachedto just one edge. We assume that each edge has the same capacity,C. We let E8(T)denote the set of edges in a given tree T and use 8 (t) to denote the set of edges that are incidentwith the node t. For each node t E T, we define XA(t) := {t' E T : tt' e 8(t)} to be the set of immediateneighbours of t. We allow two types of calls in the network. Single link calls, requiringcapacity on just a single edge, arriveat each edge as a Poisson process of rate k. Single link calls connect two neighbouring nodes. Multicast calls, on the other hand, are centred at a node, and connect that node to all of its immediateneighbours. Thus a call centredat an internalnode t e SL requirescapacity on each of the q + 1 edges in JN(t), while a call centredat a terminalnode requirescapacity on just one edge. We assume that multicastcalls arriveat each node t as a Poisson process of rate vt = v. We will assume that all calls requirejust a single unit of capacity on each edge of their connection. Throughoutthis paperwe will be concentratingon assessing the blocking probabilitiesfor large sphericaltree networks. Much of the treatmentbelow can be applied in the more complex multiratesetting, but in orderto clarify the presentationwe do not do so here. Althoughthe tree networkthat we have describedhere is clearlyunrealistic, it is nevertheless a useful tool for studying some of the behavioursthat multicastingmight introduceto a loss network.We show below thatmulticasting can introduceunfairness in the symmetricspherical tree that we study, as the blocking probabilitiesseen by the same call types at differentnodes This content downloaded from 128.148.231.12 on Wed, 18 Sep 2013 14:55:03 PM All use subject to JSTOR Terms and Conditions 60 K. RAMANANETAL. may vary considerably. These differences are a consequence of a phase transitioneffect that arises when the arrivalrate for multicast calls is sufficientlyhigh and that of unicast calls is sufficiently low. Thus we observe that the presence of unicast calls assists in maintaining fairness and homogeneity of the blocking probabilitiesin the system. In addition, we study the trade-offbetween unicast and multicastcall arrivalrates for a given fixed unicastblocking probability.Finally, we also shed some light on approximationsof networkswith multicasting. A commonly made assumptionwhen calculating approximateblocking probabilitiesis that resources (edges or links) behave as though they are blocking independentlyof one another. This assumptionunderpins approximations such as the Erlangfixed-point approximation, also knownas the reducedload approximation,and refinements of it-approximations thathave been widely andvery successfullyused, even whereit is clearthat the
Recommended publications
  • Methods of Monte Carlo Simulation II
    Methods of Monte Carlo Simulation II Ulm University Institute of Stochastics Lecture Notes Dr. Tim Brereton Summer Term 2014 Ulm, 2014 2 Contents 1 SomeSimpleStochasticProcesses 7 1.1 StochasticProcesses . 7 1.2 RandomWalks .......................... 7 1.2.1 BernoulliProcesses . 7 1.2.2 RandomWalks ...................... 10 1.2.3 ProbabilitiesofRandomWalks . 13 1.2.4 Distribution of Xn .................... 13 1.2.5 FirstPassageTime . 14 2 Estimators 17 2.1 Bias, Variance, the Central Limit Theorem and Mean Square Error................................ 19 2.2 Non-AsymptoticErrorBounds. 22 2.3 Big O and Little o Notation ................... 23 3 Markov Chains 25 3.1 SimulatingMarkovChains . 28 3.1.1 Drawing from a Discrete Uniform Distribution . 28 3.1.2 Drawing From A Discrete Distribution on a Small State Space ........................... 28 3.1.3 SimulatingaMarkovChain . 28 3.2 Communication .......................... 29 3.3 TheStrongMarkovProperty . 30 3.4 RecurrenceandTransience . 31 3.4.1 RecurrenceofRandomWalks . 33 3.5 InvariantDistributions . 34 3.6 LimitingDistribution. 36 3.7 Reversibility............................ 37 4 The Poisson Process 39 4.1 Point Processes on [0, )..................... 39 ∞ 3 4 CONTENTS 4.2 PoissonProcess .......................... 41 4.2.1 Order Statistics and the Distribution of Arrival Times 44 4.2.2 DistributionofArrivalTimes . 45 4.3 SimulatingPoissonProcesses. 46 4.3.1 Using the Infinitesimal Definition to Simulate Approx- imately .......................... 46 4.3.2 SimulatingtheArrivalTimes . 47 4.3.3 SimulatingtheInter-ArrivalTimes . 48 4.4 InhomogenousPoissonProcesses. 48 4.5 Simulating an Inhomogenous Poisson Process . 49 4.5.1 Acceptance-Rejection. 49 4.5.2 Infinitesimal Approach (Approximate) . 50 4.6 CompoundPoissonProcesses . 51 5 ContinuousTimeMarkovChains 53 5.1 TransitionFunction. 53 5.2 InfinitesimalGenerator . 54 5.3 ContinuousTimeMarkovChains .
    [Show full text]
  • 12 : Conditional Random Fields 1 Hidden Markov Model
    10-708: Probabilistic Graphical Models 10-708, Spring 2014 12 : Conditional Random Fields Lecturer: Eric P. Xing Scribes: Qin Gao, Siheng Chen 1 Hidden Markov Model 1.1 General parametric form In hidden Markov model (HMM), we have three sets of parameters, j i transition probability matrix A : p(yt = 1jyt−1 = 1) = ai;j; initialprobabilities : p(y1) ∼ Multinomial(π1; π2; :::; πM ); i emission probabilities : p(xtjyt) ∼ Multinomial(bi;1; bi;2; :::; bi;K ): 1.2 Inference k k The inference can be done with forward algorithm which computes αt ≡ µt−1!t(k) = P (x1; :::; xt−1; xt; yt = 1) recursively by k k X i αt = p(xtjyt = 1) αt−1ai;k; (1) i k k and the backward algorithm which computes βt ≡ µt t+1(k) = P (xt+1; :::; xT jyt = 1) recursively by k X i i βt = ak;ip(xt+1jyt+1 = 1)βt+1: (2) i Another key quantity is the conditional probability of any hidden state given the entire sequence, which can be computed by the dot product of forward message and backward message by, i i i i X i;j γt = p(yt = 1jx1:T ) / αtβt = ξt ; (3) j where we define, i;j i j ξt = p(yt = 1; yt−1 = 1; x1:T ); i j / µt−1!t(yt = 1)µt t+1(yt+1 = 1)p(xt+1jyt+1)p(yt+1jyt); i j i = αtβt+1ai;jp(xt+1jyt+1 = 1): The implementation in Matlab can be vectorized by using, i Bt(i) = p(xtjyt = 1); j i A(i; j) = p(yt+1 = 1jyt = 1): 1 2 12 : Conditional Random Fields The relation of those quantities can be simply written in pseudocode as, T αt = (A αt−1): ∗ Bt; βt = A(βt+1: ∗ Bt+1); T ξt = (αt(βt+1: ∗ Bt+1) ): ∗ A; γt = αt: ∗ βt: 1.3 Learning 1.3.1 Supervised Learning The supervised learning is trivial if only we know the true state path.
    [Show full text]
  • Pdf File of Second Edition, January 2018
    Probability on Graphs Random Processes on Graphs and Lattices Second Edition, 2018 GEOFFREY GRIMMETT Statistical Laboratory University of Cambridge copyright Geoffrey Grimmett Geoffrey Grimmett Statistical Laboratory Centre for Mathematical Sciences University of Cambridge Wilberforce Road Cambridge CB3 0WB United Kingdom 2000 MSC: (Primary) 60K35, 82B20, (Secondary) 05C80, 82B43, 82C22 With 56 Figures copyright Geoffrey Grimmett Contents Preface ix 1 Random Walks on Graphs 1 1.1 Random Walks and Reversible Markov Chains 1 1.2 Electrical Networks 3 1.3 FlowsandEnergy 8 1.4 RecurrenceandResistance 11 1.5 Polya's Theorem 14 1.6 GraphTheory 16 1.7 Exercises 18 2 Uniform Spanning Tree 21 2.1 De®nition 21 2.2 Wilson's Algorithm 23 2.3 Weak Limits on Lattices 28 2.4 Uniform Forest 31 2.5 Schramm±LownerEvolutionsÈ 32 2.6 Exercises 36 3 Percolation and Self-Avoiding Walks 39 3.1 PercolationandPhaseTransition 39 3.2 Self-Avoiding Walks 42 3.3 ConnectiveConstantoftheHexagonalLattice 45 3.4 CoupledPercolation 53 3.5 Oriented Percolation 53 3.6 Exercises 56 4 Association and In¯uence 59 4.1 Holley Inequality 59 4.2 FKG Inequality 62 4.3 BK Inequalitycopyright Geoffrey Grimmett63 vi Contents 4.4 HoeffdingInequality 65 4.5 In¯uenceforProductMeasures 67 4.6 ProofsofIn¯uenceTheorems 72 4.7 Russo'sFormulaandSharpThresholds 80 4.8 Exercises 83 5 Further Percolation 86 5.1 Subcritical Phase 86 5.2 Supercritical Phase 90 5.3 UniquenessoftheIn®niteCluster 96 5.4 Phase Transition 99 5.5 OpenPathsinAnnuli 103 5.6 The Critical Probability in Two Dimensions 107
    [Show full text]
  • Mean Field Methods for Classification with Gaussian Processes
    Mean field methods for classification with Gaussian processes Manfred Opper Neural Computing Research Group Division of Electronic Engineering and Computer Science Aston University Birmingham B4 7ET, UK. opperm~aston.ac.uk Ole Winther Theoretical Physics II, Lund University, S6lvegatan 14 A S-223 62 Lund, Sweden CONNECT, The Niels Bohr Institute, University of Copenhagen Blegdamsvej 17, 2100 Copenhagen 0, Denmark winther~thep.lu.se Abstract We discuss the application of TAP mean field methods known from the Statistical Mechanics of disordered systems to Bayesian classifi­ cation models with Gaussian processes. In contrast to previous ap­ proaches, no knowledge about the distribution of inputs is needed. Simulation results for the Sonar data set are given. 1 Modeling with Gaussian Processes Bayesian models which are based on Gaussian prior distributions on function spaces are promising non-parametric statistical tools. They have been recently introduced into the Neural Computation community (Neal 1996, Williams & Rasmussen 1996, Mackay 1997). To give their basic definition, we assume that the likelihood of the output or target variable T for a given input s E RN can be written in the form p(Tlh(s)) where h : RN --+ R is a priori assumed to be a Gaussian random field. If we assume fields with zero prior mean, the statistics of h is entirely defined by the second order correlations C(s, S') == E[h(s)h(S')], where E denotes expectations 310 M Opper and 0. Winther with respect to the prior. Interesting examples are C(s, s') (1) C(s, s') (2) The choice (1) can be motivated as a limit of a two-layered neural network with infinitely many hidden units with factorizable input-hidden weight priors (Williams 1997).
    [Show full text]
  • MCMC Learning (Slides)
    Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions MCMC Learning Varun Kanade Elchanan Mossel UC Berkeley UC Berkeley August 30, 2013 Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Outline Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Uniform Distribution Learning • Unknown target function f : {−1; 1gn ! {−1; 1g from some class C • Uniform distribution over {−1; 1gn • Random Examples: Monotone Decision Trees [OS06] • Random Walk: DNF expressions [BMOS03] • Membership Query: DNF, TOP [J95] • Main Tool: Discrete Fourier Analysis X Y f (x) = f^(S)χS (x); χS (x) = xi S⊆[n] i2S • Can utilize sophisticated results: hypercontractivity, invariance, etc. • Connections to cryptography, hardness, de-randomization etc. • Unfortunately, too much of an idealization. In practice, variables are correlated. Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Markov Random Fields • Graph G = ([n]; E). Each node takes some value in finite set A. n • Distribution over A : (for φC non-negative, Z normalization constant) 1 Y Pr((σ ) ) = φ ((σ ) ) v v2[n] Z C v v2C clique C Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Markov Random Fields • MRFs widely used in vision, computational biology, biostatistics etc. • Extensive Algorithmic Theory for sampling from MRFs, recovering parameters and structures • Learning Question: Given f : An ! {−1; 1g. (How) Can we learn with respect to MRF distribution? • Can we utilize the structure of the MRF to aid in learning? Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Learning Model • Let M be a MRF with distribution π and f : An ! {−1; 1g the target function • Learning algorithm gets i.i.d.
    [Show full text]
  • Markov Random Fields: – Geman, Geman, “Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images”, IEEE PAMI 6, No
    MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Conglin Lu Medical Image Display and Analysis Group University of North Carolina, Chapel Hill MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Outline: Background; Definition and properties of MRF; Computation; MRF m-reps models. MarkovMarkov RandomRandom FieldsFields Model a large collection of random variables with complex dependency relationships among them. MarkovMarkov RandomRandom FieldsFields • A model based approach; • Has been applied to a variety of problems: - Speech recognition - Natural language processing - Coding - Image analysis - Neural networks - Artificial intelligence • Usually used within the Bayesian framework. TheThe BayesianBayesian ParadigmParadigm X = space of the unknown variables, e.g. labels; Y = space of data (observations), e.g. intensity values; Given an observation y∈Y, want to make inference about x∈X. TheThe BayesianBayesian ParadigmParadigm Prior PX : probability distribution on X; Likelihood PY|X : conditional distribution of Y given X; Statistical inference is based on the posterior distribution PX|Y ∝ PX •PY|X . TheThe PriorPrior DistributionDistribution • Describes our assumption or knowledge about the model; • X is usually a high dimensional space. PX describes the joint distribution of a large number of random variables; • How do we define PX? MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Outline: 9 Background; Definition and properties of MRF; Computation; MRF m-reps models. AssumptionsAssumptions •X = {Xs}s∈S, where each Xs is a random variable; S is an index set and is finite; • There is a common state space R:Xs∈R for all s ∈ S; | R | is finite; • Let Ω = {ω=(x , ..., x ): x ∈ , 1≤i≤N} be s1 sN si R the set of all possible configurations.
    [Show full text]
  • Anomaly Detection Based on Wavelet Domain GARCH Random Field Modeling Amir Noiboar and Israel Cohen, Senior Member, IEEE
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 45, NO. 5, MAY 2007 1361 Anomaly Detection Based on Wavelet Domain GARCH Random Field Modeling Amir Noiboar and Israel Cohen, Senior Member, IEEE Abstract—One-dimensional Generalized Autoregressive Con- Markov noise. It is also claimed that objects in imagery create a ditional Heteroscedasticity (GARCH) model is widely used for response over several scales in a multiresolution representation modeling financial time series. Extending the GARCH model to of an image, and therefore, the wavelet transform can serve as multiple dimensions yields a novel clutter model which is capable of taking into account important characteristics of a wavelet-based a means for computing a feature set for input to a detector. In multiscale feature space, namely heavy-tailed distributions and [17], a multiscale wavelet representation is utilized to capture innovations clustering as well as spatial and scale correlations. We periodical patterns of various period lengths, which often ap- show that the multidimensional GARCH model generalizes the pear in natural clutter images. In [12], the orientation and scale casual Gauss Markov random field (GMRF) model, and we de- selectivity of the wavelet transform are related to the biological velop a multiscale matched subspace detector (MSD) for detecting anomalies in GARCH clutter. Experimental results demonstrate mechanisms of the human visual system and are utilized to that by using a multiscale MSD under GARCH clutter modeling, enhance mammographic features. rather than GMRF clutter modeling, a reduced false-alarm rate Statistical models for clutter and anomalies are usually can be achieved without compromising the detection rate. related to the Gaussian distribution due to its mathematical Index Terms—Anomaly detection, Gaussian Markov tractability.
    [Show full text]
  • Probability on Graphs Random Processes on Graphs and Lattices
    Probability on Graphs Random Processes on Graphs and Lattices GEOFFREY GRIMMETT Statistical Laboratory University of Cambridge c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Geoffrey Grimmett Statistical Laboratory Centre for Mathematical Sciences University of Cambridge Wilberforce Road Cambridge CB3 0WB United Kingdom 2000 MSC: (Primary) 60K35, 82B20, (Secondary) 05C80, 82B43, 82C22 With 44 Figures c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Contents Preface ix 1 Random walks on graphs 1 1.1 RandomwalksandreversibleMarkovchains 1 1.2 Electrical networks 3 1.3 Flowsandenergy 8 1.4 Recurrenceandresistance 11 1.5 Polya's theorem 14 1.6 Graphtheory 16 1.7 Exercises 18 2 Uniform spanning tree 21 2.1 De®nition 21 2.2 Wilson's algorithm 23 2.3 Weak limits on lattices 28 2.4 Uniform forest 31 2.5 Schramm±LownerevolutionsÈ 32 2.6 Exercises 37 3 Percolation and self-avoiding walk 39 3.1 Percolationandphasetransition 39 3.2 Self-avoiding walks 42 3.3 Coupledpercolation 45 3.4 Orientedpercolation 45 3.5 Exercises 48 4 Association and in¯uence 50 4.1 Holley inequality 50 4.2 FKGinequality 53 4.3 BK inequality 54 4.4 Hoeffdinginequality 56 c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 vi Contents 4.5 In¯uenceforproductmeasures 58 4.6 Proofsofin¯uencetheorems 63 4.7 Russo'sformulaandsharpthresholds 75 4.8 Exercises 78 5 Further percolation 81 5.1 Subcritical phase 81 5.2 Supercritical phase 86 5.3 Uniquenessofthein®nitecluster 92 5.4 Phase transition 95 5.5 Openpathsinannuli 99 5.6 The critical probability in two dimensions 103 5.7 Cardy's formula 110 5.8 The
    [Show full text]
  • Arxiv:1504.02898V2 [Cond-Mat.Stat-Mech] 7 Jun 2015 Keywords: Percolation, Explosive Percolation, SLE, Ising Model, Earth Topography
    Recent advances in percolation theory and its applications Abbas Ali Saberi aDepartment of Physics, University of Tehran, P.O. Box 14395-547,Tehran, Iran bSchool of Particles and Accelerators, Institute for Research in Fundamental Sciences (IPM) P.O. Box 19395-5531, Tehran, Iran Abstract Percolation is the simplest fundamental model in statistical mechanics that exhibits phase transitions signaled by the emergence of a giant connected component. Despite its very simple rules, percolation theory has successfully been applied to describe a large variety of natural, technological and social systems. Percolation models serve as important universality classes in critical phenomena characterized by a set of critical exponents which correspond to a rich fractal and scaling structure of their geometric features. We will first outline the basic features of the ordinary model. Over the years a variety of percolation models has been introduced some of which with completely different scaling and universal properties from the original model with either continuous or discontinuous transitions depending on the control parameter, di- mensionality and the type of the underlying rules and networks. We will try to take a glimpse at a number of selective variations including Achlioptas process, half-restricted process and spanning cluster-avoiding process as examples of the so-called explosive per- colation. We will also introduce non-self-averaging percolation and discuss correlated percolation and bootstrap percolation with special emphasis on their recent progress. Directed percolation process will be also discussed as a prototype of systems displaying a nonequilibrium phase transition into an absorbing state. In the past decade, after the invention of stochastic L¨ownerevolution (SLE) by Oded Schramm, two-dimensional (2D) percolation has become a central problem in probability theory leading to the two recent Fields medals.
    [Show full text]
  • A Representation for Functionals of Superprocesses Via Particle Picture Raisa E
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector stochastic processes and their applications ELSEVIER Stochastic Processes and their Applications 64 (1996) 173-186 A representation for functionals of superprocesses via particle picture Raisa E. Feldman *,‘, Srikanth K. Iyer ‘,* Department of Statistics and Applied Probability, University oj’ Cull~ornia at Santa Barbara, CA 93/06-31/O, USA, and Department of Industrial Enyineeriny and Management. Technion - Israel Institute of’ Technology, Israel Received October 1995; revised May 1996 Abstract A superprocess is a measure valued process arising as the limiting density of an infinite col- lection of particles undergoing branching and diffusion. It can also be defined as a measure valued Markov process with a specified semigroup. Using the latter definition and explicit mo- ment calculations, Dynkin (1988) built multiple integrals for the superprocess. We show that the multiple integrals of the superprocess defined by Dynkin arise as weak limits of linear additive functionals built on the particle system. Keywords: Superprocesses; Additive functionals; Particle system; Multiple integrals; Intersection local time AMS c.lassijication: primary 60555; 60H05; 60580; secondary 60F05; 6OG57; 60Fl7 1. Introduction We study a class of functionals of a superprocess that can be identified as the multiple Wiener-It6 integrals of this measure valued process. More precisely, our objective is to study these via the underlying branching particle system, thus providing means of thinking about the functionals in terms of more simple, intuitive and visualizable objects. This way of looking at multiple integrals of a stochastic process has its roots in the previous work of Adler and Epstein (=Feldman) (1987), Epstein (1989), Adler ( 1989) Feldman and Rachev (1993), Feldman and Krishnakumar ( 1994).
    [Show full text]
  • A Novel Approach for Markov Random Field with Intractable Normalising Constant on Large Lattices
    A novel approach for Markov Random Field with intractable normalising constant on large lattices W. Zhu ∗ and Y. Fany February 19, 2018 Abstract The pseudo likelihood method of Besag (1974), has remained a popular method for estimating Markov random field on a very large lattice, despite various documented deficiencies. This is partly because it remains the only computationally tractable method for large lattices. We introduce a novel method to estimate Markov random fields defined on a regular lattice. The method takes advantage of conditional independence structures and recur- sively decomposes a large lattice into smaller sublattices. An approximation is made at each decomposition. Doing so completely avoids the need to com- pute the troublesome normalising constant. The computational complexity arXiv:1601.02410v1 [stat.ME] 11 Jan 2016 is O(N), where N is the the number of pixels in lattice, making it computa- tionally attractive for very large lattices. We show through simulation, that the proposed method performs well, even when compared to the methods using exact likelihoods. Keywords: Markov random field, normalizing constant, conditional indepen- dence, decomposition, Potts model. ∗School of Mathematics and Statistics, University of New South Wales, Sydney 2052 Australia. Communicating Author Wanchuang Zhu: Email [email protected]. ySchool of Mathematics and Statistics, University of New South Wales, Sydney 2052 Australia. Email [email protected]. 1 1 Introduction Markov random field (MRF) models have an important role in modelling spa- tially correlated datasets. They have been used extensively in image and texture analyses ( Nott and Ryden´ 1999, Hurn et al. 2003), image segmentation (Pal and Pal 1993, Van Leemput et al.
    [Show full text]
  • Markov Random Fields and Stochastic Image Models
    Markov Random Fields and Stochastic Image Models Charles A. Bouman School of Electrical and Computer Engineering Purdue University Phone: (317) 494-0340 Fax: (317) 494-3358 email [email protected] Available from: http://dynamo.ecn.purdue.edu/»bouman/ Tutorial Presented at: 1995 IEEE International Conference on Image Processing 23-26 October 1995 Washington, D.C. Special thanks to: Ken Sauer Suhail Saquib Department of Electrical School of Electrical and Computer Engineering Engineering University of Notre Dame Purdue University 1 Overview of Topics 1. Introduction (b) Non-Gaussian MRF's 2. The Bayesian Approach i. Quadratic functions ii. Non-Convex functions 3. Discrete Models iii. Continuous MAP estimation (a) Markov Chains iv. Convex functions (b) Markov Random Fields (MRF) (c) Parameter Estimation (c) Simulation i. Estimation of σ (d) Parameter estimation ii. Estimation of T and p parameters 4. Application of MRF's to Segmentation 6. Application to Tomography (a) The Model (a) Tomographic system and data models (b) Bayesian Estimation (b) MAP Optimization (c) MAP Optimization (c) Parameter estimation (d) Parameter Estimation 7. Multiscale Stochastic Models (e) Other Approaches (a) Continuous models 5. Continuous Models (b) Discrete models (a) Gaussian Random Process Models 8. High Level Image Models i. Autoregressive (AR) models ii. Simultaneous AR (SAR) models iii. Gaussian MRF's iv. Generalization to 2-D 2 References in Statistical Image Modeling 1. Overview references [100, 89, 50, 54, 162, 4, 44] 4. Simulation and Stochastic Optimization Methods [118, 80, 129, 100, 68, 141, 61, 76, 62, 63] 2. Type of Random Field Model 5. Computational Methods used with MRF Models (a) Discrete Models i.
    [Show full text]