Technische Universiteit Delft Faculteit Elektrotechniek, Wiskunde en Informatica Faculteit Technische Natuurwetenschappen Delft Institute of Applied Mathematics
Dissipation in the Abelian Sandpile Model
Verslag ten behoeve van het Delft Institute of Applied Mathematics als onderdeel ter verkrijging
van de graad van
BACHELOR OF SCIENCE in Technische Wiskunde en Technische Natuurkunde
door
HENK JONGBLOED
Delft, Nederland Augustus 2016
Copyright c 2016 door Henk Jongbloed. Alle rechten voorbehouden.
BSc verslag TECHNISCHE WISKUNDE en TECHNISCHE NATUURKUNDE
“Dissipation in the Abelian Sandpile Model”
HENK JONGBLOED
Technische Universiteit Delft
Begeleiders
Prof. dr.ir. F.H.J. Redig Dr.ir. J.M. Thijssen
Overige commissieleden
Dr. T. Idema Dr. J.L.A. Dubbeldam Dr. J.A.M. de Groot
Augustus, 2016 Delft
Abstract
The Abelian Sandpile model was originally introduced by Bak, Tang and Wiesenfeld in 1987 as a paradigm for self-organized criticality. In this thesis, we study a variant of this model, both from the point of view of mathematics, as well as from the point of view of physics. The effect of dissipation and creation of mass is investigated. By linking the avalanche dynamics of the infinite-volume sandpile model to random walks, we derive some criteria on the amount of dissipation and creation of mass in order for the model to be critical or non-critical. As an example we prove that a finite amount of conservative sites on a totally dissipative lattice is not critical, and more generally, if the distance to a dissipative site is uniformly bounded from above, then the model is not critical. We apply also applied a renormalisation method to the model in order to deduce its critical exponents and to determine whether a constant bulk dissipation destroys critical behaviour. Numerical simulations and a statistical analysis are performed to estimate critical exponents. Finally, we give a short discussion on self-organized criticality.
Preface
Sitting here in the pittoresque Zillertal in Austria, the beauty of nature intrigues me. Moreover, the symphony of light, sound and movement has triggered existential sensations in many people in history. No wonder that the great Greek philosopher Aristotle has said ‘Wonder is the beginning of all philosophy’. The research documented in this report forms my Bachelor Project at the TU Delft for the pro- grammes Applied Mathematics and Applied Physics. When choosing a project, I wanted to study a subject contained a robust mathematical aspect as well as an interesting physical interpretation. Also, numerical simulation was a preference. These properties are all found in the subject of this research, which is the Abelian Sandpile Model. Proposed for the first time in 1987 by Bak, Tang and Wiesenfeld in a physical context, it served as an illustration of ‘self-organized criticality’. Since then, the model and related concepts have been studied both in the mathematics and physics literature. Although much of the model has been understood thus far, there remain open questions as to how complex structure can arise out of some elementary local interaction rules. This touches on the philosophical concept of emergence. These three disciplines combined -physics, mathematics and philosophy- make the Abelian Sandpile model a beautiful subject for me to study. Some personal notes are in place here. During this project, I had a lot of other activities, like three courses, extracurricular activities and preparations for next year, which will be a year on the board of a student association. All of this made it difficult to really focus on the project, especially since it is on my own. In fact, I have come to better know myself, my interests and my weaknesses in studying. Projects are different from courses in a lot of ways and require a great deal of independence. To double-degree bachelor students, the difference can be even greater due to the relative small number of projects we have had before and necessity of ‘efficient course studying’ in order to graduate in time. In a project, efficiency is a lot harder to achieve, at least in my case. Delay of thesis defence was one of the consequences of my behaviour. I do not know exactly how to say it, but the last year of my bachelor was a lot harder that expected. This project was a real challenge for me; I have learned a lot but also disappointed myself due to my bad skills in planning. These are all lessons I will take with me in the future. I would like to thank my supervisors Frank Redig and Jos Thijssen for the time and effort they offered to help me in my project. Making appointments always went smoothly and I have learned a lot from them. The mathematical background of this project I discussed with Frank, and many times during these meetings we ran into problems that I could not even begin to solve with my relative small experience in the field of probability. Luckily, Frank always found a way to tackle a specific problem. With Jos, I mostly discussed numerical simulation, renormalisation theory and general theory of critical phenomena. These meetings were always very informative and relaxed, for which I thank him greatly. I also thank dr. Johan Dubbeldam, dr. Timon Idema and dr. Joost de Groot for taking place in my thesis committee. Finally, I would like to thank my family, roommates and close friends for giving me advice and helping me. Henk Jongbloed, August 15th, 2016
Contents
1 Introduction 1
2 The classical Abelian Sandpile Model 3
3 Introduction to mixed dissipative/source systems 7
4 Markov Processes, semigroups and generators 11 4.1 Definitions ...... 11 4.2 Deriving the generator of Markov Processes ...... 13 4.3 The Feynman-Kac formula for countable state space Markov Processes ...... 16
5 Toppling numbers, Avalanches and Random Walks 18 5.1 Linking avalanche dynamics to random walks ...... 20 5.2 Towards infinite volume ...... 23 5.3 Estimating critical behaviour ...... 26 5.4 From a CRW to a DRW ...... 26
6 Mathematical results 29 6.1 Conditions on D to obtain non-criticality ...... 29 6.2 Adding sources ...... 32 6.3 Finitely many sources ...... 32
7 A Renormalisation Approach to the BTW model 35 7.1 Introduction: Critical phenomena and renormalisation ...... 35 7.2 General remarks ...... 36 7.3 Renormalisation equations ...... 37 7.4 Fixed points ...... 43 7.5 Critical exponents ...... 43 7.6 Introducing dissipation ...... 45
8 Numerical simulation 46 8.1 Simulating the ASM ...... 46
9 Critical avalanche data analysis 53 9.1 The BTW approach ...... 53 9.2 Likelihood estimation of τ ...... 55 9.3 Truncated power law MLE estimation ...... 58 9.4 Our recommendation ...... 60
10 Self-Organized Criticality 61
11 Conclusions, Discussion, Recommendations and personal notes 64
Appendices 68
A Project Description: Dissipation in the Abelian Sandpile Model 68
B Code 69
1 Introduction
It has been almost 30 years since Per Bak, Chao Tang and Kurt Wiesenfeld (BTW) proposed the sandpile model as a paradigm of self-organized criticality (SOC) [1]. It serves as the simplest and best- studied example of a non-equilibrium system, driven at a slow steady rate by adding particles, with local threshold relaxation rules, which in its critical state shows power-law behaviour obtained without fine-tuning of any control parameters. BTW claimed that the concept of SOC is an explanation of many different physical phenomena: from the formation of the Earth’s crust to the dynamics of solar flares to the distribution of skyscrapers in the world’s biggest cities [2].
Figure 1: Per Bak, Chao Tang and Kurt Wiesenfeld.1
This immediately explains the relevance of studying ‘toy models’ such as the sandpile model. Through detailed understanding of these relatively easy models, we may be able to make predictions about certain phenomena in the real world. The model was originally defined as follows. In one dimension (d = 1), consider a connected subset of Z, which we will call Vn. Without loss of generality we can take Vn = [−n, n] ∩ Z, and thus 2n + 1 is the size of the system. We denote by x ∈ Vn a site, and we define a height function η on Vn. Choosing a site x at random, we consequently keep adding particles to the system, increasing local heights by one unit.
η(x) → η(x) + 1
Given the fixed threshold value 2, when a local height becomes greater or equal to 1, one particle tumbles, which is referred to as a toppling event.
η(x) → η(x) − 2 η(x ± 1) → η(n ± 1) + 1 if η(x) ≥ 2
Boundaries can be either absorbing or periodic. With absorbing boundaries, the one-dimensional BTW model naturally evolves to a minimally stable state, where all but one site have a height of 1. The toppling rules in higher dimensions are essentially the same as in one dimension. In most papers, a d sandpile on a simply connected subset of Z is defined as a height function η with a critical height of 2d determines the toppling condition. Upon toppling, a site distributes its height to its 2d nearest- neighbours. On boundary sites, particles are lost. Thereafter, if another site becomes unstable from a previous toppling, it topples, until the time that no unstable sites exist in the system anymore. At this point, a particle is again added to a random location in Vn. A series of connected toppling events is called an avalanche. However, the dynamics of higher-dimensional BTW models is fundamentally
1Sources: http://www.eoht.info/page/Per+Bak, https://en.wikipedia.org/wiki/Chao_Tang, https://www. physics.gatech.edu/user/kurt-wiesenfeld 1 INTRODUCTION different from the one-dimensional model. Rather than evolving to its minimally stable configuration, it naturally evolves to a critical state in which avalanches of all sizes, up to the size of the system itself, occur. In the years following the proposal of the model, many researchers began working on it. Numerical simulations have been performed by Manna [4], who numerically identified some parameters governing the critical state of the 2D model. Not long after the BTW paper, Dhar and others [3] began formalising the model in terms by introducing the concept of addition operators, their group structure and spanning trees. He therefore called the original BTW model the ‘Abelian Sandpile Model’ (ASM). Abelianness in the dynamics stems from the fact that first adding a particle at a site x, let the relaxation dynamics take place, and then add a particle at y and letting relax results in the same final configuration as first adding at y and thereafter at x. Later, Frank Redig considered the infinite-volume limit of the ASM as the limit of the finite model. Critical exponents in various dimensions have been computed by mean-field approximations as well as by partial rigorous computations [6] and renormalisation methods [20], but in most cases these derivations are highly non-trivial. Methods of statistical physics provide another way to analyse the BTW model. The theory of critical phenomena, such as phase transitions, has been increasing greatly since the 1930’s. Since then, numerous critical phenomena have been analysed. Like water at a phase transition, the behaviour of large ensembles of particles dramatically change when approaching a so-called ‘critical point’: charac- teristic time and length scales vanish and highly non-linear dynamics are observed. The scale-invariance that characterises the critical state of such systems gives rise to renormalisation theory. In this method, the dynamics at a generic scale are linked to the dynamics at another scale via the renormalisation transformation. Scale-invariance then allows calculation of critical exponents and formulation of scaling laws. In this thesis, we will look at the effect of dissipation and formation of mass in the Abelian Sandpile Model. We will study variations of the BTW model with sinks (dissipative vertices where mass disap- pears) and sources (vertices where mass is added). Analytically we will focus on the question above which level of dissipation criticality is lost. Physically we will apply a renormalisation approach to the infinite-volume BTW model in two dimensions, and numerically simulate the finite-volume BTW model with sinks and sources. This report is organised as follows. Firstly, we will introduce the Abelian Sandpile Model in a mathematical way. We then continue by considering micro-scale dynamics in the one-dimensional model and the effect of sources and sinks. Thereafter, we develop the general mathematical theory concerning Markov processes, their semigroups and generators. Via the famous Feynman-Kac equation d combined with Dhar’s formula, we will relate avalanche dynamics to a simple random walk on Z , derive a characterisation of criticality in the d-dimensional model and consequently derive conditions for criticality or non-criticality. The more physical part of this paper begins by introducing critical phenomena and related concepts. Thereafter, a renormalisation approach will be applied to the BTW model in which we try to incorporate dissipation. The concept of self-organised criticality and emergence is discussed. Thereafter, we present results from numerical simulations. It will come as no surprise if the reader has not fully understood the matter in this introduction. It mostly serves as a quick overview of the sandpile model and its history in various disciplines of science. We encourage the reader to enjoy what follows!
2 2 THE CLASSICAL ABELIAN SANDPILE MODEL
2 The classical Abelian Sandpile Model
This section provides a brief discussion of the general mathematical aspects of the classical Abelian Sandpile model, sometimes abbreviated as ASM. By ‘classical’, we mean that we introduce the model the same way as Bak, Tang and Wiesenfeld [1], which was later described mathematically by Dhar [3]. This section is based on chapter 3 of the paper by Redig [5]. We consider as a ‘sandpile base’ the d simply connected finite set Vn ⊆ Z . Many other lattices and sets exist where sandpile automata can be defined upon, but we will restrict us to this class of lattices. We define the sandpile basis by
d d Vn = [−n, n] ∩ Z
with n ∈ N fixed. As the toppling or system matrix, we take minus the lattice Laplacian: ∆x,x = 2d, ∆x,y = −1, for x, y ∈ Vn, |x − y| = 1, and ∆xy = 0 for x, y ∈ Vn, x 6= y, |x − y|= 6 1. A height configuration η is a map η : Vn → N, and the set of all height configurations is denoted by H. A height configuration η ∈ H is called stable if for every x ∈ Vn, η(x) < ∆xx. The set of all stable configurations is denoted by Ω = {η ∈ H : η(x) < ∆xx for all x ∈ Vn}. A site x ∈ Vn where η(x) ≥ ∆x,x is called an unstable site. The toppling of a site x ∈ Vn is now conveniently defined by
Tx(η)(y) = η(y) − ∆x,y (1)
This means that the site x will lose 2d grains and will distribute them to its nearest neighbours in Vn. Note that mass is conserved, except when boundary sites topple, i.e. sites with less than 2d neighbours in Vn. By convention, boundary sites lose 2d grains upon toppling and distribute these to nearest neighbours, if present, and other grains enter a global sink. The toppling of a site x is called legal if x is unstable, otherwise it is called illegal. Furthermore, if x, y ∈ Vn are both unstable sites of η, then
TxTy(η) = η − ∆x,· − ∆y,· = TyTx(η) (2)
This is the elementary abelian property of the Abelian Sandpile Model, which follows directly from V the commutativity of addition in Z n . From the elementary abelian property it follows that any finite sequence of the same legal topplings yields the same end result, independent of the order of toppling. This is very important in both mathematical and computational perspective. Also, the elementary abelian property motivates defining the following operator:
S (η) = Tx1 ...Txn (η) (3)
S : H → Ω is called the stabilization operator. In this definition, two additional requirements must be made: S (η) must be stable and for all i ∈ {1, .., n}, the toppling at site xi must be legal. For η ∈ H
and a sequence Tx1 ...Txm of legal topplings we define the toppling numbers of that particular sequence as m X nx = I(xi = x) (4) i=1 Now the configuration resulting from that sequence of topplings can be written as
Tx1 ...Txm (η) = η − ∆n (5) where n is the column indexed by x ∈ Vn with elements nx. This formula is very important in the mathematics of the ASM. It is proven in [5] that in the classical case of the finite-volume ASM, S is well-defined. This means that when η ∈ H is given, Tx1 ...Txn and Ty1 ...Tym are two sequences of legal topplings leading to a stable configuration, the resulting configuration will be the same. Furthermore, nx = mx for all x ∈ Vn. In the proof of the well-definedness of S , it it very important that the toppling numbers resulting
3 2 THE CLASSICAL ABELIAN SANDPILE MODEL
from the stabilizing process via legal topplings are finite. That is, for every η0 ∈ H there exists a finite N sequence of sites (xi)i=1 ∈ Vn such that
S (η0) = Tx1 ...TxN (η0) = η0 − ∆n is stable. This is a very important feature of the classical Abelian Sandpile Model. The well-definedness of the stabilization operator S immediately implies that the addition operator
axη = S (η + δx) (6) is well defined, and abelianness holds:
axayη = ayaxη = S (η + δx + δy), (7) which is why the ASM is called Abelian. The Abelian Sandpile as described by BTW consists of a sequence of stable configurations. At each time n ∈ N, a ‘particle’ is added at a certain x ∈ Vn. Thereafter, the configuration is stabilized. This process continues in time, resulting in a sequence of stable configurations evolving in time. More mathematically, let p = p(x) be a probability distribution on Vn, thus p(x) ≥ 0 for all x ∈ Vn and P p(x) = 1. Starting from an initial height configuration η ∈ H, the configuration at time n is x∈Vn 0 given by the random variable n Y ηn = aXi η n ∈ N (8) i=1 iid where X1, ..., Xn ∼ p. The process (8) is a Markov chain. Ω is the state space of the Markov chain, and the Markov transition operator defined on functions f :Ω → R is then given by X P f(η) = E(f(η1|η0 = η)) = p(x)f(axη) (9) x∈Vn Now, the configurations η ∈ Ω can be divided in two classes: recurrent and transient, as is always the case with Markov processes. However, we can prove that the set of transient configurations is non-empty. To see this, consider a one-dimensional finite system with Vn = {−2, −1, 0, 1, 2} and initial height function 1 1 0 0 1. We claim that this configuration is transient. Indeed, the two zeros will never come back: Adding a particle to any site in this configuration causes one of the two middle sites to gain a particle. The only way this site can have zero particles again is by toppling, but then its neighbouring site will have height one. This is also true in d > 1. Fortunately, we have a characterisation of configurations which are transient, and equivalently a characterisation of recurrent configurations.
DEFINITION 2.1. Let η ∈ H. For W ⊆ Vn, W 6= ∅, we call the pair (W, ηW ) a forbidden subconfigu- ration (FSC) if for all x ∈ W : X η(x) < (−∆xy) y∈W \{x}
If for η ∈ Ω there exists a FSC (W, ηW ), then we say that η contains a FSC. A configuration η ∈ Ω is called allowed if it does not contain forbidden subconfigurations. The set of stable allowed configurations is denoted by R0. Note that Let us denote the (unique) set of recurrent configurations by R. This set is unique because the Markov chain restricted to this set is irreducible: any element of R can be reached from every other element in R. If we denote by the set A the set of all finite products of addition operators on Ω, (A , ·) is an abelian semigroup. Furthermore, restricted to R, G ≡ (A , ·) is an abelian group. Some very useful theorems can be proven about the set of recurrent configurations in the ASM. A particularly interesting example of this is the following.
4 2 THE CLASSICAL ABELIAN SANDPILE MODEL
THEOREM 2.1. A stable configuration η ∈ Ω is recurrent if and only if it is allowed. So R = R0.
PROOF. The proof can be found in [5]. Theorem 2.1 is a rigorous statement of the phenomenon we have encountered before by deducing that 1 0 0 1 is a transient configuration. Indeed, (W, ηW ) = ({1, 2}, {0, 0}) is a FSC. Therefore 1 0 0 1 is a transient configuration. A bijection can be made between rooted spanning trees and recurrent configurations, as has been illustrated by Dhar [3] using Kirchhoff’s matrix tree theorem and Dhar’s burning algorithm. Thereby, we arrive at the identity |R| = det (∆) (10) Because of the group structure of addition operators under composition, Dhar shows that G is iso- morphic to the group of equivalence classes of configurations gotten by reducing modulo the toppling operation, which can be written as V V G ' Z n /∆Z n (11) where ∆ again denotes the lattice Laplacian. It is known that an irreducible Markov chain has a stationary distribution if and only if all of its states are recurrent. Restricting ηn to R, the Markov chain is irreducible. Therefore a stationary distribution on R exists. Since there is strictly positive probability that a recurrent configuration will be reached in the process ηn, n ∈ N, the Markov chain will eventually reach that class and remain in R forever. Now, denote the stationary distribution on R by µ(η).
THEOREM 2.2. The process ηn, n ∈ N with transition operator X P f(η) = p(x)f(axη) (12) x has a stationary distribution 1 X µ(η) = µ = δη (13) |R| η∈R i.e. it is simply uniform on R.
PROOF. X 1 X X P f(η)µ(η) = f(axη) |R| η∈R η∈R x∈Vn 1 X X = p(x) f(axη) |R| x∈Vn η∈R 1 X X 0 = p(x) f(η ), since ax : R → R is a bijection |R| 0 x∈Vn η ∈R ! X 1 X = p(x) f(η0) |R| 0 x∈Vn η ∈R 1 X = f(η0) |R| η0∈R X = f(η0)µ(η) η0∈R
Previous considerations were all made for finite system size Vn. An interesting question is to look at the
5 2 THE CLASSICAL ABELIAN SANDPILE MODEL
d infinite volume limit. What happens if we let Vn ↑ Z by letting n → ∞? Bak, Tang and Wiesenfeld [1] define criticality by the emergence of power-law behaviour in avalanche sizes. Redig [5] takes another perspective which has to do with expected avalanche size in the large volume limit. We turn to the large volume limit later in this report.
6 3 INTRODUCTION TO MIXED DISSIPATIVE/SOURCE SYSTEMS
3 Introduction to mixed dissipative/source systems
In the previous section, we have seen that the ‘classical’ finite-volume Abelian Sandpile model is called ‘abelian’ for a reason. The elementary abelian property of toppling TxTy = TyTx, together with the well-definedness of the stabilization operator S guarantees the abelian property of addition operators: axay = ayax. This property is very important for the mathematics and simulation of the Abelian Sandpile model. Regardless of the order of legal topplings, the resulting stable configuration S (η) ∈ Ω is always the same, starting from some η ∈ H. Let us now define a general finite-volume toppling d matrix for systems in which sources and sinks are present. Again, Vn = [−n, n] ∩ Z is our finite lattice and we define by Dn ⊂ Vn the set of dissipative sites and Fn ⊂ Vn the set of source sites. When d viewing D as the set of dissipative sites in infinite volume and F as the source set in Z , we have Cn = C ∩ Vn, Dn = D ∩ Vn and Fn = F ∩ Vn. By denoting Cn the set of normal, conservative sites, we have Vn = Cn ∪ Dn ∪ Fn as a union of pairwise disjoint sets.
Dn,Fn d DEFINITION 3.1. The finite-volume toppling matrix ∆x,y for x, y ∈ Vn = [−n, n] ∩ Z is defined as −1 for x, y ∈ V , |x − y| = 1 n 2d + 1 for x = y ∈ Dn ∆Dn,Fn = (14) x,y 2d for x = y ∈ C n 2d − 1 for x = y ∈ Fn
Upon toppling at a site x ∈ Dn, which happens at a height 2d + 1, the site distributes 2d grains to its nearest neighbours, and looses one grain to an invisible sink. Conversely, upon toppling at a site x ∈ Fn, which happens at a height 2d − 1, the site distributes 2d grains to its nearest neighbours, and creates one grain ex nihilo. In this section, we focus on mixed sink/source systems, with normal, dissipative and source sites. We begin by observing some behaviour in one dimension.
Stabilizability We aim to provide some conjectures about finite-system configurations that are al- ways stabilizable in a finite time. That is, via a finite sequence of topplings, one arrives at a unique stable final configuration. Can we characterize a class of subsystems that are always stabilizable? Firstly, the definition of stabilizability, metastabilizability and unstabilizability is given below. Note that a system is fully characterized by the sets Cn,Dn,Fn.
DEFINITION 3.2. Characterization of systems
Given an initial configuration η0 ∈ H and a legal toppling sequence (xn)n∈N, a system is called
N 1. stabilizable, if for all unstable η0 ∈ H there exists a finite legal toppling sequence (xn)n=1(η0) such
that ξ = ηN = TxN TxN−1 ...Tx1 (η0) satisfies ξ(x) < ∆xx for all x ∈ Vn.
∞ 2. metastabilizable, if there exists η0 ∈ H such that for every legal toppling sequence (xn)n=1(η0) and
for all N ∈ N, ηN = TxN TxN−1 ...Tx1 (η0) is unstable, and there exists M ∈ N such that the total P mass is bounded: x ηN (x) < M.
∞ 3. unstabilizable, if there exists η0 ∈ H such that for every legal toppling sequence (xn)n=1(η0) and P for all N ∈ N, ηN = TxN TxN−1 ...Tx1 (η0) is unstable, and the total mass x η(x) diverges. For example, if a system only contains only normal and dissipative sites, it is always stabilizable. Mass cannot build up in the interior of Vn if no sources are present. At every toppling of a dissipative site, up to one grain is lost, so the total mass of the system P η(x) is a decreasing sequence. x∈Vn Eventually, it will be stable. Some examples are given below. The bold-faced numbers indicate the sequence of decreasing heights on the source site, indicating stabilizability. This is all in d = 1, and
7 3 INTRODUCTION TO MIXED DISSIPATIVE/SOURCE SYSTEMS because small system sizes are considered, we can write it out by hand. Now, we denote a dissipative site by δ, a source site by σ and a normal site by ν.
Table 1: A stabilizable δσδ system with initial configuration 0n0.
δ σ δ Sum 0 n 0 n 1 n-1 1 n+1 2 n-2 2 n+2 3 n-3 3 n+3 0 n-1 0 n-1
It is clear that this small configuration is stabilizable, since it is losing mass at a steady rate and will eventually reach a stable configuration.
Table 2: A stabilizable νδσδν system with initial configuration 00n00.
ν δ σ δ ν Sum 0 0 n 0 0 n 0 1 n-1 1 0 n+1 0 2 n-2 2 0 n+2 0 3 n-3 3 0 n+3 1 1 n-2 1 1 n+2 1 2 n-3 2 1 n+3 1 3 n-4 3 1 n+4 2 1 n-3 1 2 n+3 0 3 n-4 3 0 n+2 1 1 n-3 1 1 n+1
The stabilizability of this configuration can clearly be seen.
Table 3: A metastabilizable νσν system with initial configuration 0n0. This time, sites are toppled conveniently to demonstrate the periodicity of configurations, which indicate the metastabilizability of this system.
ν σ ν Sum 0 n 0 n 1 n-1 1 n+1 2 n-2 2 n+2 0 n 0 n
8 3 INTRODUCTION TO MIXED DISSIPATIVE/SOURCE SYSTEMS
Table 4: An unstabilizable δσσδ system with initial configuration 0010. Mass builds up in the system because two sources next to each other in d = 1 are never stabilizable.
δ σ σ δ Sum 0 0 1 0 1 0 1 0 1 2 1 0 1 1 3 1 1 0 2 4 2 0 1 2 5 2 1 0 3 6 3 0 2 0 5
Among other observations, we see that νσν is periodic in its configurations. After a finite number of legal topplings, the same configuration is reached. From this, we can conjecture that a subsystem νσν embedded in some larger system is never stabilizable when given a configuration of 0n0, with n ≤ 2, or n0m, with n + m ≤ 3. That is, a source site has to be next to at least one dissipative site in order for the configuration to be stabilizable in general. From this, one can also deduce that the system δσν is stabilizable. But, as we shall see, the embedding of δσν in a larger system is rather complicated. We can therefore formulate some conjectures about the behaviour of one-dimensional mixed sink/source systems as follows.
• A system containing subsystem σσ is unstabilizable.
• A system consisting of configurations δσδ and further only ν (and δ) is always stabilizable.
• A system with source boundaries, and furthermore only ν or δσδ configurations, conserves mass and is therefore metastabilizable.
• Generally, a system containing the subsystem νσν is unstabilizable.
• A system consisting of only ν and δ sites is always stabilizable.
• When embedded in a finite lattice of dissipative sites, the subsystem σν is sometimes stabilizable. The stabilizing behaviour highly depends on the distance of the subsystem σν to the boundary of the lattice. When multiple σν subsystems are present, the behaviour also depends on the distance between them. This behaviour is rather chaotic.
In fact, we can summarize these statements in the following conjecture, which is rather surprising.
CONJECTURE 3.1. Characterizations of stabilizability For a general system consisting of normal, dissipative and source sites, with associated system matrix ∆Dn,Fn , the following statements are true. A system is
1. stabilizable ⇔ ∀λ ∈ λ(∆Dn,Fn ): λ > 0
2. metastabilizable ⇔ ∃λ ∈ λ(∆Dn,Fn ): λ = 0
3. unstabilizable ⇔ ∃λ ∈ λ(∆Dn,Fn ): λ < 0
Motivation. This conjecture was obtained while proving Dhar’s formula, documented in following sections. When simulating different systems, we started to see a one-to-one correspondence between the inverse toppling matrix and the behaviour of the stabilizing process. This was in the one-dimensional case. For example, consider the one-dimensional system σννν...νννσ. We know that this system is metastabilizable: when all sites topple at once, no mass is lost. We then see that det (∆Dn,Fn ) = 0.
9 3 INTRODUCTION TO MIXED DISSIPATIVE/SOURCE SYSTEMS
Also the system ννν...σσ...ννν is unstabilizable. We then see that det (∆Dn,Fn ) < 0. However, to link the determinant of ∆Dn,Fn directly to the stabilizability of a system with toppling matrix ∆Dn,Fn may be too pretentious. We therefore state this weaker conjecture, knowing that the determinant of a matrix is the product of its eigenvalues.
Meta- and unstabilizability It turns out that we cannot conclude much about meta- or unstabi- lizable systems. The notion of extended abelianness has been disproved. By extended abelianness we mean that a metastabilizable system shows a finite set of configurations in the stabilizing process where only a finite number of ‘recurrent’ unstable configurations are encountered. Regardless of the order of topplings, the same configurations would emerge and one could define a measure on these recurrent unstable configurations. However, this is not the case, as basic numerical simulations show. A first example shows that the system order is very important. Consider a system δσσδδ with initial height function 0010. This configuration is not stabilizable, even though the total mass of 1 is less than the mass of the minimally stable configuration 20022 which has total mass 6. A system consisting of the same vertices but in different order, say δσδσδ with initial configuration 01010 is stable in one step: when using the ‘left first’ toppling rule, we have 01010 → 10110 → 10201. Via the ‘select most unstable vertex’ rule, we have the same result. This makes it very difficult, in general, to make predictions about the behaviour of large-size mixed sink-source systems, since the behaviour depends largely on the micro-structure of the system. As an example of the failed generalized abelian property conjecture, consider the metastabilizable dis- tribution νσσσν with initial configuration 22222. It depends on the toppling algorithm which set of unstable configurations is seen. At this point, we therefore have to restrict ourselves to stabilizable systems.
10 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
4 Markov Processes, semigroups and generators
In this section, we will develop the theory to support our mathematical investigation of the Abelian Sandpile Model considering dissipation and anti-dissipation. First, some definitions are needed. We will begin by defining stochastic processes and in particular Markov processes. Associated to Markov processes are semigroups and generators. In various textbooks, such as [8], these definitions are very precise. In order to analyze the theoretical framework of stochastic processes, this is necessary. We choose to take a slightly less robust way of defining and reasoning. The most important thing to remember is that under certain conditions, there is a natural one-to-one correspondence between a Markov process, its semigroup and it generator. This introduction is largely based on [9].
4.1 Definitions We begin by defining Markov processes, their associated semigroups and corresponding generators. A Markov process is a specific example of a stochastic process.
DEFINITION 4.1. Stochastic Process Given a probability space (X , F ,P ) and a measurable space (Ω, Σ), a stochastic process is a collection of Ω-valued random variables indexed by a totally ordered set T :
{Xt : t ∈ T }
The space Ω is referred to as the state space of the process.
Stochastic processes come in a wide variety of shapes and forms: among other things, they are extensively used in physics, economics and biology. We will focus on a class of stochastic processes with a special property, called the Markov property, named after the famous Russian mathematician Andrey Markov (1856-1922). Stochastic processes with the Markov property are called Markov processes. From now on, we take T = [0, ∞). Furthermore, we assume that Ω is countable.
DEFINITION 4.2. Markov Property A stochastic process {Xt : t ≥ 0} on a state space Ω is called to have the Markov property if for all t > s ∈ [0, ∞) and x ∈ Ω the equality
E (f(Xt+s)|Fs) = E [f(Xt+s)|Xs]
where Fs ≡ σ{Xr : 0 ≤ r ≤ s}, holds. Alternatively, we can formulate the Markov property as follows: A stochastic process {Xt : t ≥ 0} is called a Markov process if for all t > 0, and for all 0 < t1 <, ..., < tn < t, n ∈ N
E[f(Xt)|Xt1 , ..., Xtn ] = E[f(Xt)|Xtn ] for all measurable and bounded f :Ω → R. One can see this property as a form of loss of memory. Conditioning on the entire past of the process until a time s is equivalent to conditioning only on the value of the process at time s. Also, given the state of the process at time s, the past of the process and future of the process are independent. An family of operators naturally associated to a Markov process {Xt, t ≥ 0} is the so-called semigroup {St, t ≥ 0}.
DEFINITION 4.3. Semigroup For measurable and bounded f ∈ Cb(Ω), the semigroup associated to a Markov process {Xt, t ≥ 0} with state space Ω is defined as
Stf(x) = E[f(Xt)|X0 = x]
11 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
We will also use the notation Stf(x) = E[f(Xt)|X0 = x] = Ex[f(Xt)]. St has a number of useful properties, justifying the name ‘semigroup’. These are, among others for f ∈ Cb(Ω):
1. S0f = f
2. St1 = 1
3. f ≥ 0 implies Stf ≥ 0
4. for s, t > 0, we have the semigroup property: StSs = St+s
5. St is a contraction semigroup: ||Stf||∞ ≤ ||f||∞. Properties 1-3 are trivial to see. Property 4 follows from the observations
St+sf(x) = E[f(Xt+s)|X0 = x] = Ex[f(Xt+s)] = Ex[E[f(Xt+s)|Xs]]
= Ex[EXs [f(Xt)]], by the Markov property ofXt = Ex[(Stf)(Xs)] = (Ss(Stf))(x) = (SsStf)(x)
This expression is symmetric in t and s, from which it follows that St+s = StSs = SsSt. We are now ready to define the generator of a Markov process {Xt, t ≥ 0}.
DEFINITION 4.4. Generator Let St be the semigroup of a Markov process {Xt, t ≥ 0}. For suitably measurable f :Ω → R, we define the generator L by S f − f Lf = lim t (15) t↓0 t for functions f for which this limit exists.
One can interpret the generator of the semigroup St as the ‘derivative’ of St at t = 0, or equivalently as the expected behaviour of the process after a very short time. Indeed, for f ∈ Cb(Ω) we have
Ex(f(Xt)) = f(x) + tLf(x) + o(t), as t ↓ 0.
The Markov process {Xt, t ≥ 0}, the semigroup operator St and the generator L are closely and exclusively related. In the case of finite state space Ω, we can write the semigroup as a matrix where
(St)x,y = P (Xt = y|X0 = x) (16) Also, when L is a bounded operator, the semigroup property allows the semigroup with generator L to be written as ∞ X tnLn S = etL = (17) t n! n=0
The boundedness of L ensures that St is well-defined in this case. Indeed, we have
∞ n ∞ n n ∞ n n X tnL X t ||L || X t ||L|| k k ≤ ≤ = et||L|| < ∞. n! n! n! n=0 n=0 n=0 In the case of a general generator L, one cannot simply write the semigroup as in equation (17), because this expression is not guaranteed to have meaning in case of unbounded L. However, under certain conditions, a natural connection can be made between a Markov generator L and its semigroup {St : t ≥ 0}, a result that is known as the Hille-Yosida theorem. For these conditions, we refer the reader to for example [8].
12 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
THEOREM 4.1. Hille-Yosida The one-to-one connection between a Markov generator and a contrac- tion semigroup is given by S f − f Lf = lim t (18) t→0 t and vice versa t −n St = lim I − L (19) n→∞ n
These definitions, relations and theorems can be viewed in a much wider sense than we have done here. For a profound discussion, see for example [8]. We are almost finished introducing general definitions and properties associated to Markov processes. It is useful to define the notion of an invariant measure.
DEFINITION 4.5. Invariant measure Given a family (St)t≥0 of Markov semigroup operators on a measurable space (Ω, Σ) as before, a (pos- itive) σ-finite measure µ on (Ω, Σ) is said to be invariant for (St)t≥0 if for every bounded measurable function f :Ω → R and every t ≥ 0: Z Z Stfdµ = fdµ (20) Ω Ω
In general, the invariant measure is only defined up to a multiplicative constant. When it is finite, it is a natural choice to normalize it by a positive constant to construct a probability measure. Then, it has a clear probabilistic meaning for the associated Markov process {Xt, t ≥ 0}. If the process starts at time t = 0 from X0 with initial distribution µ, then it keeps this distribution for all t, since by the law of total expectation and the Markov property it is true that for any bounded and measurable function f :Ω → R:
E [f(Xt)] = E [E [f(Xt)|X0]] = E [Stf(X0)] Z = Stfdµ Ω Z = fdµ Ω = E [f(X0)]
It can be proven rather easily that the following holds: R µ is an invariant measure for a family of Markov semigroups (St)t≥0 ⇔ Ω Lfdµ = 0 for all f :Ω → R bounded and measurable.
4.2 Deriving the generator of Markov Processes
Poisson process Let us take a short break and compute the Markov generator of a Poisson process. This is a process in space Ω = N, where at random times the process ‘jumps one step up’. The process starts from some x ∈ N, so that X0 = x. After a time T1 ∼ exp (λ) has elapsed, the process jumps one step up by an fixed amount. This is repeated in time, where the elapsed times T1, ..., Tn are exponentially distributed with rate λ. The memorylessness of the exponential distribution is important. In fact, this memoryless property guarantees the Markov property of the process, and thereby the existence of its semigroup and generator.
13 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
Poisson Processes with various rates 50
45 λ =2 λ =4 λ =8 40
35
30
t 25 X
20
15
10
5
0 0 5 10 15 20 25 30 35 t
Figure 2: A realization of a Poisson process with rates 1, 4 and 8.
For small enough t, we may only keep order t terms. Let Nt denote the random variable of the num- ber of jumps of the process starting at time t = 0. Because waiting times are distributed exponentially, we can calculate
Stf(x) = Ex[f(x + Nt)] (21) 2 = f(x)P (Nt = 0) + f(x + 1)P (Nt = 1) + O(t ) for small t (22) = e−tλf(x) + (1 − e−tλ)f(x + 1) + O(t2) (23) It follows from l’Hopital’s rule that e−tλf(x) + (1 − e−tλ)f(x + 1) − f(x) Lf(x) = lim + O(t) t→0 t = λ (f(x + 1) − f(x)) This is the generator for a Poisson process with rate λ.
LEMMA 4.1. Distribution of Nt When defining a continuous Markov process {Xt, t ≥ 0} with fixed rate λ on a state space Ω, we postulate that the transition times T1,T2, ... are i.i.d. exponentially −1 distributed with parameter λ, such that E[Ti] = λ for all i ∈ N. From this, it is easy to see that the random variable Nt representing the number of transitions in [0, t) is distributed as Nt ∼ P ois(λt).
PROOF. Let T1,T2, ... ∼ exp (λ) be independent identically distributed random variables. It is known Pn that Mn ≡ i=1 Ti ∼ Gamma(n, λ). Now, the probability that exactly n transitions of the Markov process occur in the interval [0, t) is equal to the probability that Mn ≤ t and Mn+1 > t. Thus
P (Nt = n) = FMn (t) − FMn+1 (t) (24) where FMn = P (Mn ≤ t). Since Mn is Gamma distributed with parameters n and λ, we have n Z t n+1 Z t λ n−1 −λs λ n −λs P (Nt = n) = s e ds − s e ds (25) Γ(n) 0 Γ(n + 1) 0 By partial integration, we have the identity Z t 1 h i Z t sn−1e−λsds = tne−λt + λ sne−λsds 0 n 0 Substituting this expression back in (25), many terms cancel and eventually (λt)n P (N = n) = e−λt (26) t n! which is the Poisson distribution function with parameter λt.
14 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
The generator of a discrete Markov jump process Lemma 4.2 states a very important property of continuous-time Markov processes, something which we have used a number of times earlier in this section.
LEMMA 4.2. In order to have the Markov property, a continuous-time stochastic process {Xt, t ≥ 0} has to make transitions from state to state at exponentially distributed times.
PROOF. Let {Xt, t ≥ 0} be a Markov process on a state space Ω. Furthermore, let X0 = x. Let T denote the random variable representing the first time of transition from x ∈ Ω to another state. Then
P (T > t + s|T > s) = P (Xr = x ∀r ∈ [0, t + s]|Xr = x ∀r ∈ [0, s])
= P (Xr = x ∀r ∈ [0, t + s]|Xs = x), , by the Markov property ofXt
= P (Xr = x ∀r ∈ [0, t]), , by the Markov property ofXt = P (T > t)
The only memoryless continuous distribution functions are the family of exponential distributions, so memorylessness completely characterises the exponential distribution of T . The rate of T can be state- dependent. In a more general setting therefore, for x, y ∈ Ω, we define the rate function c :Ω × Ω → R. The numbers c(x, y) are taken as the rates to get from x to y, and can be viewed as an ‘probability per unit time’. Accordingly, the transition probability to reach a state x to state y is given by
c(x, y) X Π(x, y) ≡ , where Λx ≡ c(x, y) (27) Λx y6=x
A natural parameter for this is Λx, which obviously depends on x. This way, if Λx is small, because exit rates from x to any y ∈ Ω are small, the expected resting time of the process in x which is given −1 by Λx is large. Let us now proceed in deriving a general expression for the generator of a Markov process {Xt, t ≥ 0}. To obtain it, we expand Stf(x) to first order in t, just as in the derivation of the generator of the Poisson process. Again, we let Nt denote the stochastic variable of a the number of transitions of the process. So, for small t:
2 Stf(x) = Ex[f(Xt)] = Ex[f(Xt) · I(Nt ∈ {0, 1})] + O(t ) X 2 = P (Nt = 0)f(x) + P (Nt = 1) Π(x, y)f(y) + O(t ) y∈Ω X = e−Λxtf(x) + (1 − e−Λxt) Π(x, y)f(y) + O(t2) y∈Ω
Let me explain what is happening here. If Nt = 0, the process has stayed in x for a time t ≤ T1, where T1 ∼ exp (Λx). This happens with probability P (Nt = 0) = P (T1 ≥ t) = 1 − P (T1 ≤ t) = 1 − (1 − exp (−Λxt)) = exp (−Λxt). If Nt = 1, this means that the process has taken one step. This happens with probability P (Nt = 1) = Λxt exp (−Λxt). Furthermore, in one step the process can enter state y ∈ Ω with probability Π(x, y). We continue by expanding exp (−Λxt): for small t we have 2 exp (−Λxt) = 1 − Λxt + O(t ). We obtain
X 2 Stf(x) = f(x) − Λxf(x)t + Λxt Π(x, y)f(y) + O(t ) y∈Ω X X 2 X = f(x) − tΛx Π(x, y)f(x) + Λxt Π(x, y)f(y) + O(t ), because Π(x, y) = 1 y∈Ω y∈Ω y X = f(x) + t c(x, y)(f(y) − f(x)) + O(t2). y∈Ω
15 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
The last equation follows because c(x, y) = ΛxΠ(x, y). Consequently, for the Markov generator of this general Markov process L we can immediately conclude X Lf(x) = c(x, y)(f(y) − f(x)). (28) y∈Ω This is the general expression for a generator for discrete Markov jump processes. In matrix form (finite state space), L contains the transition rates on its off-diagonals and its rows sum up to zero.
4.3 The Feynman-Kac formula for countable state space Markov Processes The Feynman-Kac formula is very important when solving certain kinds of partial differential equa- tions. Named after the prominent American theoretical physicist Richard Feynman (1918-1988) and the influential Polish American mathematician Mark Kac (1914-1984) who published the formula in 1949, it has been widely used in the analysis of stochastic partial differential equations, most notably partial differential equations, such as diffusion equations.
Figure 3: Richard Feynman and Mark Kac.2
It is often the case that one has to look at differential operators of the form
Lf + V f (29) where L is the generator of a Markov semigroup and V f is multiplication by some non-constant potential V . This happens in particular when solving Schr¨odinger-type equations. In the case of the sandpile model, the reader may guess that V has something to do with sinks and sources on top of the original model. Applied to our case, we have the following version of the Feynman-Kac formula. Its importance shall soon be clear.
THEOREM 4.2. Feynman-Kac formula Let {Xt, t ≥ 0} be a Markov process with generator L on a countable state space Ω, and let V :Ω → R be a bounded function. Define the diagonal ‘matrix’ Vxy = V (x)δxy. Then L+V is an infinitesimal generator and the semigroup Σt generated by A satisfies
h R t i t(L+V ) V (Xs)ds Σtf(x) = e f(x) = Ex f(Xt)e 0 (30) This is a version of the Feynman-Kac formula.
2Sources: http://www.inspiremeyouth.com/richardfeynman/, https://en.wikipedia.org/wiki/Mark_Kac
16 4 MARKOV PROCESSES, SEMIGROUPS AND GENERATORS
PROOF. Expanding the right hand side of the Feynman-Kac formula to first order in t gives for the R t integral in the exponent 0 V (Xs)ds ≈ tV (x), and thus
h R t i V (Xs)ds Σtf(x) = Ex f(Xt)e 0
h tV (x)+O(t2)i = Ex f(Xt)e
h tV (x)i 2 = Ex f(Xt)e + O(t ) 2 = Ex [f(Xt)] + tV (x)Ex [f(Xt)] + O(t )
Because L is the generator of {Xt, t ≥ 0}, for small t it holds that tL + f = Ex[f(Xt)]. Therefore, we continue by saying
h R t i V (Xs)ds 2 Ex f(Xt)e 0 = tV (x)f(x) + f(x) + tLf(x) + O(t ). (31)
Subtracting f(x) from this equation, dividing by t and letting t ↓ 0 we obtain our result
h R t i Σtf(x) − f(x) 1 V (Xs)ds lim = lim Ex f(Xt)e 0 − f(x) t↓0 t t↓0 t 1 2 = lim tV (x)f(x) + f(x) + tLf(x) − f(x)O(t ) t↓0 t = (L + V )f(x)
W conclude that L + V is the generator of the semigroup Σt. To prove that Σt has the semigroup property, we show that for t, s > 0, Σt+s = ΣtΣs.
h R t+s i V (Xr)dr Σt+sf(x) = Ex f(Xt+s)e 0 h h R t+s ii V (Xr)dr = Ex E f(Xt+s)e 0 |Fs , where Fs ≡ σ{Xr, 0 ≤ r ≤ s} h R s h R t+s ii V (Xr)dr V (Xr)dr = Ex e 0 E f(Xt+s)e s |Fs h R s h R t ii 0 V (Xr)dr 0 V (Xr)dr = Ex e EXs f(Xt)e , by the Markov property h R s i V (Xr)dr = Ex e 0 (Σtf)(Xs)
= Σs(Σtf)(x) as required. Because every semigroup has a unique generator and vice versa, we conclude that L + V is the generator of the semigroup Σt. We are now ready to apply the theory we have seen in this section thus far to the Abelian Sandpile Model.
17 5 TOPPLING NUMBERS, AVALANCHES AND RANDOM WALKS
5 Toppling numbers, Avalanches and Random Walks
To analyse the Abelian Sandpile Model in terms of stochastic processes we have to define the state space, semigroup and generator of the model. One fundamental identity is called Dhar’s Formula, which is proven in [3]. It expresses the expected number of topplings of a site y ∈ Vn upon addition of a Dn,Fn particle at x ∈ Vn in a recurrent configuration η ∈ R entirely in terms of ∆ . In this section, we will make this formula precise. Furthermore, we will relate the ASM to the more general theory of Markov Processes, whereby the Feynman-Kac formula will help us greatly. Notice that we are still working d with finite system size Vn = [−n, n] ∩ Z . The following notions will bring us further in analysing the ASM with the help of the Feynman-Kac formula and Dhar’s formula.
PROPOSITION 5.1. When Vn is finite and a system is stabilizable in the sense of Definition 3.2, there will be a finite set of recurrent configurations R ⊆ Ω. The number of configurations in the recurrence class is given by |R| = det (∆Dn,Fn )
PROOF. This is proven in [5], using the burning algorithm and the matrix tree theorem. Redig proves the theorem for the case V = C, but since it is done in full generality with respect to the toppling matrix ∆Dn,Fn , the proof also applies here, in the case of stabilizable systems.
DEFINITION 5.1. Toppling numbers function
For a legal sequence of topplings Tx1 Tx2 ... resulting after addition of a particle at x ∈ Vn (η → η + δx) Q such that the resulting configuration i Txi (η + δ) is stable, the toppling numbers function n : Vn × Vn × H → Z is defined as X n(x, y, η) = I(xi = y) i
In other words, n(x, y, η) is the number of topplings of a site y ∈ Vn necessary to stabilize η + δx.
PROPOSITION 5.2. Finiteness of toppling numbers If a system is stabilizable, the addition operator ax :Ω → Ω is well-defined. That is, it can be written as a composition of finitely many toppling operators and the resulting configuration axη is unique. We can therefore write Dn,Fn axη = S (η + δx) = η + δx − ∆ n(x, ·, η) where n(·, x, η) denotes the vector of toppling numbers resulting from the stabilizing of η + δx.
PROOF. This is proven in by Redig [5], in the classical case. It is entirely analogous to the general case. Two important properties used in the proof remain the same in the case of general systems: firstly, stabilizability guarantees the finiteness of toppling numbers and the ‘off-diagonal’ elements are still negative. Since the addition operator is well-defined for stabilizable systems and group properties are preserved, we have again an invariant measure on the Markov chain of recurrent configurations. This fact is exploited in Dhar’s Formula.
THEOREM 5.1. Dhar’s Formula For a stabilizable system and for η ∈ R,