PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES

HARRY CRANE

Abstract. We study both time-invariant and time-varying Gibbs distributions for configura- tions of particles into disjoint clusters. Specifically, we introduce and give some fundamental properties for a class of partition models, called permanental partition models, whose distribu- tions are expressed in terms of the α-permanent of a similarity matrix parameter. We show that, in the time-invariant case, the permanental partition model is a refinement of the cele- brated Pitman-Ewens distribution; whereas, in the time-varying case, the permanental model refines the Ewens cut-and-paste Markov chains [J. Appl. Probab. (2011), 3:778–791]. By a special property of the α-permanent, the partition function can be computed exactly, allowing us to make several precise statements about this general model, including a characterization of exchangeable and consistent permanental models.

1. Introduction Consider a collection of n distinct particles (labeled i = 1,..., n) located at distinct positions x1,..., xn in some space . Each particle i = 1,..., n occupies an internal state yi, which may change over time. Jointly,X the state of the system at time t 0 depends on the ≥ positions of the particles x1,..., xn and the interactions among these particles. We assume that, with the exception of their positions in , particles have identical physical properties X and, thus, the pairwise interaction between particles at positions x and x0 is determined only by the pair (x, x0). In this setting, we model pairwise interactions by a symmetric function K : [0, 1] with unit diagonal, K(x, x) = 1 for all x , where K(x, x ) X × X → ∈ X 0 quantifies the interaction between particles at locations x and x0. Qualitatively, we regard K as the environment, or medium, equipped on . For a fixed configuration x ,..., xn, we X 1 simply represent the interactions among x ,..., xn by a matrix X := (Xij, 1 i, j n) for 1 ≤ ≤ which Xij := K(xi, xj). For example, the Ising model describes a system with internal states = 1 and interactions determined by the adjacency matrix of a weighted graph. More generally,Y {± } the above scheme describes an interacting particle system. In this paper, we consider interacting particle systems where the configuration is a partition of the particles 1,..., n into disjoint clusters. In particular, we study a family of Gibbs measures on partitions of a finite set. Previously, Gibbs measures for fragmentation processes have appeared [4] in the context of modeling degradation of mass over time. More generally, models of agglomeration and fragmentation have been studied under the heading of coagulation and fragmentation processes, see e.g. [3, 5, 6, 33]. The exchangeable, or mean-field, case has garnered special attention; and, in our setting, the mean-field case

Date: September 19, 2013. 1991 Mathematics Subject Classification. Primary 82C22, secondary 60C99, 60J10. Key words and phrases. Boltzmann-Gibbs measure; canonical Gibbs ensemble; Markovian Gibbs structure; permanental parition model; α-permanent; permanental process. This author is partially supported by NSF grant DMS-1308899 and NSA grant H98230-13-1-0299. 1 2 HARRY CRANE corresponds to an environment that is uniformly homogeneous, i.e. K(x, x0) 1 for all x, x0 ; see Aldous [1] for a review of mean-field models for stochastic coalescence.≡ ∈ X Given an energy function H : R and a temperature β R, a Gibbs distribution on has the form Y → ∈ Y ( ) 1 1 (1) Pβ,H Y = y = exp H(y) , y . { } Zβ,H β ∈ Y Boltzmann-Gibbs distributions arise in various settings in both statistical mechanics and elsewhere. In the special case where is the collection of partitions of [n]:= 1,..., n , the energy function H can be assumedY to have the product partition form; that is,{ for any} partition π, H(π) can be expressed as a sum over the blocks of π. Product partition models were initially introduced by Hartigan [18] in the context of astronomical modeling and have since appeared in several applications involving partition models, including Bayesian nonparametrics. Because of its general form, many common distributions are of product partition type, including the Ewens distribution [14] from population genetics; however, the more general Pitman-Ewens distributions [33] are not strictly of product partition type. In this paper, we modify (1) so as to include both the Pitman-Ewens distributions and the more general permanental partition model, our principle model of interest. Extending (1) to a time-varying model, we define a Markovian Gibbs structure on as a transition probability of the form Y ( ) 1 1 (2) pβ,H(y, y0) = exp H(y, y0) , y, y0 , Zβ,H(y) β ∈ Y where H : R is called the joint energy function and p H(y, y ) gives the probability Y × Y → β, 0 of a transition from y to y0. In specializing to Markovian Gibbs structures on partitions, we restrict H to functions of (π, π ) that can be expressed as a sum over the blocks of π π , the 0 ∧ 0 infimum of π and π0 in the partition lattice. We take up certain statistical and mathematical properties of permanental partition models. In particular, we consider conditions under which permanental models are ex- changeable, consistent and, in the time-varying case, reversible. Reversible models are important in many scientific contexts of temporally evolving physical systems [22]. Many physical phenomena are, at least intuitively, reversible, and so models for such phenomena should reflect this property. Consistency (under subsampling) is important in statistical inference, and reflects a coherence among a collection of probability distributions that per- mits out-of-sample statistical inference. Consistency also results in beautiful mathematical theory involving limiting stochastic processes as the size of the system grows to infinity; see e.g. [6] for an overview of the mathematical theory underlying consistent systems of partition-valued processes and see [27] for a discussion of consistency in statistical models. Consistency under subsampling belies an assumed lack of interference among particles in a physical system. As such, consistency may not be appropriate in some applications and, in keeping, the permanental model is consistent only under special choices of K.

1.1. The α-permanent of a matrix. Gibbs models for both time-invariant and time-varying partition structures have a history in diverse settings and here we focus primarily on perma- nental partition models. Specifically, we introduce both the permanental partition distribution and its related family of permanental partition transition probabilities, whose energy function entails the α-permanent of some similarity matrix. PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 3

The permanent of any real-valued matrix X := (Xij, 1 i, j n) is defined by ≤ ≤ X Yn (3) per X := Xj,σ(j),

σ n j=1 ∈S where n is the acting on [n]:= 1,..., n . In pure mathematics, the permanentS arises in representation theory of the symmetric{ group} as the immanant associ- ated to the trivial representation ρ 1 of n [16, 31]. In probability theory and statistical physics, permanents and determinants≡ appearS in spatial point process models of bosons and fermions, respectively. In contrast to Poisson point processes, permanental and de- terminantal processes incorporate dependence among the points in an aggregated point pattern with clumping (permanental) or repellence (determinantal). Relevant to fermions is the Pauli exclusion principle, by which no two particles can occupy the same quantum state, and, consistent with this principle, any matrix with two or more rows identical has zero determinant; this principle does not apply to bosons, for which the permanent is an appropriate kernel, see [11, 19, 29, 35]. In permanental point processes, X plays the role of a correlation matrix–a higher value of Xij indicates correlation between measurements on particles labeled i and j. In the permanental partition models we study, X plays the role of a similarity matrix–a higher value of Xij indicates that particles i and j are more likely to occupy the same internal state. In a different context, Gyires [17] studied a family of discrete probability distributions related to the permanent; and, more recently, a generalization of the permanent, called the α-permanent, has been connected to some well-known distributions in probability and statistics [37, 38]. For α R, the α-permanent of X is defined by ∈ X Yn #σ (4) perα X := α Xj,σ(j), σ n j=1 ∈S where #σ denotes the number of cycles of σ n. Valiant [36] first showed the computa- tional complexity of the permanent is #P-complete;∈ S and, since this seminal work, several algorithms have been proposed for its approximation, e.g. [21, 26, 28]. It has been recently conjectured [10] that, aside from α a negative integer, computation of the α-permanent is at least as complex as the permanent.

1.2. Permanental Gibbs models. While the permanent has received substantial attention in the literature, the α-permanent has been mostly neglected, except by a few authors. In this paper, we draw on ideas from statistical mechanics (Bolzmann-Gibbs distributions and permanental processes), applied probability (the Ewens sampling formula) and linear alge- bra (the α-permanent) to derive a natural family of distributions and transition probabilities on partitions of a finite set [n]. Throughout the paper, we generically assume that particles i = 1,..., n occupy fixed positions in some abstract medium , and X := (Xij, 1 i, j n) X ≤ ≤ defines a similarity matrix so that Xij quantifies the similarity (or interaction) between particles i and j. Throughout the paper, we assume that X satisfies

Xii = 1 for all 1 i n, • ≤ ≤ 0 Xij 1 for 1 i , j n, and • ≤ ≤ ≤ ≤ Xij = Xji for all 1 i, j n. • ≤ ≤ For every α > 0, the α-permanent of any similarity matrix is strictly positive, as is the α-permanent of X[b] for every subset b [n], where X[b]:= (Xij, i, j b) is the submatrix of ⊆ ∈ 4 HARRY CRANE

X with rows and columns indexed by b. Therefore, we can use the α-permanent to weight each partition of [n], which determines a probability measure by normalization. Let π be a partition of [n] for which we write b π to denote that b is a block of π. Then, in the Boltzmann-Gibbs form, the permanental∈ distribution can be expressed as   X  α,β,X a(β, π)   (5) Pn (π) = exp  log perα X[b] . Zα,β,X   b π  ∈ In a similar way, we define the permanental transition probability from partition π to π0 by    X  α,β,X a(β, π0)   (6) pn (π, π0) = exp  log perα X[b] . Zα,β,X(π)   b π π  ∈ ∧ 0 In both of these expressions, the additional factor a( , ) arises when considering Gibbs measures on the space of set partitions, and is consistent· · with previous literature on Gibbs partitions, e.g. Chapter 1.5 of [33].

1.3. Computing the partition function. In statistical mechanics, the partition function Zβ,H in (1) plays an important role in understanding Gibbs measures in thermodynamic equilibrium. However, except in special cases, the partition function cannot be computed explicitly. A remarkable aspect of the permanental partition models is that the partition function can always be evaluated in closed form because of the Permanent Decomposition Theorem from [10] (reproduced as Theorem 3.1). In this case, (5) and (6) are given by Q α,β,X #π b π perα X[b] Pn (π):= β↓ ∈ and(7) perαβ X Q Y per X[b b ] α,β,X #π b0 π0 α 0 (8) p (π, π0):= β↓ 0 ∈ ∩ , n per X[b] b π αβ ∈ respectively. Knowledge of the partition function allows us to study this family of models in more detail than general Gibbs measures. For example, we observe an intimate connection between α,β,X Pn in (7) and the Pitman-Ewens distribution from population genetics [14, 33]; we discuss this connection further in Section 4. The permanental distributions can, therefore, be viewed as a three-parameter extension to the Pitman-Ewens two-parameter model. Analogously, the permanental transition probabilities (8) refine the self-similar cut-and-paste chains studied in [8, 9]; we discuss this further in Section 5.

1.4. Summary of main theorems. Far from a merely algebraic observation, the permanental partition families exhibit some fundamental modeling properties, adding to their practical appeal. Furthermore, any potential applications further emphasize the importance of understanding the α-permanent, which has not been well-studied in the literature. Before going through our main discussion, we highlight our main theorems and explain the structure of the paper. Our main results establish fundamental properties of both (7) and (8). Though we have not yet formally defined the parameter space of either (7) or (8), we assume, throughout this section, that (α, β, X) is a valid parameter choice, i.e. (7) and/or (8) determine a on [n]. We formally define the parameter space for these models in Definitions 4.6 and 5.8. P PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 5

Note that, in Theorems 4.9 and 5.12, we say X corresponds to an infinite partition of N if Xij := I i j for some equivalence relation on N. { ∼ } ∼

α,β,X 1.4.1. Gibbs partition structures. Clearly, by (5), Pn in (7) has the Gibbs form. Furthermore, α,β,1n for X = 1n, the matrix of all ones, we see that P coincides with the Pitman-Ewens( α, αβ)- n − law on [n], discussed in detail in Section 4.1. In general, the Pitman-Ewens distribution characterizesP the class of consistent canonical Gibbs ensembles on ( , n N). P[n] ∈ X Theorem 4.8: The permanental partition structure Pα,β, in (7) is exchangeable if • n and only if there is some 0 d 1 such that X satisfies Xij = d for all i , j and Xii = 1 for all i = 1,..., n. ≤ ≤ Theorem 4.9: Let X := (Xij, i, j 1) be an infinite array corresponding to an infinite • ≥ partition of N and, for each n N, let Xn := (Xij, 1 i, j n) denote its leading n n ∈ α,β,Xn ≤ ≤ × submatrix. Then the collection (Pn , n N) is consistent under subsampling. α,β,Xn ∈ Corollary 4.11: Let (Pn , n N) be a collection of permanental partition struc- • ∈ α,β,Xn tures such that, for each n N, Pn is a probability distribution on [n]. Then α,β,X ∈ P (P n , n N) is both exchangeable and consistent if and only if β N and, for n ∈ ∈ either d = 0 or d = 1, Xn := (Xij, 1 i, j n), where (Xij, i, j 1) is an infinite array ≤ ≤ ≥ with entries Xij = d if i , j and Xii = 1 for all i N. ∈ 1.4.2. Markovian Gibbs partition structures. α,β,X Theorem 5.9: For all (α, β, X), the permanental Markovian Gibbs structure pn in (8) • is reversible with respect to the permanental partition structure (7) with parameter (αβ, β, X). X Theorem 5.10: The permanental Markovian Gibbs structure pα,β, in (8) is exchange- • n able if and only if there is some 0 d 1 such that X satisfies Xij = d for all i , j ≤ ≤ and Xii = 1 for all i = 1,..., n. Theorem 5.12: Let X := (Xij, i, j 1) be an infinite array corresponding to an infinite • ≥ partition of N and, for each n N, let Xn := (Xij, i, j 1) denote its leading n n α,β,∈ Xn ≥ × submatrix. Then the family (pn , n N) is consistent under subsampling. α,β,Xn ∈ Corollary 5.13: Let (pn , n N) be a family of permanental Markovian Gibbs • ∈ α,β,Xn structures such that, for each n N, pn is a transition probability on [n]. Then α,β,X ∈ P (p n , n N) is exchangeable and consistent if and only if β N and, for either n ∈ ∈ d = 0 or d = 1, Xn := (Xij, 1 i, j n), where (Xij, i, j 1) is an infinite array with ≤ ≤ ≥ entries Xij = d if i , j and Xii = 1 for all i N. ∈ In addition to the above properties, the permanental partition models are equivariant with respect to the symmetric group, and we can construct an exchangeable model by mixing σ with respect to a random matrix X for which X = X for all σ n, where = denotes equality in law. We take these latter two propertiesL up briefly in∈ Section S 6. BasedL on this summary, the permanental partition model has no a priori obvious modeling defects; that is, it is equivariant with respect to the natural action of the symmetric group by relabeling particles and, specific to the Markov chain setting, the permanental Markovian Gibbs structure is reversible. Moreover, we can derive conditions under which exchangeability and consistency are satisfied (both jointly and separately). Finally, far from an ad hoc choice 6 HARRY CRANE based on the identity (19), the model extends the celebrated Pitman-Ewens family and, furthermore, the extra parameter X offers an intuitive interpretation in several applications. In Section 2, we give some preliminaries for partitions and random partition models. In Section 3, we present some preliminary properties of the α-permanent. In Section 4, we α,β,X discuss the permanental family of distributions Pn introduced above. In Section 5, we discuss the permanental family of transition probabilities. In Section 6, we briefly discuss the properties of equivariance and lack of interference in the context of the permanental partition model, as well as a related family of distributions on the space of k-colorings.

2. Preliminaries: random set partitions

2.1. Partitions. For n N, a partition π of [n]:= 1,..., n is a collection π , . . . , πr of ∈ { S } { 1 } non-empty, non-overlapping subsets (blocks) such that πj = [n]. We write to denote j P[n] the collection of all partitions of [n]. Equivalently, we can define a partition of [n] as an equivalence relation , where ∼π i j i and j are in the same block of π. ∼π ⇐⇒ From the definition of , we may represent π as an n n matrix (πij, 1 i, j k) with ∼π × ≤ ≤ πij = 1 if i π j and πij = 0 otherwise. Throughout the paper, the equivalence of these three representations∼ (as a collection of subsets, as an equivalence relation and as a matrix) is understood, and we need not specify to which representation we are referring. For each n N, [n] is partially ordered by , called sub-partition. In particular, for π, π , we∈ defineP the infimum of π and π by≤ 0 ∈ P[n] 0 π π0 := b b0 : b π, b0 π0 . ∧ { ∩ ∈ ∈ } \ {∅} We then say that π is a sub-partition, or refinement, of π , written π π , if and only if 0 ≤ 0 π = π π0. We also write #π to denote the number of blocks of π [n]. Given∧ π and m n, we define the restriction of π to [m] by∈ P ∈ P[n] ≤ (9) Dm,n π = π [m] := b [m], b π , | { ∩ ∈ }\{∅} the partition obtained by restricting each block of π to [m] and removing any empty sets. Under this restriction operation, the system ( , n N) is projective and determines the P[n] ∈ limiting space of partitions of N. The projective structure of ( [n], n N) is logically important in statisticalP∞ applications and allows one to build processesP on the∈ abstract space by specifying a consistent collection of finite-dimensional marginal distributions. P∞ 2.2. Random partitions. Random partitions and partition-valued processes arise in popu- lation genetics [14, 24, 25] to model evolutionary processes as well as in the physical sciences, e.g. [1], to model clustering of particles. In addition to these applications, the theory of exchangeable partitions is deeply connected to classical theory of stochastic processes, such as Brownian motion and Levy´ processes, see [6, 33]. Specializing (1) to , we define a Gibbs measure on by the form P[n] P[n]   X  a(π)   (10) Pn Π = π := exp  H(b) , π [n] . { } Za,H   ∈ P b π  ∈ The additional factor a(π) in (10) arises when we induce a Gibbs measure on [n] by projecting from a richer space. For example, if the internal states of a system areP given PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 7

1 n [n] [n] explicitly by x x Z , a Gibbs measure on Z induces a measure on [n] through the relation ··· ∈ P i and j are in the same block of π xi = xj. ⇐⇒ In projecting to [n], the Gibbs form from (1) becomes (10). A special case of (10) is the so-called canonicalP Gibbs ensemble on which, for sequences of non-negative coefficients P[n] v := (vn, n N) and w := (wn, n N), has the form • ∈ • ∈ v#π Y (11) Pn(π; v , w ):= w#b, π [n] . • • Bn(v , w ) ∈ P • • b π ∈ In (11), Bn(v , w ) is the nth Bell polynomial associated to (v , w ). We call Pn( ; v , w ) • • • • · • • the Gibbs[n](v , w ) distribution on [n]. We also call Gibbs distributions of the form (11) • • P mean-field, meaning that the probability of any partition π [n] depends only on its block sizes. In the probability literature, mean-field models are∈ called P exchangeable and reflect invariance of Pn( ) under relabeling of the particles. In particular, for any permutation · σ :[n] [n], we define the image of π by σ by the relation σ , where → ∼π i σ j σ(i) σ(j). ∼π ⇐⇒ ∼π σ A distribution Pn on [n] is called exchangeable if Pn Π = π = Pn Π = π for all permutations σ :[n] [n]. Clearly,P a canonical Gibbs ensemble{ (11)}is exchangeable{ } for any choice of (v , w ).→ •When• the weight sequences v and w are comprised of non-negative integers, the • • canonical Gibbs[n](v , w ) ensemble has a natural physical interpretation ([33], Chapter 1.5): • • any Π Gibbs[n](v , w ) can be regarded as a random (V W)-composite structure. By (V W)-composite∼ structure• • we mean that Π is taken as a uniform◦ draw from the set of all marked◦ partitions of [n], where each π with k blocks and block sizes n n > 0 ∈ P[n] 1 ≥ · · · ≥ k is assigned one of vk V-structures, and each block of π of size ni is assigned one of wni W-structures, independently of other blocks. The distribution of Π is then obtained by forgetting the (V W)-structure and looking only at the induced partition of [n]. ◦ 2.3. Integer partitions. A partition of the integer n N, which we shall call an integer ∈ partition of n, is a list n := (m1,..., mn) of multiplicities for which mj 0, for all j = 1,..., n, Pn ≥ and jmj = n. We write n to denote the collection of integer partitions of n. Any j=1 P π [n] induces an integer partition n(π) of n N through the map n : [n] n, ∈ P ∈ PPn → P π n(π):= (m ,..., mn), where mj is the number of blocks of π of size j and mj = #π. 7→ 1 j=1 Aside from its natural combinatorial and physical interpretations, a canonical Gibbs[n](v , w ) • • distribution (11) depends on π only through n(π). To each n n, there are ∈ P[n] ∈ P n! Qn mj j=1(j!) mj! set partitions of [n] in π : n(π) = n . Hence, given a probability distribution P N = { ∈ P[n] } { ·} on n, we can generate a random partition of [n] by first drawing its block sizes from P NP= and, given N = n, we draw uniformly from π : n(π) = n . In some cases, { ·} { ∈ P[n] } as we see below, this induced distribution on [n] can be expressed as a canonical Gibbs ensemble. P Example 2.1 (Ewens sampling formula). The Ewens sampling formula was first introduced by Ewens [14] as a probability distribution on integer partitions induced by allele frequencies in 8 HARRY CRANE population genetics. For fixed θ > 0, the Ewens sampling formula on n is given by P P n m θ j=1 j n! (12) Pn N = n = , n := (m1,..., mn) n . n Qn mj { } θ↑ j=1 j mj! ∈ P From the above discussion, the Ewens sampling formula induces the Ewens(θ) distribution on , P[n] #π 0,θ θ Y P Π = π = (#b 1)!, π [n] . n { } θ n − ∈ P ↑ b π ∈ Note that the Ewens(θ) distribution is of canonical Gibbs type for v := θ• and w := ( 1)!. The Pitman-Ewens distribution is a two-parameter extension of• the Ewens distribution• • − which, for suitable choices of (α, θ), is defined by #π Y α,θ (θ/α)↑ #b P Π = π = ( α)↑ , π [n], n { } θ n − − ∈ P ↑ b π ∈ where α j := α(α + 1) (α + j 1). Note that the Pitman-Ewens distribution is a canonical Gibbs ↑ ··· − ensemble with v := (θ/α)↑• and w := ( α)↑•. We discuss the Pitman-Ewens family in more detail in Section• 4.1. • − − 2.4. Markovian Gibbs partition structures. Following the discussion in Section 2.2, we introduce Markovian Gibbs structures for time-varying Gibbs partitions. A Markovian Gibbs structure is a transition probability measure on of the form P[n]   X X  a(π0)   (13) pn(π, π0):= exp  H(b, b0) , π, π0 [n], Za,H(π)   ∈ P b π b π  ∈ 0∈ 0 where H : 2[n] 2[n] R is the joint energy function. The form in (13) is natural as, in some cases, it admits× an interpretation→ of independent partitioning of each of the blocks of π and, subsequently, merging blocks together to obtain π0. The self-similar cut-and-paste chain is an example of a Markovian Gibbs structure with this interpretation.

Example 2.2 (Self-similar cut-and-paste chain). For k 1, let [n]:k denote the collection of partitions of [n] having at most k blocks. For α > 0, Crane [8≥] introducedP a family of Markov chains on with transition probabilities P[n]:k Q Y (α/k) #(b b0) #π0 b0 π0 ↑ ∩ (14) pn(π, π0) = k↓ ∈ , π, π0 [n]:k, α #b ∈ P b π ↑ ∈ j #π where k↓ := k(k 1) (k j + 1). Note that (14) has the form of (13) for a(π0):= k↓ 0 , #(−b b0)··· − Q #b H(b, b0):= (α/k)↑ ∩ and Za,H(π):= b π α↑ . Note also that (14) can be expressed as ∈ #π #π0 k 0 Y k↓ b Y p (π, π ) = ↓ | (α/k) #(b b0), π, π , n 0 Q π #b ↑ ∩ 0 [n]:k 0b α ∈ P b π k↓ | b π ↑ b π ∈ ∈ 0∈ 0 which has a straightforward interpretation. First, we independently partition each block of π α/k,α according to the Pitman-Ewens( α/k, α) distribution (Pn− from Example 2.1); we then label each of the sub-blocks within each− block distinctly in [k] by drawing uniformly without replacement; finally, we put all sub-blocks assigned the same label into a single block in π0 and remove labels. PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 9

Motivated by the above interpretation of the cut-and-paste transition probabilities and the cut-and-paste representation of exchangeable Feller chains in [9], we focus on joint energy functions depending on the pair of subsets (b, b0) only through the intersection b b . Under this convention, we define a canonical Markovian Gibbs structure, or mean-field ∩ 0 Markovian Gibbs structure, on [n] by the form P Y v#π0 (15) pn(π, π0; v , w ):= w#b, π, π0 [n] . • • Bn(v , w , π) ∈ P • • b π π ∈ ∧ 0 In the case where both v and w are integer-valued, (15) has an analogous interpretation, in terms of (V W)-composite• structures,• as in the one-dimensional case above. We call ◦ a Markovian Gibbs structure pn( , ) exchangeable if it is invariant under joint relabeling of · · σ σ the argument (π, π0), that is, pn(π, π0) = pn(π , π0 ) for all permutations σ :[n] [n], for all σ σ σ → π, π0 [n]. Because, for any σ :[n] [n], π π0 = (π π0) , it is clear that any canonical Markovian∈ P Gibbs ensemble (15) is exchangeable.→ ∧ ∧ 2.5. Model properties. We consider several fundamental properties of partition models in this paper. We have already mentioned exchangeability in Sections 2.2 and 2.4. Later, in Section 6, we discuss the property of equivariance, which pertains to models that associate covariates to the particles of the system. In the permanental models we study, the covariates are the positions of particles in space, resulting in a parameter X that is not invariant under relabeling. However, if we relabel the particles and the similarity matrix jointly by the same permutation, the model is unchanged. Other properties include consistency under subsampling and reversibility, which we now discuss. 2.5.1. Consistency under subsampling. A consideration that arises in statistical inference, and elsewhere when studying incompletely observed systems, is consistency under subsampling. A family of measures Pn, n N , where Pn is a probability measure on [n] for every n N, is consistent if, for every{ m ∈n, } P ∈ ≤ 1 (16) Pm(π) = Pn(D− (π)), for all π . m,n ∈ P[n] Particularly in Bayesian posterior predictive inference, a consistent family Pn, n N { ∈ } determines a predictive statistical model in that, given an observation π [m], the model gives a set of conditional probability distributions on the unobserved part∈ of P the population by ( Pn(π0)/Pm(π), π = Dm,n π0 Pn m(π0 π):= | | 0, otherwise. Consistent collections on ( [n], n N) play a particularly important role in Bayesian classification and clusteringP [30]. In∈ classical probability theory, consistent collections of probability measures determine a unique probability measure on some limiting space, in this case the space of partitions of N. It is known that Gibbs distributions are not consistent in general,P see∞ e.g. [34] for an illustration in the case of the Exponential Random Graph Model. For a Markov chain, consistency under subsampling corresponds to the condition that, for every m n, the restriction to of a -valued Markov chain is, itself, a Markov chain. ≤ P[m] P[n] Since the projection operation Dm,n : [n] [m] is many-to-one, it does not immediately follow that the image of a Markov chainP by→ restriction P will maintain the Markov property. By a theorem of Burke and Rosenblatt [7], it is necessary and sufficient that, for every m n, ≤ 1 1 (17) Pm(π, π0) = Pn(π∗, D− (π0)) for all π∗ D− (π), m,n ∈ m,n 10 HARRY CRANE for every π, π0 [m]. Under this condition, there exists a transition probability measure on and, hence,∈ P a Markov chain on with that transition probability measure. P∞ P∞ 2.5.2. Reversibility. A further property, relevant in the time-varying setting, is reversibility, whereby the law governing the evolution of a physical system is invariant under time- reversal. Given a Markov transition probability pn( , ) on and a measure µn( ) on , · · P[n] · P[n] pn is reversible with respect to µn if it satisfies the detailed-balance equation

(18) µn(π)pn(π, π0) = µn(π0)pn(π0, π) for all π, π0 . ∈ P[n] If (18) holds for a probability measure µn, then µn is the stationary distribution for any Markov chain governed by pn( , ). · · 3. Preliminaries: the α-permanent Throughout this section, let X be an R-valued n n matrix and, for α R, recall the definition of the α-permanent in (4). We note that some× authors, e.g. [19], prefer∈ to define the α-determinant detα X by X Yn n #σ detα X := α − Xj,σ(j),

σ n j=1 ∈S n which relates to (4) through perα X = α det1/α X. For all α 0, any X with Xij 0 for all 1 i, j n has a non-negative α-permanent and, moreover,≥ the α-permanent≥ of every submatrix≤ ≤ of X is non-negative. Non-negativity of perα X[b] for all b [n] plays the same role as positive definiteness in the related study of determinantal distributions⊆ on subsets of [n]. In particular, when X is positive definite, det X[b] 0 for all b [n] because all leading minors of a positive definite matrix are positive and the determinant≥ ⊆ is invariant under conjugation by a permutation matrix. Furthermore, P the normalizing constant is given by b [n] det X[b] = det(In + X), where In is the n n identity matrix (see e.g. Theorem 1.2 of [10]).⊆ × Theorem 3.1 (Theorem 1.1 in [10]). Let α, β C and M be any n n C-valued matrix. Then the α-permanent satisfies the identity ∈ × X #π (19) per M = β↓ per (M π), αβ α · π ∈P[n] where π := (πij) is regarded as a 0-1 valued matrix representing a partition of [n], and M π := · (Mijπij, 1 i, j n) is the Hadamard product of M and π. In particular, ≤ ≤ Y per (M π):= per M[b]. α · α b π ∈ From a purely algebraic perspective, (19) is interesting because it gives an explicit relationship between the α-permanent and the determinant: X n #π per X = ( 1) ( α)↓ det(X π). α − − · π ∈P[n] From a statistical mechanics and probabilistic perspective, (19) gives an explicit formula for the normalizing constant of a family of probability distributions on set partitions: provided each term on its right-hand side is non-negative and perαβ X > 0, (19) determines X X a probability distribution Pα,β, on , as given in (7). We call Pα,β, the permanental partition n P[n] n PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 11 distribution, or permanental partition structure, with parameter (α, β, X). In general, many combinations of α, β and X determine a probability distribution on [n] through (7); however, in the coming sections, we focus on a parameterization that admitsP a straightforward interpretation. Similarly, we can define a transition probability on as in (8). P[n] 4. Permanental partition structures 4.1. The Pitman-Ewens model. The Pitman-Ewens distribution is, perhaps, the most well- known and well-studied family of distributions on set partitions. In surprising ways, the Pitman-Ewens distribution arises in diverse applied and theoretical problems involving exchangeable random partitions. Here we show its close connection to the α-permanent, which connects the Pitman-Ewens family to the permanental partition structures in (7). For θ > 0 and n N, let us define ∈ #π 0,θ θ Y (20) P (π):= (#b 1)!, π [n], n θ n − ∈ P ↑ b π ∈ n 0,θ where θ↑ := θ(θ+1) (θ+n 1). The distribution Pn is called the Ewens distribution with parameter θ (as in Example··· 2.1).− Viewed as a model in population genetics, the parameter θ relates to the mutation rate, as higher values of θ correspond to a larger expected number of blocks and, hence, more mutation. There are numerous connections between (20) and topics in probability theory, number theory and . In probability theory alone, there are connections between the Ewens distribution and Brownian motion, Levy´ processes and the Poisson-. See [6, 33] for recent surveys of this work. The next two examples illustrate the diverse scenarios in which (20) arises. Example 4.1 (Ewens distribution in statistics). In statistics, the Ewens distribution arises, at least implicitly, in a species sampling context; see e.g. [13, 15]. In this setting, let X1, X2,... be independent Poisson random variables so that Xj has mean θ/j, for θ > 0. For each j 1, we interpret X as the number of species appearing j times in a sample of n specimens. The conditional≥ j P distribution of X1,..., Xn, given j jXj = n, is easily computed by   P  X  P j xj   θ 1/j θ P (X ,..., X ) = (x ,..., x ) jx = n e− j ,  1 n 1 n j  Q xj  |  ∝ j xj! j j which normalizes to the Ewens sampling formula (12) on integer partitions of n; see Example 2.1. The Ewens sampling formula with parameter θ gives the distribution of the block sizes of Π Ewens(θ), from which the mean number of blocks of Π is readily observed to be ∼ Xn Xn E Xj = θ 1/j θ log n. ∼ j=1 j=1 Example 4.2 (Ewens partitions in number theory, see [12]). For a number theoretic connection, consider drawing a uniform random integer N from [n]. For each i = 1,..., n, let Fi be the ith largest prime factor of N, where we put Fi 1 if i exceeds the number of prime factors of N. Then the ≡ 1 logarithmically transformed sequence (log n)− (L1, L2,...), where Li = log Fi, converges in law to the Poisson-Dirichlet distribution with parameter (0, 1) [2]. In general, the Ewens(θ) distribution in (20) corresponds to the Poisson-Dirichlet distribution with parameter (0, θ) through Kingman’s paintbox correspondence [23]. 12 HARRY CRANE

In addition to the previous two examples, the Ewens distribution is observed in other diverse mathematical and scientific contexts. In this paper, we highlight a new connection between permanental partition structures and the two-parameter Pitman-Ewens law. For (α, θ) satisfying either α < 0 and θ = kα for some k N or • 0 α 1 and θ− > α, ∈ • ≤ ≤ − the Pitman-Ewens distribution with parameter (α, θ) is given by #π Y α,θ (θ/α)↑ #b (21) P (π) = ( α)↑ , π [n] . n θ n − − ∈ P ↑ b π ∈ α,θ For (α, θ) in the parameter range of the Pitman-Ewens model, the collection (Pn , n N) is both exchangeable and consistent, and, therefore, determines a unique probability measure∈ Pα,θ on . P∞ 4.2. Gibbs partitions. In specializing the Boltzmann-Gibbs measure (1) to the space of set partitions, we recall that Gibbs distributions on [n] have the form  P    a(π) X  (22) Pn(π):= exp  H(b) , π [n], Za,H   ∈ P b π  ∈ [n] for functions a : [n] R and H : 2 R. Immediately, comparing (21) and (22), we see P → → #π that the Pitman-Ewens(α, θ) distribution is a Gibbs partition measure with a(π):= (θ/α)↑ , #b n H(b):= log[ ( α)↑ ] and Za,H := θ↑ . In the special case α = 0, the Pitman-Ewens(α, θ)- law reduces− to− the Ewens(θ)-law (20). Furthermore, the Pitman-Ewens(α, θ)-law (21) is a canonical Gibbs[n](v , w ) measure (11) with v := (θ/α)↑• and w := ( α)↑•. The Ewens(θ) • • • • − − distribution is a Gibbs[n](v , w ) measure with v = 1• and w := ( 1)!. In the following theorem, we recall the characterization• • of the Pitman-Ewens• (α,• θ) distribution• − as the unique exchangeable and consistent Gibbs model on . P[n] Theorem 4.3. For weight sequences v , w , let (Pn( ; v , w ), n N) be a family of canonical Gibbs • • · • • ∈ measures. Then (Pn( ; v , w ), n N) is consistent if and only if there exists (α, θ) in the parameter · • • ∈ space of the Pitman-Ewens model such that v = (θ/α)↑• and w = ( α)↑•. • • − − v ,w Proof. Regard Pn• • as a generic Gibbs[n](v , w ) distribution with weights v , w : • • • • v ,w v#π Y Pn• • (π):= w#b, π [n] . Zn ∈ P b π ∈ v ,w Consistency of the family (P • • , n N) requires that, for every n N, n ∈ ∈ v ,w X v ,w P • • (π) = P • • (π0) for all π . n n+1 ∈ P[n] π D 1 (π) 0∈ n−,n+1 In the case of the canonical Gibbs[n](v , w ) distribution (11), consistency reduces to • • v Y X v Y #π w = #π0 w Z #b Z #b n 1 n+1 b π π D (π) b π0 ∈ 0∈ n−,n+1 ∈    Y Y X  1   = v#π+1w1 w#b + v#π w#b W#b , Zn+1   b π b π b π ∈ ∈ ∈ PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 13 where Wi := wi+1/wi for i 1. Now, suppose #π = k 1, then the above identity is identical to ≥ ≥ Zn+1 X = Vk + W#b, Zn b π ∈ for all k 1, where Vk := vk+1/vk. Since the left-hand side does not depend on π, we must have, for≥ every i, j 1, ≥ Wi + Wj = Wi 1 + Wj+1, − and, hence, Wi Wi 1 = Wj+1 Wj for all i, j 1. Therefore, Wi is of the form Wi = α + βi for real constants− α and− β. Thus,− ≥

Zn+1 = Vk + αk + βn, Zn and so Vk + αk is constant in k. We, therefore, may write Vk = θ αk, for some constant θ such that V 0 for all k 1 and V > 0 for k = 1. Hence, − k ≥ ≥ k k 1 ( k k Y− α (θ/α)↓ , α , 0 v := Vj = (θ α(k 1)) θ = k − − ··· θk, α = 0. j=0

For 0 < α < , vk 0 for all k 1 if and only if θ/α is a positive integer. Furthermore, for each j 1, ∞ ≥ ≥ ≥ j 1 ( Y− j  (j 1) β α/β + 1 ↑ − , α , 0 wj = w Wi = β(α + β) (α + (j 1)β) = 1 ··· − (j 1)!, α = 0, i=1 − where we have written w1 = β. By the form of (11), we can force β = w1 = 1 so that (j 1) wj = (α + 1) for α , 0. We must have α + 1 0 and, hence, α 1. For 1 α 0, the ↑ − ≥ ≥ − − ≤ ≤ condition v1 > 0 forces θ > α. This completes the proof.  In addition to Theorem 4.3, the Ewens(θ) distribution is of type with canonical sufficient statistic #π and natural parameter log θ. These observations point to why the Ewens distribution appears so widely in applications involving partitions. Not only are exponential family models widely used in statistical applications, but exchangeability and consistency are also important to the logical structure of statistical models [27]. Above, we have also recounted how the Ewens and Pitman-Ewens families arise out of purely mathematical concerns (Example 4.2). In the next section, we point out another appearance of the Pitman-Ewens family that relates to the permanental partition distribution (7).

4.3. Permanental partition structures. In (7), the permanental partition structure with parameter (α, β, X) is defined by

α,β,X #π perα(X π) (23) Pn (π):= β↓ · , π [n], perαβ X ∈ P for some choice of α, β R and X Rn n. By identity (19), (23) is a probability distribution ∈ ∈ × on [n], provided (23) is non-negative for every π [n] and perαβ X > 0. P ∈ P #π In comparision to (22), (23) is of Gibbs-type with a(π):= β↓ , H(b):= log perα X[b] and Z X := per X. Writing 1n to denote the n n matrix of all ones, we observe that α,β, αβ × 14 HARRY CRANE

n perα 1n = α↑ ; hence, for X = 1n and an appropriate choice of (α, β), (23) reduces to Q #b α,β,1n #π b π α↑ P ( ) = ∈ n π β↓ n , π [n], (αβ)↑ ∈ P the Pitman-Ewens( α, αβ) distribution in (21). This observation highlights yet another appearance of the Pitman-Ewens− distribution out of a seemingly unrelated algebraic obser- vation.

Proposition 4.4. For n N, let α > 0, β [n] [n, ) and X be an n n R+-valued matrix such that per X > 0. Then (23)∈ is a probability∈ distribution∪ ∞ on . × αβ P[n] #π Proof. If β [n] [n, ), then β↓ 0 for all π [n]. By (19) and the assumption per X > 0,∈ (23) is∪ a probability∞ distribution≥ on .∈ P  αβ P[n] For the rest of the paper, we always assume α > 0 and β [n] [n, ) when discussing ∈ ∪ ∞ the permanental distribution (23) on [n]. However, by our next proposition, the family α,β,X P P X of distributions on is not identifiable when X is allowed to vary over the { n }α,β, P[n] entire space of R+-valued matrices. For example, for fixed α, β, the matrices X and aX, where a > 0, determine the same probability distribution through (23), prompting our restriction of X to the space of similarity matrices. In particular, we call an R+-valued matrix X := (Xij, 1 i, j n) a similarity matrix if ≤ ≤ (i) Xii = 1 for all i = 1,..., n, (ii)0 Xij 1 for all 1 i , j n, and ≤ ≤ ≤ ≤ (iii) Xij = Xji for all 1 i, j n. ≤ ≤ Although there is, in general, no harm in allowing off-diagonal values to exceed diagonal values, the interpretation of X as a measurement of pairwise similarity suggests that Xij Xii Xjj for all i and j. Symmetry of X is also needed for identifiability. To be clear, only≤ (i) and∧ (iii) are necessary to make the model identifiable; item (ii) arises out of modeling concerns and not mathematical necessity. Proposition 4.5. Let X be any n n non-negative matrix with strictly positive diagonal entries × X11,..., Xnn and define X0 from X by

(24) X0 := Xij/Xii, 1 i, j k. ij ≤ ≤ Let α, β satisfy the conditions of Proposition 4.4. Then the triples (α, β, X) and (α, β, X0) determine the same probability law on through (23). P[n] a Proof. For a generic matrix M := (Mij, 1 i, j n) and constants a1,..., an > 0, let M := (Ma , 1 i, j n) denote the matrix with values≤ ≤ ij ≤ ≤ a M := aiMij, 1 i, j k. ij ≤ ≤ a Qn Then, for α R, per M := per (M) aj. Therefore, for X defined in (24), we have ∈ α α j=1 Qn per (X π) per (X0 π) j=1 Xjj α,β,X #π α #π α · α,β,X0 Pn (π) = β↓ · = β↓ Qn = Pn (π), perαβ X perαβ(X0) j=1 Xjj for every π .  ∈ P[n] PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 15

Definition 4.6 (Permanental partition distribution). For n N, α > 0, β [n] [n, ) and α,β,X ∈ ∈ ∪ ∞ any similarity matrix X, we call Pn in (23) the Pern(α, β, X)-partition structure, or Pern(α, β, X) α,β,X distribution, with diversity α, mixing coefficient β and similarity matrix X. We call Pn the permanental family of distributions on . { } P[n]

4.3.1. Parameterizing the Pern(α, β, X)-family. There are some natural sub-parameter choices for X that may be appropriate in specialized settings. Per our discussion in Section 1, we interpret X as a parameterization of the medium, or environment, in which particles are suspended. For example, we refer to the case X = 1n as a uniformly homogeneous medium, which reflects the assumption that every particle, regardless of its location, interacts with every other particle (including itself) in the same way. We have seen that the permanental partition model in a uniformly homogeneous medium corresponds to the Pitman-Ewens model. We can generalize the uniformly homogeneous model to a homogeneous medium, in which case X corresponds to some symmetrically homogeneous matrix, H[1, d; n], 0 d 1, where ≤ ≤ 1 d d   d 1 ··· d   (25) H[1, d; n]:= . . ···. . . . . .. . . . . d d 1 ···

The special case d = 0 corresponds to X = In, the n n identity matrix, which we refer to as a discrete medium. In a discrete medium, each particle× interacts only with itself. Not covered by the above homogeneous media cases is the case of non-overlapping uniformly homogeneous sub-populations, in which we can let X correspond to a partition π of [n] by the specification ( 1, i j X = π ij 0, otherwise.∼

In this case, each block of the partition acts as a uniformly homogeneous medium over a sub-population, but particles in different sub-populations do not interact. By a further generalization to the non-overlapping sub-population case, we can define, for each b π, a parameter 0 d 1 so that, within each sub-population, particles interact as if suspended∈ ≤ b ≤ in a homogeneous medium described by H[1, db;#b], but particles in disjoint blocks do not interact. As we progress, we discuss these parameter choices in some detail. Another choice for X, natural in phylogenetic modeling, is the case where X corresponds to a rooted-tree matrix. Within our parameterization of X, a rooted-tree matrix is a similarity matrix satisfying the additional three-point condition

Xij X X for all 1 i, j, k n. ≥ ik ∧ jk ≤ ≤ Under the three-point condition, the matrix X admits the interpretation as a rooted tree with leaves labeled in [n] and distance Xij from the root to the point at which elements i and j branch apart in the tree. In this case, we regard X as a parameter in a phylogenetic model, not a medium in a physical sense. This choice of X hints at another potential application of permanental partition structures, which fits well with previous applications of partition models. We do not treat this case here. 16 HARRY CRANE

4.3.2. Mean-field permanental partition structures. The α-permanent of H[1, d; n], defined in (25), was computed in [10] by Xn Xn k n l (26) perα H[1, d; n] = c(n, k, l)α d − , k=1 l=0 where c(n, k, l), called a generalized rencontres number, is the number of permutations of [n] having exactly k cycles and exactly l fixed points. For 0 l k n, the generalized rencontres numbers satisfy the recursion ≤ ≤ ≤ c(n, k, l) = c(n 1, k 1, l 1) + (n l 1)c(n 1, k, l) + (l + 1)c(n 1, k, l + 1), − − − − − − − and, otherwise, c(n, k, l) = 0. Therefore, for 0 d 1, we can express the Pern(α, β, H[1, d; n]) distribution by ≤ ≤ Q P#b P#b k #b l α,β,d b π k=1 l=0 c(#b, k, l)α d − P (π) = β #π ∈ , π . n ↓ Pn Pn k n l [n] k=1 l=0 c(n, k, l)(αβ) d − ∈ P

Proposition 4.7. For fixed n N, α > 0, β [n] [n, ) and 0 d 1, the Pern(α, β, H[1, d; n]) ∈ ∈ ∪ ∞ n ≤ ≤ Pn Pn k n l distribution on is a canonical Gibbs ensemble with vn := β and wn := c(n, k, l)α d . P[n] ↓ k=1 l=0 − Proof. By Proposition 3.2 in [10], for d [0, 1] and α R, the α-permanent of H[1, d; n] is ∈ ∈ given by (26). The correspondence between the Pern(α, β, H[1, d; n])-law and the canonical Gibbs[n](v , w ) ensemble with v and w as specified follows.  • • • • Because the Pern(α, β, H[1, d; n])-law has the canonical Gibbs form (11), it is exchangeable for every 0 d 1. In particular, the Pern(α, β, H[1, 1; n])-law corresponds to the permanen- ≤ ≤ tal partition structure in a uniformly homogeneous medium and the Pern(α, β, H[1, 0; n])-law corresponds to the permanental partition structure in a discrete medium.

Theorem 4.8. The Pern(α, β, X)-law is exchangeable if and only if X = H[1, d; n] for some 0 d 1. ≤ ≤ Proof. One direction, that the Pern(α, β, H[1, d; n])-law is exchangeable for all d [0, 1], α,β,X ∈ follows from Proposition 4.7. In the other direction, suppose Pn is exchangeable. In this case, for n 4, consider the partition π with blocks 1, 2 , 3 , 4,..., n , and let σ n be the permutation≥ that transposes elements 2 and 3. Then{ πσ} has{ } { blocks }1, 3 , 2 , 4,...,∈ S n and, under exchangeability, must have the same probability of π. In this{ case,} { we} { have } Q Q α,β,X #π b π perα X[b] #πσ b πσ perα X[b] σ Pn (π) = β↓ ∈ = β↓ ∈ = Pn(π ); perαβ X perαβ X whence, per X[ 1, 2 ] = per X[ 1, 3 ], from which we conclude X = X . Since the above α { } α { } 12 13 argument holds for 1, 2, 3 replaced by any distinct i, j, k, we must have Xij = Xi0 j0 for all i , j and i0 , j0. For n 3, this result can be shown by omitting the last block 4,..., n in the above line of reasoning.≤ This completes the proof. { } 

4.3.3. Non-overlapping uniformly homogeneous subpopulations. In this section, let X := (Xij, i, j 1) be a 0-1 valued array representing a partition of N. That is, for each n N, the leading≥ ∈ n n submatrix Xn := (Xij, 1 i, j n) corresponds to some π as in Section 2.1. × ≤ ≤ ∈ P[n] Theorem 4.9. Let X := (Xij, i, j 1) be an array representing a partition of N and let α > 0, k 1. ≥ ≥ For each n N, let Xn := (Xij, 1 i, j n) be the leading n n submatrix of X. Then the collection ∈ ≤ ≤ × (Pα,k,Xn , n N) of permanental partition distributions in (7) is consistent under subsampling. n ∈ PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 17

Proof. It is enough to establish that for m n N, π and π D 1 (π), the identity ≤ ∈ ∈ P[m] ∗ ∈ m−,n P #π #π0 1 β 00 per (π π ) β↓ per (π π0) π D− (π ) ↓ α ∗ 00 (27) α · = 00∈ m,n 0 · perαβ π perαβ π∗ holds for all π . By induction, we need only show this for m = n 1. 0 ∈ P[m] − Fix n N and let π , π D 1 (π). Then, for any π , each π D 1 (π ) ∈ ∈ P[n] ∗ ∈ n−,n+1 0 ∈ P[n] 00 ∈ n−,n+1 0 is obtained from π either by inserting element n + 1 into some b π or adding n + 1 as a 0 ∈ 0 { } singleton to π0. We write b π to denote the block of π to which n + 1 is added to obtain ∗ ∈ { } π∗. We have per π∗ = (αβ + #b ) per π and αβ ∗ αβ X #π00 β↓ perα(π∗ π00) = 1 · π00 Dn−,n+1(π0) ∈ X #π #π = β↓ 0 (#(b b0) + α) per (π π0) + (β #π0)β↓ 0 α per (π π0) ∗ ∩ α · − α · b π 0∈ 0 #π = β↓ 0 per (π π0)(#b + αβ), α · ∗ which completes the proof.  α,β,d Corollary 4.10. When d = 0, 1, the collection (Pn , n N) of Pern(α, β, H[1, d; n]) distributions is consistent. ∈

When X corresponds to a partition of N, Kolmogorov’s extension theorem and Theo- rem 4.9 imply the existence of a unique probability measure Pα,k,X on whose finite- P∞ dimensional distributions correspond to the Pern(α, k, Xn)-law for each n N. To our knowledge, in the literature on random partitions of N, precise descriptions∈ of probabil- ity distributions on N by consistent finite-dimensional marginals has been restricted to exchangeable and partially exchangeable random partition structures, see e.g. [32]. In Theorem 4.9, we encounter an explicit collection of finite-dimensional distributions that are not exchangeable, but nevertheless determine a unique probability measure on . P∞ α,β,Xn Corollary 4.11. Let (Pn , n N) be a collection of permanental partition structures such that, for α,β,X ∈ α,β,X each n N, P n is a probability distribution on . Then (P n , n N) is both exchangeable ∈ n P[n] n ∈ and consistent if and only if β N and, for either d = 0 or d = 1, Xn := (Xij, 1 i, j n), where ∈ ≤ ≤ (Xij, i, j 1) is an infinite array with entries Xij = d if i , j and Xii = 1 for all i N. ≥ ∈ Proof. Clearly, by Theorems 4.8 and 4.9, β N and X = H[1, d; n], 0 d 1, determines an exchangeable and consistent family of permanental∈ partition structures.≤ ≤ Conversely, α,β,X suppose (P n , n N) is exchangeable and consistent. Then, by Theorem 4.8, there is n ∈ some d [0, 1] such that Xn = H[1, d; n] for all n N. Furthermore, by the definition ∈ ∈ of the Pern(α, β, X)-partition structure, β must be a positive integer. Under consistency, α,β,X (P n , n N) must satisfy n ∈ X α,β,Xn α,β,Xn+1 P (π) = P (π0) for all π , n n+1 ∈ P[n] π D 1 (π) 0∈ n−,n+1 18 HARRY CRANE for each n N. In the case of the permanental partition model (23) with β = k N, consistency∈ becomes ∈ X #π perα(Xn π) (#π+1) perα(Xn π) #π perα(Xn π) perα(Xn+1[b n + 1 ]) k↓ · = k↓ α · + k↓ · ∪ { } ; per Xn per Xn+1 per Xn+1 per (Xn[b]) αk αβ αβ b π α ∈ whence, perαβ Xn+1 X per (Xn+1[b n + 1 ]) = (k #π)α + α ∪ { } . per Xn − per (Xn[b]) αβ b π α ∈ Since the left-hand side does not depend on π, we must have that the ratio per (Xn [b α +1 ∪ n + 1 ])/ per (Xn[b]) is of the form y + z#b, z 0, for every b [n], in which case { } α ≥ ⊆ per (Xn[b]) = (y + z(#b 1)) (y + z)y for all b [n]. α − ··· ⊆ Clearly, perα(Xn[[1]]) = α forces y = α. Furthermore, if z > 0, then 2 2 perα(Xn[[2]]) = α + d α implies d = 1. On the other hand, when z = 0, we have α2 + d2α = α2, which implies d = 0.  4.3.4. Similarity functions. Rather than specifying a fixed similarity matrix X, we could, instead, parameterize a permanental partition structure by a similarity function K : [0, 1], where is assumed to be some abstract space in which particles are positioned,X × X→ or suspended. MimickingX our definition of a similarity matrix, a similarity function K satisfies (i’) K(x, x) = 1 for all x , ∈ X (ii’)0 K(x, x0) 1 for all x, x0 , and (iii’) K≤(x, x ) = K≤(x , x) for all x, x∈ X . 0 0 0 ∈ X This description gives a more vivid illustration of our prototypical phenomenon. We assume there are n particles at positions x1,..., xn in space. The pairwise interactions among these particles are determined by the medium K. Of course, for fixed K and x1,..., xn, we can define a similarity matrix X := K[x1,..., xn] = (Xij, 1 i, j n) with Xij := K(xi, xj). Some benefits of this description are that it allows us to generate≤ ≤ positions randomly, consider adding points, consider motion of particles in , etc. We discuss some generalization of this in Section 6, but leave its detailed study to futureX work.

5. Markovian Gibbs structures and the permanental partition process We now discuss a Gibbs model for time-varying particle configurations, which we call Markovian Gibbs structures. We specifically consider Markov chain models in discrete-time. In general, a Markov chain on any finite space is determined by an initial distribution Υ0 and a transition probability measure p( , ), whereY p(y, y ) gives the probability of a transition · · 0 from state y to y0. In this section, we study the transition probabilities given in (8). 5.1. Self-similar cut-and-paste chains. In [8], we introduced a family of exchangeable and consistent Markov chains on the system ( [n]:k, n N) of set partitions with at most k blocks. Here we focus on a special sub-familyP of Ewens∈ (α) cut-and-paste Markov chains, with α > 0, which is reversible with respect to the Pitman-Ewens( α, kα)-laws: − Q #b α,kα #π b π α↑ P− ( ) = k ∈ n π ↓ n , π [n] . (kα)↑ ∈ P PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 19

The asymptotic behavior of these transition probabilities as both α and k vary is summarized in the following proposition, whose proof is straightforward. α,k Proposition 5.1. For α > 0 and k 1, let pn be the finite-dimensional transition probabilities (14). The following asymptotic transition≥ laws hold. For fixed α > 0 and k , (14) converges to • → ∞ Q ( Q Γ(#(b b0)) #(π π0) b0 π0 α ∧ b π ∈ #b ∩ , π0 π pn(π, π0; α) = ∈ α↑ ≤ 0, otherwise, the law of a Ewens(α)-fragmentation chain. For fixed k 1 and α , (14) converges to • ≥ → ∞ #π0 n (28) pn(π, π0; k) = Pn(π0; k) = k↓ /k . For fixed k 1 and α 0, (14) converges to • ≥ → ( k #π0 /k#π, π π p (π, π ; k) = ↓ 0 n 0 0, otherwise≤ , a discrete-time coalescent chain. For α/k 0, k such that α θ > 0, (14) converges to • → → ∞ → Q ( Q (Γ(#(b b0)) #(π π0) b0 π0 θ ∧ b π ∈ #b ∩ , π0 π pn(π, π0; θ) = ∈ θ↑ ≤ 0, otherwise, the law of a Ewens(θ)-fragmentation chain. For α 0 and k , (14) converges to the unit mass: • → → ∞ pn(π, π0) = δπ(π0). Remark 5.2. Note that the asymptotic transition law as α in (28) is independent of π. The weak limit of a sequence of partitions under α is that→ ∞ of an independent and identically distributed sequence of partitions governed by the→ coupon-collector ∞ distribution on . P[n]:k Remark 5.3. We call α the diversity parameter, as its magnitude relates to the dependence in the model: larger values of α correspond to more erratic behavior (the sequence is asymptotically i.i.d.), while small values of α lead to more controlled behavior, which tends toward local one-step transitions (pure coalescence). The transition probabilities in (14) resemble the form of a Gibbs measure (1) and motivates our consideration of Markovian Gibbs structures in the next section. 5.2. Markovian Gibbs structures. For n N, let H : 2[n] R be an energy function for the Markovian Gibbs transition probability ∈ →    X  a(π0)   (29) pn(π, π0):= exp  H(b) , π, π0 [n] . Za,H(π)   ∈ P b π π  ∈ ∧ 0 #π #b By putting a(π0):= k↓ 0 and H(b):= log(α/k)↑ , (29) reduces to the Ewens(α) cut-and-paste transition probability (14). We define a canonical Markovian Gibbs structure as any transition probability of the form Y v#π0 (30) pn(π, π0; v , w ) = w#b, π, π0 [n], • • Bn(v , w , π) ∈ P • • b π π ∈ ∧ 0 20 HARRY CRANE for sequences of non-negative weights v := (vn, n N) and w := (wn, n N). A canonical Markovian Gibbs structure may also be• called mean-field∈ or exchangeable• , as∈ (30) satisfies the exchangeability criterion σ σ pn(π, π0) = pn(π , π0 ) for all σ n, for all π, π0 . ∈ S ∈ P[n] Note that in the special case (14), the functions a( ) and H( ) depend on π0 and b, respectively, only through their cardinalities; hence, (14) is a· canonical· Markovian Gibbs structure with v := k↓• and w := (α/k)↑•. • • α,k Corollary 5.4. For any α > 0 and k 1, the collection (pn , n N) of Ewens(α) cut-and- paste transition probabilities is a consistent≥ collection of canonical Markovian∈ Gibbs structures on ( , n N). P[n]:k ∈ Proof. This proof is analogous to that of Corollary 4.11.  Analogous to the characterization of canonical Gibbs ensembles by Pitman-Ewens(α, θ) (Theorem 4.3), the Ewens(α) cut-and-paste chain is the unique consistent canonical Mar- kovian Gibbs structure.

Theorem 5.5. For non-negative sequences v and w , let (pn( , ; v , w ), n N) be a collection of • • · · • • ∈ canonical Markovian Gibbs structures as in (30). Then (pn( , ; v , w ), n N) is consistent if and · · • • ∈ only if v := k↓• and w := α↑•, for some k N and α > 0. • • ∈ Proof. For convenience, we write (29) in the equivalent form

a(#π0) Y Y pn(π, π0) = w(#(b b0)), Z(π) ∩ b π b π ∈ 0∈ 0 1 Q for w(0) = 1, where w(n):= wn. Let π, π0 [n], π∗ Dn− n+1(π) and write w(π) = b π w#b. ∈ P ∈ , ∈ The Gibbs form (29) and consistency (17) imply

a(#π0) a(#π0 + 1) X a(#π0) w(π π0) = w(π π0)w(1) + w(π∗ π00), Z(π) ∧ Z(π ) ∧ Z(π ) ∧ b ∗ b π ∗ ∈ 0 where π is obtained by inserting element n + 1 in block b π ; whence, for any π in the b00 ∈ 0 0 support of pn(π, ), · Z(π∗) X C(π∗):= = A(#π0)w(1) + W(#(b0 b∗)), Z(π) ∩ b π 0∈ 0 where W(i):= w(i + 1)/w(i), A(r):= a(r + 1)/a(r) and b π is the block into which n + 1 ∗ ∈ is inserted to obtain π∗. The left side depends only on π∗, which implies W(i + 1) W(i) is constant for all i 0, and we may write W(i + 1) W(i) = β. For the , assume− W(i) = α + βi. We have≥ − Yi 1 − i i w(i) = W(i 1)W(i 2) W(0) = (α + βj) = β (α/β)↑ . − − ··· j=0 i There is no loss of generality in putting β = 1 so that w(i) = α↑ , i 0, giving X ≥ C(π∗) = A(r)w(1) + W(#(b∗ b0)) = αA(r) + αr + #b∗, r inf m N : a(m + 1) = 0 ; ∩ ≤ { ∈ } b π 0∈ 0 PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 21 whence, A(r):= δ r for some δ R. Without loss of generality, we assign a(1) = δ so that − ∈ Yr 1 Yr 1 − − r a(r) = a(1) A(j) = δ (δ j) = δ↓ . − j=1 j=1

Since the Markovian Gibbs form (29) depends on π0 and π π0, but not π, both α and δ must be strictly positive. As a result, A(r) will be negative for∧ large enough r; and so we must have #π0 < k for some integer k < and a(k + 1) = 0 implies δ = k. In this case, the conditional Gibbs splitting∞ rule is

#π0 Y Y k↓ #(b b ) p (π, π0) = α↑ ∩ 0 . n Z(π) b π b π ∈ 0∈ 0

By (19), the normalizing constant Z(π) = perαk π. The following asymptotic cases arise. k < and 0 < γ < : In this case, (29) is exactly the Ewens(γk) cut-and-paste • transition∞ probability∞ (14). k < and γ = 0: γ = 0 occurs if either β = or α = 0 and, in this case, (29) is zero • unless∞ π π = π and the splitting rule has∞ the form ∧ 0 #π k↓ 0 pn(π, π0) = I π π=π , k#π { 0∧ } a discrete-time coalescent chain. i k < and γ = : γ = corresponds to β = 0, for which w(i) = α and C(π∗) = • αA(r∞) + αr ; hence,∞ A(r ) +∞r = δ > 0 and A(r ) = δ r as above. In this case, 0 0 0 0 − k #π0 p (π, π ) = ↓ . n 0 kn k = , γ = 0 and 0 < γk = θ < : We see that (29) becomes • ∞ ∞ Q Y Γ(#(b b )) #(π π0) b0 π0 0 pn(π, π0; θ) = θ ∧ ∈ ∩ I π π =π . θ #b { ∧ 0 0} b π ↑ ∈

k = and 0 < γ : This is the deterministic split into singletons, pn(π, π ) = • ∞ ≤ ∞ 0 I π =0n . Indeed, when 0 < γ the limit of pn(π, π0) in (29) as k is the { 0 } ≤ ∞ → ∞ at #π = n = π = 0n . This is the limit of (14) as k . { 0 } { 0 } → ∞ k = , γ = 0 and γk = 0; or k = , γ = 0 and γk = : In this case, pn(π, π0) = I π=π0 • is the∞ degenerate distribution at∞π. ∞ { } 

Remark 5.6. In comparision to Theorem 4.3, the characterization in Theorem 5.5 only permits the case corresponding to α < 0 and θ = kα for k N; the canonical Markovian Gibbs ensemble − ∈ cannot be supported on all of [n], for all n N. The difference between the cases arises because, when inserting a new element Pn + 1 into partitions∈ (π, π ), a new block can form in π π without 0 ∧ 0 forming in either π or π0. 22 HARRY CRANE

5.3. Permanental Markovian Gibbs structures. For a certain choice of α, β R and X n n ∈ ∈ R × , we defined the permanental transition probability on [n] in (8). The transition X P α,β, #π0 probability pn in (8) has the form of a Markovian Gibbs structure (29) with a(π0):= β↓ , H(b):= log per X[b] and Za H(π):= per (X π). An important special case of (8) arises α , αβ · when α > 0, β = k N and X = 1n, the n n matrix of all ones, in which case (8) coincides with the Ewens(α)∈ Markovian Gibbs structure× (14).

Proposition 5.7. Let α > 0, β [n] [n, ) and X be a similarity matrix, then, for every π [n], ∈ ∪ ∞ α,β,X ∈ P pn(π, ) in (8) is a probability distribution on [n]. It follows that pn in (8) is a transition probability· on . P P[n] Proof. This follows from Proposition 4.4 since, for any similarity matrix X, the Hadamard product X π is also a similarity matrix, for every π .  · ∈ P[n] Definition 5.8 (Permanental Markovian Gibbs structures). For n N, α > 0, β [n] [n, ) α,β,X ∈ ∈ ∪ ∞ and any similarity matrix X, we call pn in (8) the Pern(α, β, X)-Markovian Gibbs structure with α,β,X diversity α, mixing coefficient β and similarity matrix X. Alternatively, we may refer to pn as the permanental Markovian Gibbs structure, or permanental transition probability, with parameter (α, β, X).

Theorem 5.9. The Pern(α, β, X)-transition measure is reversible with respect to the Pern(αβ, β, X)- distribution.

Proof. For α > 0 and β [n] [n, ), αβ > 0, and so, by Proposition 4.4, (αβ, β, X) is in the ∈ ∪ ∞ α,β,X parameter space of the permanental partition distribution. It is immediately clear that pn αβ,β,X in (8) satisfies the detailed-balance condition (18) with respect to Pn in (7).  Theorem 5.10. The permanental transition probability in (8) is a canonical Markovian Gibbs α,k,X structure (30) if and only if X is homogeneously symmetric. In this case, pn satisfies (30) for P P k l v := k↓• and w := k• l• c( , k, l)α d•− , for some d [0, 1]. • • =1 =0 • ∈ Proof. This proof is analogous to that of Theorem 4.8. 

5.3.1. The case β N. We now specialize to the case β N which, by Definition 5.8, is in the ∈ ∈ parameter space of the Pern(α, β, X)-transition probability for every n N. For k 1, the ∈ ≥ Pern(α, k, X)-transition probability in (8) is a transition probability on the subspace . P[n]:k Corollary 5.11. For every n N, the Ewens(α) cut-and-paste transition probability in (14) ∈ coincides with the Pern(α/k, k, 1n)-transition probability in (8).

Another special case arises when X = In, the n n identity matrix. In this case, (8) simplifies to ×

α,k,In #π n p (π, π0):= k↓ 0 /k , π, π0 , n ∈ P[n]:k which is independent of π and α. Hence, when X = In, the permanental transition probabilities govern an i.i.d. sequence of coupon collector partitions on k coupons. We also point out that, in the cases X = 1n and In, the Pern(α, k, X)-transition probability is consistent. We discuss consistent collections of transition probabilities in the next section. PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 23

5.4. Consistent Markovian Gibbs structures. Recall, from Section 2.4, that a family of transition probabilities (pn, n N) on ( , n N) is consistent if and only if, for every ∈ P[n] ∈ π, π0 [m], ∈ P 1 1 pm(π, π0) = pn(π∗, D− (π0)) for all π∗ D− (π). m,n ∈ m,n Theorem 5.12. Let X be an infinite array corresponding to a partition of N, and, for each n N, α,k,Xn ∈ let Xn := (Xij, 1 i, j n) be its leading n n submatrix. If, for each n N, pn denotes ≤ ≤ × α,k,Xn ∈ the Pern(α, k, Xn)-transition probability on [n]:k, then (pn , n N) is a consistent collection of Markovian Gibbs structures. P ∈ Proof. This proof is analogous to that of Theorem 4.9. 

By Theorem 5.12, we can construct a Markov chain on [ ]:k, partitions of N with at most k blocks, that is not exchangeable. By combining TheoremP ∞ 5.10 and 5.12, we arrive at a characterization of the class of exchangeable and consistent permanental transition probabilities.

α,β,Xn Corollary 5.13. Let (pn , n N) be a family of permanental Markovian Gibbs structures such α,β,X ∈ α,β,X that, for each n N, p n is a transition probability on . Then (p n , n N) is exchangeable ∈ n P[n] n ∈ and consistent if and only if β N and, for either d = 0 or d = 1, Xn := (Xij, 1 i, j n), where ∈ ≤ ≤ (Xij, i, j 1) is an infinite array with entries Xij = d if i , j and Xii = 1 for all i N. ≥ ∈ 6. Equivariance, exchangeability and lack of interference In this section, we discuss both the Permanental partition structures and Permanental Markovian Gibbs structures simultaneously by referring generically to the permanental model.

6.1. Equivariance. In a homogeneous enviroment, the permanental model is exchangeable; however, in any other environment, it is not. In other situations, the lack of homogeneity implies that the positions of the particles affect pairwise interactions and, hence, the law governing the configuration. In a non-exchangeable model in which particles have fixed positions in space (or, more generically, fixed covariates), the natural model invariance is with respect to the action of the symmetric group on the similarity matrix parameter, which reflects a relabeling of the positions x1,..., xn, but not any substantive change to the configuration of the particles in space (as we are assuming particles are physically identical). Such a model is called equivariant with respect to the symmetric group. Note that exchangeability is also defined as an invariance with respect to the symmetric group; however, exchangeability is a property of a distribution, whereas equivariance is a property of the family. In particular, a probability distribution Pn on [n] is exchangeable if, for each π σ P X ∈ [n], Pn(π) = Pn(π ), for all σ n. On the other hand, a family (Pn ) indexed by a similarity matrixP X is equivariant with∈ respect S to the symmetric group if, for every permutation Xσ X σ σ :[n] [n], (Pn ) determines the same family as (Pn ), where X := (Xσ(i)σ(j), 1 i, j n). → X Xσ σ ≤ ≤ Equivariance holds, for instance, if, for each π [n], Pn (π) = Pn (π ) for all σ n. By a symmetry property of the α-permanent, the permanental∈ P model is equivariant with∈ S respect to the symmetric group acting on [n], as we now show.

Proposition 6.1. For every n N, both the Pern(α, β, X)-partition structure and Markovian Gibbs structure are equivariant with∈ respect to the symmetric group. 24 HARRY CRANE

σ σ σ σ Proof. For any matrices X, Y and σ n, perα X = perα(X ) and (X ) (Y ) = (X Y) . Hence, for any π and (α, β, X) in the∈ parameter S space of the permanental· partition· model, ∈ P[n] σ σ per (X π) σ per (X π ) σ α,β,X #π α · #π α · α,β,X σ Pn (π) = β↓ = β σ = Pn (π ), perαβ X perαβ X for all σ n. The same calculation shows that the Pern(α, β, X)-family of transition measures∈ is Sequivariant.  X Equivariance is important as it ensures that the law of Π Pn depends on X only through its fundamental structure (i.e. the pairwise interactions/similarities∼ among particles) and not any superfluous variates, e.g. the labels 1,..., n assigned the particles.

6.2. Generating exchangeable Gibbs structures by mixing. Although the Pern(α, β, X) model is not exchangeable in general, we can introduce X as an exchangeable random environment governed by a distribution Σ on the space of similarity matrices. In this case, the unconditional Pern(α, β; Σ)-model (defined below) gives a Bayesian model for particle configurations. In particular, assume Σ is a weakly exchangeable probability measure, σ i.e. X Σ satisfies X = X for every σ n. In this case, we define the unconditional permanental∼ models by L ∈ S Z α,β;Σ α,β,X Pn (π):= Pn (π)Σ(dX) and Z α,β;Σ α,β,X pn (π, π0):= pn (π, π0)Σ(dX).

Proposition 6.2. Let Σ be a weakly exchangeable probability measure on n n similarity matrices. α,β;Σ × Then the unconditional Pern(α, β; Σ)-partition structure Pn and Pern(α, β; Σ)-Markovian Gibbs α,β;Σ structure pn are exchangeable. Proof. This is a straightforward consequence of weak exchangeability of Σ and Proposition 6.1.  Theorem 6.3. Let Σ be an exchangeable probability measure on . Then the permanental model with parameter (α, β; Σ) is both exchangeable and consistent and,P∞ therefore, determines a unique process on . P∞ Proof. This follows from Theorems 4.9 and 5.12 and Proposition 6.2.  The results in this section hint at probability models for particle configurations in a random enviroment. This relates to some previous literature on random walks in random environments (RWRE), see e.g. [20]. Based on the developments in Sections 4 and 5, we have a description of a RWRE with explicit marginal distributions, which is a topic for future study. 6.3. Consistency under subsampling and lack of interference. We have shown (Theorems 4.9 and 5.12) that, under certain choices of the similarity matrix X, the permanental model is consistent under subsampling. In the context of physical models, consistency under subsampling reflects an inherent lack of interference among the particles in the system. That is, given a configuration x1,..., xn, xn+1 of n + 1 particles in space, the law governing the clustering of the particles at positions x1,..., xn, ignoring the particle at position xn+1, is the PERMANENTAL PARTITION MODELS AND MARKOVIAN GIBBS STRUCTURES 25 same as the law of a system having only n particles at positions x1,..., xn. In other words, an additional particle, or particles, does not interfere with the law governing the configuration of the rest of the system. In general, the permanental partition models do not exhibit lack of interference, just as many physical systems do not possess such a property. In many physical systems, addition, or removal, of a particle perturbs how the rest of the system interacts in complicated ways; therefore, although a lack of consistency is detrimental from a statistical inference perspective, see e.g. [27], lack of interference exhibits a flexibility, rather than a defect, of the permanental partition model in modeling physical systems. 6.4. Permanental Gibbs structures on k-colorings. For fixed k 1, the permanental distri- ≥ bution and transition measures can be defined on the space [k][n] of [k]-valued sequences of length n, called k-colorings of [n]. In addition to (19), the α-permanent satisfies the following more general identity with respect to k-colorings (Lemma 3 in [29]). Let α , . . . , α R and 1 k ∈ write α := α1 + + αk. Then, for any n n matrix X, • ··· X × (31) per X = per (X[w ]) per (X[w ]), α α1 1 αk k • ··· (w1,...,wk) where the sum is over ordered collections (w1,..., wk) of disjoint subsets of [n] such that Sk i=1 wi = [n]. Note that, for k = β and α1 = = αk = α, (31) reduces to (19). In this way, (31) is a refinement of (19) in the special case where··· β is a non-negative integer. The identity (31) suggests a probability distribution on k-colorings of [n] by Qk j=1 per X[xj] α,X αj [n] (32) Pn (x):= , x [k] , perα X ∈ • i where α := (α1, . . . , αk) and xj := i [n]: x = j is the set of indices with color j in x. When { ∈ } [n] X = 1n, (32) corresponds to the Dirichlet- on [k] with parameter (α1, . . . , αk). For a k k matrix α := (αij, 1 i, j k), we define the permanental Markovian Gibbs × ≤ ≤ structure on [k][n] by

k Qk Y j=1 per X[xi x0] α,X αij ∩ j [n] (33) p (x, x0):= , x, x0 [k] , n per X[x ] ∈ αi i i=1 • Pk α0,X where αi := j=1 αij for each i = 1,..., k. Note that (33) is reversible with respect to Pn • in (32) with α0 := (α1 , . . . , αk ). Using the same arguments, one can obtain analogous outcomes for exchangeable,• consistent• and reversible permanental Gibbs structures and Markovian Gibbs structures on [k][n].

References [1] D. J. Aldous. Deterministic and stochastic models for coalescence (aggregation and coagulation): a review of the mean-field theory for probabilists. Bernoulli, 5(1):3–48, 1999. [2] R. Arratia, A. Barbour, and S. Tavare.´ Logarithmic Combinatorial Structures: a Probabilistic Approach. European Mathematical Society, 2003. [3] J. Berestycki. Exchangeable fragmentation-coalescence processes and their equilibrium measures. Electron. J. Probab., 9:no. 25, 770–824 (electronic), 2004. [4] N. Berestycki and J. Pitman. Gibbs distributions for random partitions generated by a fragmentation process. J. Stat. Phys., 127(2):381–418, 2007. [5] J. Bertoin. Homogeneous fragmentation processes. Probab. Theory Related Fields, 121(3):301–318, 2001. 26 HARRY CRANE

[6] J. Bertoin. Random fragmentation and coagulation processes, volume 102 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, Cambridge, 2006. [7] C. J. Burke and M. Rosenblatt. A Markovian function of a Markov chain. Ann. Math. Statist., 29:1112–1122, 1958. [8] H. Crane. A consistent Markov partition process generated from the paintbox process. J. Appl. Probab., 43(3):778–791, 2011. [9] H. Crane. Homogeneous cut-and-paste processes. Manuscript., 2013. [10] H. Crane. Some algebraic identities for the α-permanent. To appear in Linear Algebra and Its Applications, 2013. [11] P. Diaconis and S. N. Evans. Immanants and finite point processes. J. Combin. Theory Ser. A, 91(1-2):305–321, 2000. In memory of Gian-Carlo Rota. [12] P. Donnelly and G. Grimmett. On the asymptotic distribution of large prime factors. J. London Math. Soc, 47:395–404, 1993. [13] B. Efron and R. Thisted. Estimating the number of unseen species: How many words did Shakespeare know? Biometrika, 63:435–447, 1976. [14] W. J. Ewens. The sampling theory of selectively neutral alleles. Theoret. Population Biology, 3:87–112, 1972. [15] R. Fisher, A. Corbet, and C. Williams. The relation between the number of species and the number of individuals in a random sample of an animal population. J. Animal Ecology, 12:42–58, 1943. [16] W. Fulton and J. Harris. Representation Theory: A First Course. Graduate Texts in Mathematics/Readings in Mathematics. Springer, 1991. [17] B. Gyires. Discrete distribution and permanents. Publ. Math. Debrecen, 20:93–106, 1973. [18] J. A. Hartigan. Partition models. Comm. Statist. Theory Methods, 19(8):2745–2756, 1990. [19] J. B. Hough, M. Krishnapur, Y. Peres, and B. Virag.´ Determinantal Processes and Independence. Probability Surveys, 3:206–229, 2006. [20] B. D. Hughes. Random walks and random enviroments. Oxford University Press, 1996. [21] M. Jerrum, A. Sinclair, and E. Vigoda. A polynomial-time approximation algorithm for the permanent of a matrix with non-negative entries. Journal of the ACM, pages 671–697, 2004. [22] F. Kelly. Reversibility in Stochastic Networks. Wiley, 1979. [23] J. F. C. Kingman. The representation of partition structures. J. London Math. Soc. (2), 18(2):374–380, 1978. [24] J. F. C. Kingman. Mathematics of genetic diversity, volume 34 of CBMS-NSF Regional Conference Series in Applied Mathematics. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, Pa., 1980. [25] J. F. C. Kingman. The coalescent. Stochastic Process. Appl., 13(3):235–248, 1982. [26] S. C. Kou and P. McCullagh. Approximating the α-permanent. Biometrika, 96(3):635–644, 2009. [27] P. McCullagh. What is a statistical model? Ann. Statist., 30(5):1225–1310, 2002. With comments and a rejoinder by the author. [28] P. McCullagh. An asymptotic approximation for the permanent of a doubly stochastic matrix. Journal of Statistical Computation and Simulation, pages 1–11, 2012. [29] P. McCullagh and J. Møller. The permanental process. Adv. in Appl. Probab., 38(4):873–888, 2006. [30] P. McCullagh and J. Yang. How many clusters? Bayesian Anal., 3(1):101–120, 2008. [31] H. Minc. Permanents. Addison-Wesley, Reading, MA, 1978. [32] J. Pitman. Exchangeable and partially exchangeable random partitions. Probab. Theory Related Fields, 102(2):145–158, 1995. [33] J. Pitman. Combinatorial stochastic processes, volume 1875 of Lecture Notes in Mathematics. Springer-Verlag, Berlin, 2006. Lectures from the 32nd Summer School on Probability Theory held in Saint-Flour, July 7–24, 2002, With a foreword by Jean Picard. [34] A. Rinaldo and C. R. Shalizi. Consistency under subsampling of exponential random graph models. Annals of Statistics, 41:508–535, 2013. [35] A. Soshnikov. Determinantal random point fields. Russian Math. Surveys, 55(5):923–975, 2000. [36] L. G. Valiant. The complexity of computing the permanent. Theoretical Computer Science, 8:189–201, 1979. [37] D. Vere-Jones. Alpha-permanents and their applications to mulivariate gamma, negative binomial and ordinary binomial distributions. New Zealand J. Math., 26:125–149, 1997. [38] D. Vere-Jones. A generalization of permanents and determinants. Linear Alg. Appl., 111:119–124, 1998.

Rutgers University,Department of Statistics, 110 Frelinghuysen Road,Piscataway, NJ 08854.