Finite Operator Calculus With Applications to Linear Recursions
Heinrich Niederhausen Florida Atlantic University Boca Raton [email protected] www.math.fau.edu/Niederhausen 2 Contents
Contents i
1 Prerequisites from the Theory of Formal Power Series 1 1.1 Generating Functions and Linear Recursions ...... 2 1.1.1 Roots ...... 9 1.1.2 Exercises ...... 10 1.2 Composition and Inverses ...... 13 1.2.1 Exercises ...... 15 1.3 Multivariate Power Series ...... 17 1.3.1 Exercises ...... 20
2 Finite Operator Calculus in One Variable 21 2.1 Polynomials, Operators, and Functionals ...... 21 2.1.1 The Vector Space of Polynomials, and Their Bases . . . . . 22 2.1.2 Standard Bases and Linear Operators ...... 25 2.1.3 Exercises ...... 26 2.2 Finite Operators ...... 27 2.2.1 Translation Operators ...... 28 2.2.2 Basic Sequences and Delta Operators ...... 30 2.2.3 Special Cases ...... 32 2.2.4 Exercises ...... 40 2.3 Sheffer Sequences ...... 44 2.3.1 Initial Values Along a Line ...... 47 2.3.2 The Umbral Group ...... 51 2.3.3 Special Cases ...... 54 2.3.4 Exercises ...... 63 2.4 Transfer Theorems ...... 69 2.4.1 Umbral shifts and the Pincherle derivative ...... 73 2.4.2 Proof of the Transfer Formula ...... 74 2.4.3 Exercises ...... 76 ii CONTENTS
3 Applications 79 3.1 The Functional Expansion Theorem ...... 79 3.1.1 Some Applications of the Functional Expansion Theorem . 83 3.1.2 Exercises ...... 87 3.2 Diagonals of Riordan Matrices as Values of Sheffer Sequences . . . 89 3.2.1 Exercises ...... 93 3.3 Determinants of Hankel Matrices ...... 94 3.3.1 Exercises ...... 100 3.4 Classical Umbral Calculus ...... 102 3.4.1 The Cumulant Umbra ...... 109 3.4.2 Exercises ...... 111
4 Finite Operator Calculus in Several Variables 113 4.1 Polynomials and Operators in Several Variables ...... 114 4.1.1 Exercises ...... 117 4.2 The Multivariate Transfer Formulas ...... 119 4.2.1 Transfer with constant coeffi cients ...... 119 4.2.2 Operator based transfer ...... 121 4.2.3 The multivariate Pincherle derivative ...... 123 4.2.4 Transfer with operator coeffi cients ...... 128 4.2.5 Exercises ...... 131 4.3 The Multivariate Functional Expansion Theorem ...... 134 4.3.1 Exercises ...... 136
5 Special Constructions in Several Variables 137 5.1 Multi-indexed Sheffer Sequences ...... 137 5.1.1 Delta Operators for multi-indexed Sheffer sequences. . . . . 139 5.1.2 Translation invariance of diagonalization, and some examples.140 5.1.3 Abelization of Multi-Indexed Sequences ...... 142 5.1.4 Exercises ...... 147 5.2 Polynomials with all but one variable equal to 0 ...... 150 5.2.1 Exercises ...... 153 5.3 Cross-Sequences and Steffensen Sequences ...... 153 5.3.1 Exercises ...... 155
6 A General Finite Operator Calculus 157 6.1 Transforms of Operators ...... 157 6.2 Reference Frames, Sheffer Sequences, and Delta Operators . . . . . 162 6.2.1 Reference Frames ...... 162 6.2.2 Sheffer Sequences and Delta Operators ...... 165 6.2.3 Exercises ...... 169 6.3 Transfer Formulas ...... 171 6.3.1 General Umbral Shifts and the Pincherle Derivative . . . . 171 6.3.2 Equivalent Transfer Formulas ...... 173 CONTENTS iii
6.3.3 Exercises ...... 175 6.4 Functionals ...... 176 6.4.1 Augmentation ...... 179 6.4.2 Orthogonality ...... 179 6.4.3 Exercises ...... 184
7 Applications of the General Theory 187 7.1 The Binomial Reference Frame ...... 188 7.1.1 Orthogonal Binomial Reference Sequences ...... 189 7.1.2 Generalized Catalan Operators ...... 195 7.1.3 Dickson Polynomials ...... 197 7.1.4 Exercises ...... 200 7.2 Eulerian Differential Operators ...... 204 7.2.1 Exercises ...... 209
8 Solutions to Exercises 211
Bibliography 257 iv CONTENTS
Preface
The following text originated from various lecture notes for graduate courses in combinatorics. I would very much appreciate receiving your comments, additions and corrections. My apologies to all those mathematicians not mentioned in the text, their important contributions to the theory of the Finite Operator Calculus skipped over because of ignorance or by design in order to keep the material manageable. Your views can still be included - please let me know: [email protected].
Introduction
n Every linear operator T on polynomials has a representation T = n 0 M(pn) , ≥ D where is the derivative operator, and M(pn) stands for multiplication by the D P polynomial pn (x) (Pincherle [74, 1901]). When applied to polynomials, the oper- ator T reduces to a finite sum, of course, and may therefore be called a “finite” n operator. A special case of this concept, when T can be written as T = n 0 cn , ≥ D where c0, c1,... are scalars, is called a called a finite operator in the “Finite Op- erator Calculus” [83, 1973] by G.-C. Rota and his students D. KahanerP and A. n Odlyzko. Hence T is isomorphic to the formal power series n 0 cnt , and it is of course exactly this isomorphism that made the Finite Operator≥ Calculus so widely applicable. We adopt Rota’s approach in this book, but consider,P in the last two n chapters, also linear operators of the form T = n 0 cn , where can be any linear operator reducing the degree by 1, deg ( q) =≥ deg qR 1 for all polynomialsR q of degree larger than 0, and q = 0 for polynomialsR P of degree− 0 (we identify polyno- mials of degree 0 with scalarsR c = 0; however, we let deg (0) = , as usual). The fundamental difference between6 Rota’s approach and the generalized−∞ version we n present, due to J. M. Freeman [35, 1985], also his student, is that T = n 0 cn is translation invariant, TEc = EcT , where Ec : f (x) f (x + c) is the operator≥ D translating by c. The generalized version does not have7→ this property.P Translation invariance, however, is an important feature in many applications. The applications we have in mind are usually the solutions to recursive equa- tions. Suppose your analysis of a given enumerative problem resulted in a recursive expression for the numbers you are looking for; to be specific, let us assume you arrived at
Fm = Fm 1 + Fm 2, − − the recursion for the Fibonacci numbers, starting at F0 = F1 = 1. A computer will give us “special”answers in a very short time, even F10,000 makes no problem whatsoever. From this point of view, you will not need this book. However, finding n+1 n+1 n+1 out that Fn = 1 + √5 1 √5 / 2 √5 is as surprising as it is − − rewarding. In addition, it is even “practical”; a scientific calculator will show that 2089 F10,000 5.4 10 . It also tells you something about the ratio, Fn/Fn 1, and ≈ × − CONTENTS v its famous limit, the Golden Ratio. Generating functions are the standard tool for solving this type of linear recursion. We give a brief introduction in the first chapter. The reader familiar with formal power series may just want to browse through section 1.2 for the notation. Now suppose the recursion you found was
Fn (m) = Fn (m 1) + Fn 1 (m 2) , − − − with initial conditions Fn (n) = Fn 1 (n) for n 1, and F0 (m) = 1 for all m. The recursion is still easy, but the− initial values≥ are also “recursive” - we have to know Fn 1 (n) before we can say what the value of Fn (n) is. We will see − that F0 (x) ,F1(x),... is a sequence of polynomials, actually a basis, and that the operator T : Fn (x) Fn 1 (x) is a translation invariant operator, satisfying the operator equation 7→ − 1 2 I = E− + E− T. We will show in chapters 2 and 3.1 how to find the solution with given initial values from such an equation. Suppose you found the system of recursion
sm,n (u, v) = sm,n (u 1, v) + sm 1,n (u, v + 2) and − − sm,n (u, v) = sm,n (u, v 1) + sm,n 1 (u + 1, v + 1) sm,n 2 (u, v) , − − − − with initial values sm,n (0, 0) = 0 for all m, n 0, except s0,0 (0, 0) = 1. This system of recursions is two dimensional and linear,≥ the initial values are explicit. We show how to write sn,m (u, v) as a sum of binomial coeffi cients in chapters 3.1 and 5, dedicated to the multivariate Finite Operator Calculus. Technically the problems get more diffi cult to solve, of course, when it comes to higher dimensions. The theory, however, remains quite easy. Somewhere between the univariate and the multivariate case fall the Steffensen sequences (section 5.3) and the multi- indexed sequences (section 5.1). Rota’s goal was to create a solid foundation for the “Umbral Calculus”, to purge it of the “witchcraft”, as he called it. This was one of his favorite themes, and he wrote more papers on Umbral Calculus later. Several young (at that time) mathematicians took up his work, and showed that it had applications to a va- riety of mathematical topics, including approximation theory, signal processing, probability theory, and, of course, combinatorics. After all, the “Finite Operator Calculus”was published a part VIII in a series of papers “On the Foundations of Combinatorial Theory”. A complete survey of papers relating to Umbral Calcu- lus until the year 2000 has been compiled by Di Bucchianico and Loeb [26]. An application of the Finite Operator Calculus can also be found in Taylor [96, 1998]. We assume that after 5 chapters the reader will get interested in the theory itself. J. M. Freeman explored this generalization in some depth [35], and we follow it literally in the last two chapters. Finally, the expert may wonder why Finite Operator Calculus and not Um- bral Calculus? Umbral Calculus is a highly “symbolic” language; no operators, vi CONTENTS just umbrae! As a rule of thumb, every Finite Operator statement gets shorter in Umbral Calculus; for example, the contents of section 2.2.3, Basic sequences with polynomial coeffi cients, is reduced to α˜D χ.α in umbral notation. However, it may be exactly this brevity, achieved through≡ a multitude of special definitions, that prevents Umbral Calculus from being widely known. We give an introduction to Umbral Calculus in section 3.4. Most of the examples and exercises in this book refer to combinatorial prob- lems, with few exceptions. Yet this is not a book on enumerative combinatorics, because we begin where combinatorics ends, that is at the crucial point, where combinatorics has delivered a recursion and initial conditions. Without question, this is the hard part of combinatorics; solving the recursion is the technical part. Finite Operator Calculus can help with the technical part, when applicable. We will therefore describe the combinatorics only briefly, just enough to introduce the recursion. There is a special sequence of examples taken from [84] about the enumeration of lattice paths containing patterns; they are Example 1.2.3, 2.3.15, 2.4.5, 5.1.6, and 5.1.8. The reader of this text should have access to a computer algebra package (CAS) like Mathematica, Maple, MuPad, etc. This will allow for checking coeffi - cients of formal power series, getting conjectures on new results, and verifying the polynomial formulas in the examples and exercises. Chapter 1
Prerequisites from the Theory of Formal Power Series
The isomorphism between Finite Operator Calculus and formal power series allows us to express “everything”we can do with operators also in terms of power series. Why then do we prefer one over the other? May be the criterion should be the ease of use; how diffi cult it is to formulate, classify, and solve a given recursion in one way or the other. However, ease of use heavily depends on the availability of a commonly known (mathematical) language. For example, linear recursions are easily translated into functional relations between formal power series, or into equations between operators. We will see in section 2.4 how to solve such an equation directly for the polynomials involved in the recursion, without knowing the operators explicitly.
We give a brief introduction to the powerful method of solving recursions by generating functions. A more detailed discussion can be found in many text books; an excellent resource is the “generatingfunctionolgy” by H. Wilf [102]. If the generating function is rational, a quotient of polynomials, an explicit form for the coeffi cients can be obtained from the roots of the denominator, at least in principle. We show an example in subsection 1.1.1.
Most important for the following chapters is the Lagrange-Bürmann inver- sion formula. A proof in algebraic form can be found in Henrici’s Applied and Computational Complex Analysis, Vol. 1 [41]. See also Hofbauer [43]. Applying Lagrange inversion in several variables gets more tedious; for pedagogical reasons we separate the multivariate from the univariate case. It is, however, not the La- grange inversion that is the principle obstacle in solving linear recursions in several variables. See section 1.3 for more details. 2 Chapter 1. Prerequisites from the Theory of Formal Power Series
1.1 Generating Functions and Linear Recursions
Suppose you have computed the number Cn of certain structures on a set of f (n) elements for every n = 0, 1,... , and you want to “store”the results. For example, you counted the number Cn of sequences (c0, c1, c2, . . . , c2n) of length f (n) = 2n i consisting of +1 and 1 in equal numbers such that the partial sum k=0 ck are never below 0, − i 2n P ck 0 and ck = 0 ≥ k=0 k=0 X X for all i = 1 to 2n. The first few Cn’sare C1 = # (1, 1) = 1, { − } C2 = # (1, 1, 1, 1) , (1, 1, 1, 1) = 2, { − − − − } C3 = 5, etc. Let us assume that C0 = 1 (assuming that C0 = 0 would also make sense; see Example 1.2.1). The generating function of the sequence (Cn)n 0 is defined as the 0 1 2 ≥ “series”γ (t) = C0t + C1t + C2t + ... , a formal sum using a formal variable t. The generating function (or formal power series, which we use as synonyms) γ (t) is in essence the sequence C0,C1,... ; convergence of the series is not assumed. Some arithmetic with generating functions is obvious, like addition and scalar multiplication. Actually, the formal power series are a vector space over the integers Z, or the rational Q, or some other ring. If we call this ring k, the power series will be denoted by k [[t]]. So the ring k contains all the coeffi cients like C0,C1,... . It will turn out that a ring structure on the coeffi cients is not enough. We will assume in this chapter that k is an integral domain, i.e., we assume that for two coeffi cients a and b their product ab cannot be zero, if a and b are both different from 0. In the following chapters, we even will assume that k is a field, and we write F [[t]] for those power series. However, in section 2.4 on transfer theorems, we will need again power series that have an integral domain (and not a field) as their coeffi cient ring. By convention, there is only one “value” of t that can be substituted for t, and that is t = 0, giving γ (0) = C0. Again, this is nothing but a notational trick, but very helpful, as we will see below. Next we need a device to extract the n- th coeffi cient from a generating function. As a notation, we define the coeffi cient n n functional [t ] on k [[t]] such that Cn = [t ] γ (t), or we could write Cn = [γ]n, realizing that the name of the formal variable is not of interest. Analysis has given us a simple method to find the n-th coeffi cient of an analytic function γ (t), Cn = 1 dn n! dtn γ (t) t=0. We can also define the formal derivative of a formal power series: n d γ (t) has coeffi cients d γ = (n + 1) C , and therefore C = 1 d γ (t) dt dt n n+1 n n! dtn t=0 can be defined for formal power series as well. We saw that convergence it not an issue for generating functions, because we never evaluate them at a specific t, except for t = 0, a case which does not need convergence. However, having some positive radius of convergence is a great help if we want to use γ (t) as a storage device. Consider the above example, the 1.1. Generating Functions and Linear Recursions 3
balanced sequences of 1’s and 1’s. In this case the numbers Cn are the Catalan numbers, probably one of the most− studied sequences in combinatorics! And it is well-known that ∞ 2 C tn = (1.1) n √ n=0 1 + 1 4t X − (see Exercise 2.3.8 for a proof), a series that converges for all t < 1/4, viewed as a function in t. Writing 2/ 1 + √1 4t is clearly a very convenient notation! − We can get back the numbers Cnfrom 2/ 1 + √1 4t as C0 = 2/ 1 + √1 = 1, − d 2 4 C1 = = 2 = 1, dt 1+√1 4t 1+√1 4t √1 4t − t=0 ( ) t=0 − − 1 d 4 3 √1 4t+1 C2 = 2 = 4 − 3 = 2, 2 dt (1+√1 4t) √1 4t (1 4t)3 /2(1+√1 4t) − − t=0 − − t=0 4 d 3√1 4t+1 2√1 4t+3 10t C3 = − 3 = 16 − − 4 = 5 , etc. 3 dt (1 4t)3/2(1+√1 4t ) ( 1+4t)2√1 4t(1+ √1 4t) − − t=0 − − − t=0 The above calculations should convince you that finding the coeffi cients by the differentiation method is a recursive procedure; from the n -th derivative you calculate the n+1-th derivative. You want a computer algebra package to find C100 this way. We can solve that recursion by finding Cn explicitly, which means that we have to expand 2/ 1 + √1 4t in powers of t. Of course, 2/ 1 + √1 4t = − − 1 √1 4t / (2t), a function that is slightly easier to expand. Thus − − 1 √1 4t ∞ 1/2 n+1 2n 1 n 1 − − = ( 1) 2 − t − . 2t n n=1 − X Substitute n 1 1/2 ( 1) − 2n 1 = − − n 22n 1 (2n 1) n − − 1 2n and get Cn = n+1 n . Note that for large n this formula needs a computer also; however, Cn can be fairly accurately approximated by Stirling’s formula, 2n 2n n / (n + 1) 2 / ((n + 1) √nπ), with a relative error of roughly 1/1000 if n is around 100. ≈ Multiplication of convergent power series is defined with the help of the Cauchy product, n n [t ](γ (t) σ (t)) = cksn k, − k=0 X k k if t γ (t) = ck and t σ (t) = sk. This definition is carried over to the formal power series. Note that multiplication is commutative, because multiplication in k is commutative. The reciprocal of a formal series γ (t) can exist only when γ (0) = 0; otherwise we would obtain negative powers of t. We write 1/γ (t) for the reciprocal,6 and 4 Chapter 1. Prerequisites from the Theory of Formal Power Series
1 sometimes γ (t)− . We have 1 = γ (t) (1/γ (t)); this tells us all about the coeffi cients n of 1/γ (t) = n∞=0 c˜nt : 1 = c0c˜0, hence c˜0 = 1/c0 (showing again that c0 = 0), P 2 6 0 = c1c˜0 + c0c˜1, thus c˜1 = c1/c0, and in general n 1 − c˜n = k=0− c˜kcn k /c0, for all n 1. − − ≥ There will only be a positive power of c in the denominator of the expression P 0 for c˜n. We can now refine our statement about the existence of a reciprocal. Lemma 1.1.1. A formal power series γ has a reciprocal iff γ (0) has a nonzero multiplicative inverse in the coeffi cient ring for all n 0. ≥ For example, if the coeffi cient ring equals Z, a power series must start with 1 or 1 in order to have a reciprocal. If we return to the Catalan generating function− c (t) = 2/ 1 + √1 4t , we get the reciprocal 1/c (t) = 1 + √1 4t /2, but something ‘surprising’happens− in this special case: − 1/c (t) = 1 tc (t) . (1.2) − The order of a power series σ (t) is the smallest n 0 such that [tn] σ (t) = 0. We saw above that if σ has a reciprocal then ord σ =≥ 0. A power series β (t6 ) of order 1 is called a delta series in the Finite Operator Calculus, if ord β (t) = 1 and β(t)/t has a reciprocal. Remark 1.1.2. We defined scalar multiplication, addition, and multiplication of formal power series in purely algebraic terms. For the reader interested in the combinatorics behind all this we recommend the 1981 paper by Joyal [48], and the book on Combinatorial Species and Tree-like Structures by F. Bergeron, G. Labelle, and P. Leroux [10]. How do we “combinatorially”calculate the coeffi cient k 1 2 of β (t) if k is a positive integer and β (t) = b1t + b2t + ... ? If we write
k n β (t) = t bj bj 1 ··· k n k j1+ +jk=n X≥ ···jXi>0 we can sort the vectors (j1, . . . , jk) of positive integers and obtain vectors (λ1, . . . , λk) of sorted integers, each with a certain multiplicity. Of course, it still holds that λ1 + + λk = n. If we sort such that λ1 λk, then λ = (λ1, . . . , λk) is called··· a partition of n. In symbols, λ n. The≥ number · · · ≥ of parts is k, λ = k. There ` | | is an equivalent representation of a partition as a multiset 1`1 , . . . , n`n , where `i n n the term i means `i occurrences of i in λ. Hence i=1 `i = k, and i=1 i`i = n. For every partition of n we think of these two equivalent representations simul- taneously! The above mentioned multiplicity is theP number of permutationsP of λ, k which is . (Choose `1 places for the ones in λ, then `2 places for the twos, `1,...,`n etc.) Hence k k n k β (t) = t bλi . (1.3) `1, . . . , `n n k λ n, λ =k i=1 X≥ ` X| | Y 1.1. Generating Functions and Linear Recursions 5
For example,
k k k k k t β (t) = bλi = b1 `1 λ k, λ =k i=1 ` X| | Y k k+1 k k k 1 t β (t) = bλi = kb2b1− (1.4) `1, `2 λ k+1, λ =k i=1 ` X| | Y k+2 k k 1 k 2 k 2 t β (t) = kb b − + b b − 3 1 2 2 1 k+3 k k 1 k k 2 k 3 k 3 t β (t) = kb b − + 2 b b b − + b b − . 4 1 2 3 2 1 3 2 1 There is a second concept of partitioning in combinatorics, the partitions of an n-set. Here a set of n-elements is written as the union of k nonempty and disjoint subsets. The number of such set partitions is S (n, k), the Stirling number of the second kind (Stanley [89]). We write S (n, k) for the set of all partitions of an n- set into k parts, thus S (n, k) = Sn,k The parts B1,...,Bk in a set partition are written in no particular order; we| think| of them sorted decreasingly by magnitude, B1 Bk . Parts with the same number of elements have to be sorted in | | ≥ · · · ≥ | | some way (lexicographically). Again, the numbers B1 Bk will make a partition λ of n, with k parts, but with a certain multiplicity.| | ≥ · · · ≥ The | multiplicity| is n! , i.e., this is the number of partitions in Sn,k such that Bi = λi. λ1! λk!`1! `n! ··· ··· | | Hence (1.3 shows that β (t)k
k n k! n! = t λi!bλi n! λ1! λk!`1! `n! n k λ n, λ =k i=1 X≥ ` X| | ··· ··· Y k n k! = t λi!bλ n! i × n k λ n, λ =k i=1 ! X≥ ` X| | Y (B1,...,Bk) Sn,k such that Bi = λi × |{ ∈ | | }| k n k! = t Bi !b Bi . (1.5) n! | | | | n k (B1,...,Bk) Sn,k i=1 X≥ X∈ Y
Example 1.1.3 (Fibonacci numbers). The Fibonacci numbers Fn can be defined by F0 = F1 = 1 and Fn = Fn 1 + Fn 2 for all n 2. (Note that Z can serve as − − ≥ the coeffi cient ring.) However, defining Fn = 0 for negative n lets us express this recursion as Fn Fn 1 Fn 2 = δ0,n for all n 0. Multiplying this formula by tn and summing− up over− − all n− 0 gives the generating≥ function of the Fibonacci ≥ 6 Chapter 1. Prerequisites from the Theory of Formal Power Series numbers,
n n 2 1 = (Fn Fn 1 Fn 2) t = Fnt 1 t t − − − − − − n 0 n 0 X≥ X≥ n 2 thus n 0 Fnt = 1/ 1 t t . The roots of the denominator are easy to find; hence ≥ − − P 1 F = [tn] − n t2 + t 1 − n 1/√5 1/√5 = [t ] 1 1 1 1 t + + √5 − t + √5! 2 2 2 − 2 1 1 = n n √5 1 + 1 √5 1 √5 1 − √5 1 1 √5 1 √5 1 2 2 − 2 − 2 2 − 2 2 − 2 n+1 n+1 1 + √5 1 √5 = − − n+1 2 √ 5
On the first glance, this formula for Fn does not look integer, but a closer look will easily convince you. Of course we could also expand 1/ 1 t t2 in powers of t. − − We get F = n/2 n k , certainly an integer, and so we arrive at the identity n k=0 −k P n/ 2 n+1 n+1 n k 1 + √5 1 √5 − = − − k 2n+1√5 k=0 X n A recursion of the form σn = j=1 αjσn j + γn, where αj and γn are given − for all j 1 and n 0, is called a linear recursion for σn in one variable ( n). ≥ ≥ P The starting value in this recursion, σ0, equals γ0, and if γ1, γ2,... are different from 0, the recursion is called inhomogeneous. However, if the sequence of inho- mogeneous terms eventually becomes 0, so that the last nonzero term is γ` 1, then we usually do not say that the recursion is inhomogeneous, but has initial− values n σ0, σ1, . . . , σ` 1, and then follows the (homogeneous) recursion σn = j=1 αjσn j − − for n `. Of course, the terms γ0, . . . , γ` 1 can be recovered from the initial values, ≥ k − P as γk = σk j=1 αjσk j. − − Theorem 1.1.4.P Suppose the numbers σn solve for n 0 the (inhomogeneous) linear recursion ≥ n
σn = αjσn j + γn − j=1 X where α1, α2,... and γ0, γ1, ... are sequences of given constants. Then
n n k 1 σn = γk t − j 1 ∞ αjt k=0 j=1 X − P 1.1. Generating Functions and Linear Recursions 7 and k ∞ n k∞=0 γkt σnt = 1 ∞ α tj n=0 j=1 j X −P k k P j Proof. From ∞ γkt = ∞ σkt 1 ∞ αjt follows the Theorem. k=0 k=0 − j=1 The formP of the generating P function for (σPn) shows us that it will be rational if both (γn) and (αn) are eventually 0. This characteristic will be important in section 1.1.1. Explicit expressions can be derived from Theorem 1.1.4 in a large number of applications. We discuss Fibonacci-like sequences σn = uσn 1 + vσn 2 + γn in Exercises 1.1.2 and 1.1.3. − − n j The sequences 1, a1, a2,... and [t ] 1/ 1 j∞=1 αjt are some- − − − n 0 ≥ times called orthogonal (because they are reciprocals), P and the sequences (σn) and (γn) are an inverse pair. For examples of inverse pairs see section 2.3.3.
Example 1.1.5 (Derangement Numbers). The derangement numbers dn denote the number of permutations π of [n] that are derangements, i.e., πi = i for all i = n 6 1, . . . , n. They follow the recursion dn = ndn 1 + ( 1) ,with initial value d0 = 1. − − n In the notation of Proposition 1.1.4, σn := dn/n!, αn = δn,1, and γn = ( 1) /n!. Thus −
n k n k ( 1) n k 1 ( 1) dn = n!σn = n! 1 + − t − = n! − k! 1 α1t k! k=1 ! k=0 X − X k k n ∞ n k∞=0 ( 1) t /k! t dnt /n! = σnt = − = e− / (1 t) 1 t − n 0 n=0 P X≥ X − The generating function is not rational.
Example 1.1.6 (Bernoulli Numbers). The Bernoulli numbers Bn solve the system n n+1 1 of equations δn,0 = k=0 Bn k for n 0. Dividing by n! shows that the k+1 n+1 − ≥ numbers Bn/n! can be calculated from the linear recursion P n 1 Bn/n! = Bn k/ (n k)!. − (k + 1)! − − k=1 X
Applying the proposition with ak = 1/ (k + 1)! for k 1 gives the (exponential) generating function − ≥
B B t n tn = 0 = n! 1 + tj/ (j + 1)! t + tj+1/ (j + 1)! n 0 j 1 j 1 X≥ ≥ ≥ t t = P = . P t + (et 1 t) et 1 − − − 8 Chapter 1. Prerequisites from the Theory of Formal Power Series
The generating function is not rational. The Bernoulli polynomials ϕn (x) := n k n Bk x − t xt k=0 k! (n k)! have generating function et 1 e by convolution. Note that − − P t (x+1)t t xt t x( t) e = e = − e− − et 1 1 e t e t 1 − − − − − n hence ϕn (x + 1) = ( 1) ϕn ( x). See Exercise 2.3.2 for more details on Bernoulli numbers. An explicit− formula− for the Bernoulli numbers is obtained in Exercise 2.3.15. A detailed discussion of Bernoulli numbers (58 pages) can be found in the “Calculus of Finite Differences” by Jordan [47]. Example 1.1.7. A polyomino is a union of a finite number of squares with vertices in Z2 such that every square shares at least one side with some other square. Translations do not change a polyomino, but reflections and rotations do. In a horizontally convex polyomino P any line segment parallel to the x-axis with both end points in P must be completely in P . For example, there are 19 polyominos with 4 squares; all 19 polyominos are horizontally convex.
Hickerson [42] found a combinatorial proof in one dimension that the number f (n) of horizontally convex polyominos made of n+1 squares follows the recursion f (n + 3) = 5f (n + 2) 7f (n + 1) + 4f (n) − for n 1, with initial values f (0) = 1, f (1) = 2, f (2) = 6, f (3) = 19. By Proposition≥ 1.1.4
` 1 n 1 − k f (n) t = j σk αjσk j t 1 ∞ αjt − − n 0 j=1 k=0 j 1 X≥ − X X≥ 1 +P (2 5) t + (6 10 + 7) t2 + (19 30 + 14 4) t3 = − − − − 1 5t + 7t2 4t3 − − (1 t)3 = − 1 5t + 7t2 4t3 − − This generating function begins with 1 + 2t + 6t2 + 19t3 + 61t4 + 196t5 + 629t6 + 2017t7 + 6466t8 + 20 727t9 + 66 441t10 + 212 980t11 + 682 721t12 + 2188 509t13 + 7015 418t14 + ... 1.1. Generating Functions and Linear Recursions 9
We could expand the generating function in terms of tn, but we will find a “simple” expression for these coeffi cients in Example 1.1.8 Of course, deriving the recursion is the hard part. Stanley [89, p. 259] ob- tained the recursion from the generating function, which he found by the transfer- matrix method. His approach starts from the observation that f (n 1) = (n1 + n2 1) (n2 + n3 1) + + (ns + ns+1 1), summing over n− 1 − − ··· − all 2 − compositions of n1 + +ns+1 = n ( s = 0 contributes 1). For example, if n = 5, thenP the composition 5···(s = 0; 1 term) contributes 1, the compositions into 2 terms (s = 1) contribute 16, then 3 terms contribute also 16, 4 terms add 15, 5 terms 12, and 6 terms contributes 1. For example, f (4) = 1+16+16+15+12+1 = 61.
1.1.1 Roots
Suppose the recursion is homogeneous (γk = δ0,k and ` = 1). If there are only n finitely many factors α1, . . . , αd, the generating function n 0 σnt is rational. By Stanley’sTheorem 4.1.1 [89] ≥ P k n σn = pi (n) ri (1.6) i=1 X where the 1/ri’sare the k distinct roots of the polynomial
d k j mi 1 αjt = (1 rit) − − j=1 i=1 X Y and each pi (n) is a polynomial (in n) of degree less than the multiplicity mi of root 1/ri. The polynomials pi (n) can be determined from the the first few values of σn. We followed Stanley’s procedure in the example of the Fibonacci numbers, disguised as a partial fraction decomposition. We will do partial fraction decomposition again in the more involved next example. Example 1.1.8. Applying the root formula (1.6) is possible even when the numer- ator of the generating function is a polynomial different from 1. In Example 1.1.7 we found the generating function for the number f (n) of horizontally convex poly- ominos made of n + 1 squares,
3 2 n (1 t) (t 1) (1 t) f (n) t = − 2 3 = − − 1 5t + 7t 4t 4 (t t1)(t t2)(t t3) n 0 X≥ − − − − − With r = 3 71 + 6√177we determine the roots of the denominator as 11 r2+7r r2 11+14r r2+11 t1 = − , t2 = − i√3 , and t3 = t∗, the complex conjugate of 12rp 24r − 24r 2 t2 (the calculations were made with an algebra package; it can be cumbersome to 10 Chapter 1. Prerequisites from the Theory of Formal Power Series do this by hand!). We wrote the generating function already in a form to suggest a partial fractions decomposition,
t 1 A B C f (n) tn = − + + 4 t t1 t t2 t t n 0 2∗ X≥ − − − 2 We need A (t t2)(t t∗) +B (t t1)(t t∗) +C (t t2)(t t1) = (1 t) , thus − − 2 − − 2 − − − A + B + C = 1
A (t2 + t∗) B (t1 + t∗) C (t1 + t2) = 2 − 2 − 2 − − At2t2∗ + Bt1t2∗ + Ct1t2 = 1
This system has the solution 2 2 (5r+r2 11) ( r( 5 i√3+ 5 )+r2 11 i√3+ 11 ) A = 1 − , B = 1 − 2 2 − 2 2 , and C = B . 3 r4 11r2+121 3 (r2 11 i√3 11 )(r2+11) ∗ − − 2 − 2 If we write ξ = i√3 1 /2, ξ∗ = i√3 + 1 /2 (both are third roots of unity) 2 − 2 − (5rξ∗+r 11ξ) 1 − then B = 3 r4 11 ξr2+121ξ . The generating function can be written as − ∗ (1 t) A B f (n) tn = − + 2 Re 4 t t1 t t2 n 0 X≥ − − − For n 1 we find ≥ (1 t) A B f (n) = [tn] − + 2 Re 4 t t1 t t2 − − − 1 n A B 1 n 1 A B = [t ] + 2 Re + t − + 2 Re −4 t t1 t t2 4 t t1 t t2 − − − − A n 1 n B n 1 n = t− − t− + 2 Re t− − t− 4 1 − 1 4 2 − 2 A n 1 B n 1 = t− − (1 t1) + 2 Re t− − (1 t2) 4 1 − 4 2 −
2 3 2 3 (5r+r 11) ξ(11ξ 5rξ∗ r ) 1 − − − With A (1 t1) = 36 (r4 11r2+121)r and B (1 t2) = 36(r4 11ξr2+121ξ )r we get − − − − ∗
2 3 2 3 1 5r + r 11 n 1 ξ 11ξ 5rξ∗ r n 1 f (n) = − t− − 2 Re − − t− − . 144r r4 11r2 + 121 1 − (r2 11ξ )(r2 + 11) 2 − − ∗ !! 1.1.2 Exercises 1.1.1. Find an explicit expression for the number of horizontally convex polynomi- als, i.e., expand the generating function (1 t)3 / 1 5t + 7t2 4t3 in terms of t. − − − 1.1. Generating Functions and Linear Recursions 11
1.1.2. A Fibonacci-like sequence σ0, σ1,... solves a recurrence of the form
σn = uσn 1 + vσn 2 + γn − − for n 2, with initial values σ0, σ1, and given inhomogeneous terms γ2, γ3,... . Show≥ that k σ0 + (σ1 uσ0) t ∞ γkt − + k=2 . (1.7) 1 ut vt2 1 ut vt2 − − P− − is the generating function of σn. For example, the Fibonacci recursion Fn = Fn 1+ − Fn 2 has initial values F0 = 1, F1 = 1; this means that in Proposition 1.1.4 we − have the terms γn = δ0,n and α1 = 1, α2 = 1, αn = 0 for all n 2. Hence n n k 1 n 1 ≥ Fn = γk t j = [t ] 2 . Show that for any Fibonacci-like k=0 − 1 ∞ αj t 1 t t − j=1 − − sequence holds P P n n k − n k j n k 2j j σn = γk −j − u − − v (1.8) k=0 j=0 X X where γ0 = σ0 and γ1 = σ1 u. Derive sums like (1.8) for the following special cases: −
1. The Lucas recursion Ln = Ln 1 + Ln 2 for all n 2, L0 = 2 and L1 = 1. − − ≥ 2. The recursion σn = (a 1) σn 1 + aσn 2 + n 1 for all n 2, and a = 1. − − − − ≥ 6 3. The recursion Pn+1 (x) = 2xPn (x) + Pn 1 (x) for the Pell polynomials [69], − with P0 (x) = 1 and P1 (x) = x. The polynomials have the generating function
1 xt 1 1 + t2 P (x) tn = − = + (1.9) n 1 2xt t2 2 2 4xt 2t2 n 0 X≥ − − − − 1 1 + t2 1 = + 2 2 (1 t2) 1 2xt/ (1 t2) − − −
The Pell numbers pn are defined as pn = Pn (1). How do different values for p1 change the subsequent numbers?
4. The recursion Un (x) = 2xUn 1 (x) Un 2 (x) for the Chebychev polynomials − − − of the second kind, with U0 (x) = 1 and U1 (x) = 2x (see (7.17)). In Exercise 1.1.4 the root formula (1.6) is applied to the above cases. More related results on Chebychev polynomials in Exercise 7.1.13.
2 1.1.3. Show that for all Fibonacci-like sequences (Exercise 1.1.2) holds σn = n n σn 1σn+1 + ( 1) v c for all n 1. Determine the constant c, and show that − − ≥ 2 n 2 for the Pell polynomials holds Pn (x) = Pn 1 (x) Pn+1 (x) + ( 1) 1 + x , for 2 − n − the Fibonacci numbers Fn = Fn 1Fn+1 + ( 1) , and for the Chebychev polynomi- 2 − − als Un (x) = Un 1 (x) Un+1 (x) 1. − − 12 Chapter 1. Prerequisites from the Theory of Formal Power Series
1.1.4. Consider again the Fibonacci-like numbers defined in Exercise 1.1.2,
σn = uσn 1 + vσn 2 + γn − − for n 2, with initial values σ0, σ1, and given inhomogeneous terms γ2, γ3,... . ≥ This time we want an expression for σn in terms of the roots of the polynomial d j 2 1 αjt = 1 ut vt . Show that − j=1 − − r+1 r+1 P 2 2 r 1 r 1 u + √4v + u u √4v + u [t ] = 2− − − − 1 ut vt2 √4v + u2 − − if the discriminant 4v + u2 = 0. If 4v + u2 = 0 then 1 ut vt2 = (ut 2)2 /4 and 6 − − − 1 u r [tr] = (r + 1) . (1 ut/2)2 2 − Derive expressions in terms of the roots for the special recursions in Exercise 1.1.2 .
1.1.5. In the Fibonacci sequence let Fn = 0 for negative n. Show that Fn = k LkFn k ( 1) Fn 2k for all n 2k 1 > 0, if the numbers Lk are the Lu- − − − − ≥ − cas sequence following the Fibonacci-like recursion Lk = Lk 1 + Lk 2 with initial − − values L0 = 2,L1 = 1, hence
k k 1 + √5 + 1 √5 L = − . k 2k 1.2. Composition and Inverses 13
1.2 Composition and Inverses
Before we go deeper into the structure of formal power series, we want to introduce a notation. Remember that k is an integral domain: k has a unit element 1, is commutative and has no zero divisors, ab = 0 when a and b are in k and both 6 different from 0. We write k [t] for the polynomials in t with coeffi cients in the ring k, and k [[t]] for the formal power series with coeffi cient ring k as before. Note that we can view k [t] as being imbedded in k [[t]]. However, any element a from k can be used to evaluate a polynomial at a, giving again an element from k. In k [[t]], evaluation is only possible at a = 0. We saw that σ (0) k must have a ∈ multiplicative inverse in k if 1/σ (t) exists. Such an element of k is called a unit. Hence 1/σ (t) exists over an integral domain iff σ (0) is a unit in k. Composition of a power series γ (t) into a power series β (t) is achieved by substituting γ (t) for t in β (t). We write β (γ (t)) for the composition. Because the composition has to be in k [[t]], the term β (γ (0)) has to be in k. But only if β is a polynomial we know β (a) for a k and a = 0. Hence we will require that γ (t) is not of order 0, so γ can be substituted∈ into6 any power series β. Underlining the operational aspect we also write C(γ)β, where C (γ): k [[t]] k [[t]] is the linear operator that does the substitution of γ (t). → In analysis we have a concept of convergence; we know that eln t = t. However, ln t is not a formal power series (it is not defined at 0, the only place we can evaluate a formal power series). We can say that eln(1+t)) 1 = t, which means that C(ln (1 + t)) (et 1) = t. In other words, ln (1 + t) is the−compositional inverse of − et 1. If a power series β is of order 1, and [t] β (t) is a unit in k [t], then β − 1 1 has a unique compositional inverse β− such that β β− (t) = t. Such power series we called delta series. Often it is more convenient to use another symbol, 1 like γ, for the inverse of β. Note that the notation β− (t) for the (compositional) 1 inverse is very similar to the notation of the reciprocal β (t)− = 1/β (t), and it becomes indistinguishable, if the argument t is omitted. This shows why we like the notation 1/β for the reciprocal. Of course, they usually do not exist both for the same β k [[t]]! ∈ Example 1.2.1. Consider the sequence 0, 1, 2, 3, 4,... by looking at the formal n+1 n − − power series β (t) = n 1 ( 1) nt (a delta series in Z [[t]]); we want to know its compositional inverse.≥ We− have Pn+1 n d 1 2 β (t) = t n 0 ( 1) (n + 1) t = t dt 1+t = t/ (1 + t) , hence − ≥2 − 1 − t = 1 + 2t + t β and t = 2β 1 √1 4β 1. Therefore, the compositional P1 − − − inverse β− (t) equals the generating function of the Catalan numbers (section 1 1 1.1) without the constant term, β− (t) = 1 √1 4t 1. 2t − − − Remember that every statement about formal power series really means a statement about an infinite sequence of coeffi cients. For a series of order 1, the sequence starts with 0, and then a coeffi cient not equal to zero follows. Suppose we have a power series β with coeffi cients (0, a, b, c, . . . ). The compositional inverse 14 Chapter 1. Prerequisites from the Theory of Formal Power Series
γ has coeffi cients 0, a,˜ ˜b, c,˜ . . . such that the composition gives (0, 1, 0,... ). For calculating the first few term of the composition we need the n-th coeffi cient of k 2 ˜ ˜2 the k-th power of the inverse, γ n. For k = 2 we get 0, 0, a˜ , 2˜ab, b + 2˜ac,˜ . . . k and for = 3 we have 0, 0, 0, a˜3, 3˜a˜b + 3˜ac˜ + 3˜bc,˜ . . . . Hence β (γ) has the coeffi cient sequence (0, aa,˜ a˜b + ba˜2, ac˜ + 2ba˜˜b, ad˜ + b ˜b2 + 2˜ac˜ + ca˜3,... ), and this has to equal 1 (0, 1, 0,... ). We can now recursively find the coeffi cients of β− = γ,
0, 1/a, b/a3, 2b2 ac /a5, ad2 5abc + 5b3 /a7,... (1.10) − − − Algebraically it is important that the powers of the coeffi cient a have a nonzero multiplicative inverse in k, if we want to find the compositional inverse of a power series. This is the only coeffi cient we divide by repeatedly; the remaining terms are sums of products. Hence, a compositional inverse γ (t) will exist, if β (t) is a delta series! Lagrange-Bürmann inversion, which provides a formula for [tn] γk, has been shown for this general setting ([41, Theorem 1.9a], [43]). We only state the result, assuming that β (t) = t/φ (t), which means that φ (t) is invertible. In this way, Laurent series are not needed. For more on Lagrange-Bürmann inversion see Exercise 1.2.2. Theorem 1.2.2. If β (t) = t/φ (t) is a delta series (hence φ has a reciprocal), and γ (t) is the compositional inverse of β, then for all 0 k n holds ≤ ≤ k n n γ n = k [φ ]n k . (1.11) − Again we want to point out that the name (t) of the formal variable is of no significance. We can as well write (1.11) as
n k n k n n [s ] γ(s) = k u − φ (u) , where γ k [[s]] and φ k [[u]]. For example, in finding the inverse of β (t) = et 1 we usually∈ proceed by∈ letting u = et 1, say, and solving for t = ln (1 + u). Then− 1 − we call t = β− (u), and change the variable from u back to t. Note that the presentation (1.10) is variable free! Lagrange-Bürmann inversion is routinely applied in combinatorics. We show an example from lattice path enumeration. Example 1.2.3. We are interested in finding the number D (n; k) of , lattice path from (0, 0) to (2n, 0), staying weakly above the x-axis, and having{% &} exactly k occurrences of the pattern %& , which we also write as uddu (d = and u = ). The following path to (10&%, 0) contains the pattern uddu twice (counting& overlaps).% %& %&%& %&%& 1.2. Composition and Inverses 15
n k The generating function F (x, t) = n,k 0 D (n; k) t x is a power series in two variables, so it would be beyond the scope of≥ this chapter. However, certainly k n, so we can see F (x, t) as a power seriesP in t with coeffi cients that are polynomials≤ in x. If we set D (0; 0) = 1, then
tF 3 ((1 x) t + 1) F 2 + (1 + 2 (1 x) t) (1 x) t = 0 − − − − − (see [84]). The need for an inverse arises, because we can easily write t as a function of F , F (F 1) t = − F 3 (1 x)(F 1)2 − − − but we want F as a function of t. Remember that we can only invert a delta series, so we define u = F (x, t) 1. Thus − (u + 1) u u β (u) := t = 3 = (u + 1) (1 x) u2 (u + 1)3 (1 x) u2 / (u + 1) − − − − n and we need the inverse, u = γ (t). We have to check that the negative powers a− of the linear term in β (u) are different from 0., We could get a by differentiating, or by noting that (u + 1) / (u + 1)3 (1 x) u2 starts with 1, hence β (u) = − − u + ... and therefore a = 1. We can now find the compositional inverse γ of β from (1.11) as n 3 2 n 1 (u + 1) (1 x) u n [γ]n = u − − − u + 1 ! Routine calculations give
(n 1)/2 i b − c n 2n 3i i k i k n [γ] = − x ( 1) − n n i n 2i 1 k − i=0 k=0 X − − − X (n 1)/2 1 b − c n 2n 3i i i k D (n; k) = − ( 1) − n n i n 2i 1 k − i=k X − − − The sample path above is one of 4077 counted by D (5; 2). Remark 1.2.4. It is not true that a power series must be of order 1 to have an inverse. For example, the power series 1 + t has the inverse t 1. Add your own examples in Exercise 1.2.1. −
1.2.1 Exercises
1.2.1. Find power series in R [[t]] that are of order 0 but have a compositional inverse. 16 Chapter 1. Prerequisites from the Theory of Formal Power Series
1.2.2. Write the Lagrange-Bürmann inversion formula (1.11) in terms of β in- stead of φ (t) = t/β (t). This formulation needs negative powers of β, which only n exist in the field of Laurent series, where series of the form σ (t) = n k σnt ≥ are defined for all k Z. An important functional on Laurent series is the 1 ∈ P residue, res(σ) = t− σ (t). Note that res (σ0 (t)) = 0 for all Laurent series σ (t). k Let ρ (t) = k 0 rkt be any formal power series. Show the original Lagrange- Bürmann Theorem,≥ which says that for any delta series β (t) holds P 1 ρ (γ) = r + res ρ β n tn 0 n 0 − n 1 X≥ n 1.2.3. Show that for every power series σ (t) = n 0 σnt holds ≥ 1 t n n n σ = n 0 t k=0 σn k. P 1 t 1 t k − − − ≥ P P 1.2.4. The smallest nontrivial integral domain is Z2, the integers modulo 2. + 0 1 0 1 0 0 1 0∗ 0 0 1 1 0 1 0 1 Addition and multiplication in Z2
1 Show β (t) = β− (t) holds for more power series than just β (t) = t when k = Z2. 1.2.5. If k is an integral domain, then k [[t]] is also an integral domain. 1.2.6. Show that for the Catalan generating function c (t) = 2/ 1 + √1 4t holds n 1 − 1/c (t) = 1 tc (t). Hence Cn = k=0− CkCn 1 k for all n 1. − − − ≥ 1.2.7. [10, Chapter 3.2]Let γ(t) =Pt + G (γ(t)), where 0 = G (0) = G0 (0). Then
k k 1 j 1 k 1 j γ (t) = t + D − kt − G (t) for all k 1. j! ≥ j 1 X≥ 1.3. Multivariate Power Series 17
1.3 Multivariate Power Series
Let r be an integer larger than 1. The r-dimensional array
S = (σn ,...,n ) r 1 r (n1,...,nr ) N ∈ 0 can be represented as the formal power series
n1 nr σ (t1, . . . , tr) = σn1,...,nr t1 tr . r ··· (n1,...,nr ) N X ∈ 0
n1 nr The coeffi cient functional [t1 tr ] σ = σn1,...,nr recovers the coeffi cients of the ··· ∂ series. If they are in k, we say that σ k [t1, . . . , tr]. The partial derivatives ∈ ∂ti are defined on k [[t1, . . . , tr]] as ∂ n1 ni 1 nr σ = niσn1,...,nr t1 t − tr ∂ti r ··· ··· (n1,...,nr ) N X ∈ 0 and we have ∂n1+ +nr ··· n1 nr n1 nr σ (0,..., 0) = n1! nr![t1 tr ] σ = n1! nr!σn1,...,nr . ∂t ∂tr ··· ··· ··· 1 ··· Similar to the univariate case, evaluation of σ is only allowed by setting some or all of the ti’s equal to 0. However, there is a new concept in multivariate power series: we can equate some of the formal variables, making them equal to a new formal variable s, say. For example,
m n3 n4 [s t3 t4 ] σ (s, s, t3, t4) = σi,j,n3,n4 . i+j=m X n A univariate formal power series φ (w) = n 0 φnw , on the other hand, can be made into a multivariate series by replacing≥ the formal variable by a linear combination of new formal variables. For example,P m + n [smtn] φ (s + t) = φ . m m+n For notational simplicity, we continue the discussion of multivariate formal power series in the bivariate case. Addition of bivariate power series is defined as expected; multiplication needs the Cauchy product as in the univariate case, m n m n γ (s, t) σ (s, t) = s t γi,jσm i,n j − − m,n 0 i=0 j=0 X≥ X X m n 0 0 when [s t ] γ (t) = γm,n. We will say that γ is of order 0 iff s t γ is different from 0. If s0t0 γ is a unit in k, then γ (s, t) has a reciprocal τ (s, t) such that 1 γ (s, t) τ (s, t) = 1. We denote the reciprocal by 1/γ (s, t), or γ (s, t)− . 18 Chapter 1. Prerequisites from the Theory of Formal Power Series
Suppose the numbers σm,n solve for (m, n) N N the (inhomogeneous) linear recursion ∈ × m n γm,n = αi,jσm i,n j − − i=0 j=0 X X where (αi,j)i,j 0 and (γi,j) are double sequences of given constants, and α00 = 1. We obtain, of≥ course, the generating function identity
γ (s, t) σ (s, t) = . α (s, t)
The above linear recursion is equivalent to
σm,n = γm,n αi,jσm i,n j (1.12) − − − (0,0)<(i,j) (m,n) X≤ where we consider the partial order (i, j) (k, l) iff i j and k l on Z2. Note that (k, l 1) and (k 1, l) are both less≤ than (k, l). In≤ order to≤ calculate σ (s, t) − − as the simple expression γ (s, t) /α (s, t) the numbers σm,n must depend on the 2 numbers σi,j in a cone in N ; each σm,n depends only on σi,j for (0, 0) (i, j) (m, n) and (i, j) = (m, n). For more discussions of higher dimensional recursions≤ ≤ see Bousquet-Mélou6 and Petkovšek in [16]. Example 1.3.1. The lattice walk in the first quadrant with steps , , can cycle back to itself, so we better count the number of ways that such{% a walk← reaches↓} (m, n) in k steps, starting at (0, 0). We call the number of walks d (m, n; k). The obvious recursion
d (m, n; k) = d (m + 1, n; k 1) + d (m, n + 1; k 1) + d (m 1, n 1; k 1) − − − − − is not of the form (1.12). Bousquet-Mélou and Petkovšek show how the “kernel method” of generating functions can be applied to solve this problem. It has been solved before in other ways by Kreweras [53, 1965]. The notation for power series in r > 1 variables can be a challenge; we avoid it by considering just two variables, in some examples three, but leave the general case to the reader. A multi-series (ρ, σ) is a pair (or more general an r-tuple) of formal power 2 series in two (or r) variables, (ρ, σ) k [[s, t]] . We say that (γ1, γ2) is a delta multi- ∈ series iff γ1 (s, t) = sφ (s, t) and γ2 (s, t) = tψ (s, t) where φ (0, 0) and ψ (0, 0) are units in k, and φ (s, t) and ψ (s, t) are in k [[s, t]] (both having order 0). Thus γ1 (0, t) = 0, and γ1 (s, 0) is a power series in s. Analogously, γ2 (s, 0) = 0 and γ2 (0, t) k [[t]]. We∈ need the concept of the compositional inverse of a delta multi-series. The compositional inverse of the delta multi-series (β1, β2) is the multi-series (γ1, γ2) such that β1 (γ1 (s, t) , γ2 (s, t)) = s and β2 (γ1 (s, t) , γ2 (s, t)) = t. The inverse of 1.3. Multivariate Power Series 19
a delta multi-series is also a delta multi-series. If (γ1, γ2) is inverse to (β1, β2), then (β1, β2) is also inverse to (γ1, γ2). The inverse of (β1 (s, t) , β2 (s, t)) is usually 1 1 denoted by β1− (s, t) , β2− (s, t) . The partial derivatives ∂ φ (s, t) and ∂ φ (s, t) of a bivariate power series ∂s ∂t φ (s, t) inherit their properties (for example, the product rule) from the uni- variate case. We remember from Calculus that the derivative of a multi-series γ = (γ1, . . . , γr) is defined as γ , where γ stands for the Jacobian. In the bivariate case, |J | J
∂ ∂ ∂(γ1, γ2) ∂s γ1 (s, t) ∂s γ2 (s, t) γ = = ∂ ∂ |J | ∂ (s, t) γ1 (s, t) γ2 (s, t) ∂t ∂t
∂ ∂ ∂ ∂ = γ1 γ2 γ2 γ1 . ∂s ∂t − ∂s ∂t The following multivariate Lagrange-Good inversion formula needs the Jaco- bian determinant.
Theorem 1.3.2. If γ1 (s, t) , γ2 (s, t) is a multi-series with compositional inverse β1 (s, t) , β2 (s, t), where we can write γ1 (s, t) = s/ε1 (s, t) and γ2 (s, t) = t/ε2 (s, t) with ε1 and ε2 of order 0, then
k l m+1 n+1 β1 (s, t) β2 (s, t) = ε1 (s, t) ε2 (s, t) γ . m,n |J | m k,n l h i h i − − For an elegant proof in Finite Operator Calculus terms see Hofbauer [43]. Note that the Theorem already assumes that an inverse exists: We have written it in a form that forces γ1 and γ2 to be a pair of delta series. Hence γ = 0 |J | 6 (Exercise 1.3.1), and an inverse pair of delta series will always exist. If β1 (s, t) = s/φ1 (s, t) and β2 (s, t) = t/φ2 (s, t) with (φ1, φ2) of order 0, then the Lagrange- Good inversion formula can also be written as
k l m+1 k n+1 l φ1 (s, t) φ2 (s, t) = ε1 (s, t) − ε2 (s, t) − γ (1.13) m,n |J | m,n h i h i (Exercise 1.3.3). Example 1.3.3. A multiseries does not need to be a delta series for having an inverse. Suppose γ1 (s, t) = s and γ2 (s, t) = s+t. We find Jγ = 1, and β1(s, t) = | | s, β2 (s, t) = t s. − Example 1.3.4. The pair γ1 (s, t) = as/ (1 bt) and γ2 (s, t) = at/ (1 bs) has the inverse pair − −
1 1 1 1 a− bt 1 1 a− bs β (s, t) = a− s − and β (s, t) = a− t − 1 1 a 2b2st 2 1 a 2b2st − − − − (Exercise 1.3.4). Clearly, a needs to be a unit. Only nonnegative powers of b occur in the expansion of β1 and β2. Hence b does not has to be a unit. 20 Chapter 1. Prerequisites from the Theory of Formal Power Series
1.3.1 Exercises
1.3.1. Show that for any delta multi-series γ1 (s, t) and γ2 (s, t) holds γ = 0. |J | 6 1 1 1.3.2. Show that the compositional inverse β1− (s, t) , β2− (s, t) of a delta series (β1 (s, t) , β2 (s, t)) is also a delta series. 1.3.3. Show the inversion formula (1.13).
1.3.4. Let γ1 (s, t) = as/ (1 bt)and γ2 (s, t) = at/ (1 bs), where a is a unit. Apply Theorem 1.3.2 to show− that −
1 1 1 1 a− s 1 a− bt a− t 1 a− bs − , − 1 a 2b2st 1 a 2b2st − − − − is the pair inverse to (γ1, γ2). Convince yourself that b does not have to be invertible in k by choosing k = Z. 1.3.5. Show that the Lagrange-Good formula in Theorem 1.3.2 is equivalent to
∂ ∂ k l m n 1 γ1 ∂s ε1 γ2 ∂s ε2 β1 (s, t) β2 (s, t) = ε1 (s, t) ε2 (s, t) − ∂ − ∂ m,n γ1 ε1 1 γ2 ε2 − ∂t − ∂t m k,n l h i − −
1.3.6. Show that the multiseries (1 + s + t, 1 + s t) of order 0 has an inverse. − 1.3.7. Let φ and ψ be bivariate power series in (s, t), and let θ and η be a delta pair. Show that ∂ (φ, ψ) ∂ (φ, ψ) ∂ (θ, η) = ∂ (s, t) ∂ (θ, η) ∂ (s, t)
(hint: ∂φ = ∂φ ∂θ + ∂φ ∂η ). ∂s ∂θ ∂s ∂η ∂s 1.3.8. The mapping β (s) (β (s) , t) is an embedding of k [[s]] into k [[s, t]]. If β (s) = s/φ (s) is a delta series7→ and γ (s) its inverse, derive the univariate inver- sion formula (1.11)
k n n γ n = k [φ ]n k for 0 k n − ≤ ≤ from the bivariate Lagrange-Good formula in Theorem 1.3.2 applied to (β (s) , t). Chapter 2
Finite Operator Calculus in One Variable
The isomorphism between finite operators on polynomials, defined in section 2.2, and formal power series is central to the Finite Operator Calculus. It allows for transfer theorems, enabling us to find basic polynomial solutions to certain opera- tor equations, and for a “functional expansion theorem”, helping us to determine a specific solution under given initial conditions. All our operators act on polynomials, and it is a recursion on polynomials, that defines the operator equation. Therefore, we begin by looking at polynomials in section 2.1. Even if the set of operators we study seems to be “small”compared to all operators imaginable, this set - the delta operators - has a powerful property, the translation invariance. Rota et al. studied it in detail [83, 1973], after an earlier paper with Mullin [62, 1970], and followed by work with Roman [81, 1978].
2.1 Polynomials, Operators, and Functionals
One main ingredient in the Finite Operator Calculus are bases of polynomials. Therefore our polynomials must come from a vector space where every sequence (pn) of polynomials of degree n is a basis, and therefore their coeffi cients must lie in a field F; we must allow division by any coeffi cient except 0. We will also assume that F has characteristic 0, hence the field is infinite. We could define our polynomials as having coeffi cients in an integral domain k, and define a basis such that the leading coeffi cient in a basis polynomial is a unit. For example, if our polynomials have integer coeffi cients, we would have the basis xn : n 0 . { ≥ } However, we will embed these polynomials in Q [x], and find many more bases. 22 Chapter 2. Finite Operator Calculus in One Variable
2.1.1 The Vector Space of Polynomials, and Their Bases Pascal’s Triangle recurrence can serve as a prototype of a recurrence in two vari- ables, pn (m) = pn (m 1) + pn 1 (m) . (2.1) − − We present it in a rectangular array, shown on the left of the following tables. 1 m 1 6 21 56 126 1 1 4 1 5 15 35 70 1 2 1 3 1 4 10 20 35 1 3 3 1 2 1 3 6 10 15 1 4 6 4 1 1 1 2 3 4 5 1 5 10 10 5 1 0 1 1 1 1 1 0 1 2 3 4 n Pascal’sTriangle in its usual form pn (m) written as an array The columns on the right side of this table look like values of polynomials in n+m m, of degree n, and because we know that pn (m) = n this is easily shown. Actually, we know that the recurrence (2.1) must have a polynomial solution, as long as p0 (m) = 1 for all m 0, because of the following Theorem. This Theorem helps to decide when a given≥ difference recursion (like (2.1), written as pn (m) pn (m 1) = pn 1 (m)) has a solution in terms of polynomials. − − − Theorem 2.1.1. Let x0, x1,... be a sequence of given initial points, xi Z, and ∈ let Fn (m) be recursively defined for all integers n 0 by ≥ n
Fn (m) = Fn (m 1) + aiFn i (m + bi) − − i=1 X where ak, bk F, for k = 0, 1,... , a1 = 0, and integers m > xn. Let Fn (m) = 0 ∈ 6 for all m < xn. We assume that bi xn i xn 1 for all n 1, and i = 1, . . . , n. ≥ − − − ≥ If F0 (m) is a constant different from 0 for all m x0, and if Fn (xn) is a known ≥ “initial value” at xn for all n, then there exists a sequence of polynomials fn (x) such that Fn (m) = fn (m) for all m xn and n 0 (inside the recursive domain). ≥ ≥ Proof. We can extend F0 (m) to all x F by making it the same constant every- ∈ where, f0 (x) := F0 (x0). Note that the new values (for x < x0) have no effect on the recursion F0 (m) = F0 (m 1) for all m > x0. Now suppose Fn (m) has been extended to a polynomial of− degree n for all m and for all 0 i n. Then ≤ ≤ Fn+1 (m) Fn+1 (m 1), the backwards difference, is a polynomial of degree n in − − m for all m > xn+1, as long as the right hand side of Fn+1 (m) Fn+1 (m 1) = n+1 − − i=1 aiFn+1 i (m + bi) is a sum of polynomials with highest degree n. This is − the case if all the terms Fn+1 i (m + bi) refer to x-values above the correspond- P − ing initial points, i.e., m + bi xn+1 i for all i = 1, . . . , n, and m > xn+1, thus ≥ − bi xn+1 i xn+1 1. But if the backwards difference of a function is a polyno- mial,≥ then− the− function− itself can be chosen as a polynomial of one degree higher 2.1. Polynomials, Operators, and Functionals 23
(see [47]), Fn+1 (m) = fn+1(m), say, for all m > xn+1. We can actually express this polynomial for m > xn as
n m
fn (m) = fn (xn) + ai fn i (k + bi) (2.2) − i=1 k=x +1 X Xn and we can find fn (x) for other values of x by Lagrange interpolation (the La- grange interpolation formula is applied in the proof of Lemma 2.2.12).
m 1 7 17 21 21 6 1 6 12 13 13 5 1 5 8 8 8 4 1 4 5 5 5 3 1 3 3 3 3 2 1 2 2 2 2 1 1 1 1 1 1 0 1 1 1 1 1 0 1 2 3 4 n Fn (m) = Fn (m 1) + Fn 1 (m 2) and Fn (0) = 1 for all n 0 − − − ≥
The above table shows an example of a recursion that does not have a poly- nomial extension. The above Theorem does not apply, because the initial points are all 0, and b1 = 2 xn 1 xn 1 = 1 (but do you recognize some numbers? See Example 3.1.8).− − − − − Going back to Pascal’s Triangle we see that the Theorem above asks us for the initial points, which we can choose as xi = 1; the factors ai are 0 except for − a1 = 1, and the translations bi can all be chosen as 0. The condition b1 = 0 ≥ xn 1 xn 1 is satisfied for all i, and p0 (m) = 1 for m 0. We (implicitly) − − − ≥ assumed pn ( 1) = 0 for n 1. The discrete integral (2.2) tells us that pn (m) = m − ≥ k=0 pn 1 (k), which means − P n + m m n 1 + k = − n n 1 k=0 X − for n 1, if we would know the solution to the recurrence already! For example, ≥ suppose we keep Pascal’s recursion, but assume that p0 (x) = 1 for all x, and n 1 pn (1 n) = − pi (n 2i) for all n 1; what is the solution in this case? The − i=0 − ≥ Theorem tells us that pn (m) is a polynomial of degree n, and we will see that this information isP extremely valuable. How to make the most of it will be the topic of the following sections. 24 Chapter 2. Finite Operator Calculus in One Variable
m 1 4 10 20 35 56 m 1 4 12 35 107 344 2 1 3 6 10 15 21 2 1 3 8 23 72 237 1 1 2 3 4 5 6 1 1 2 5 25 49 165 0 1 1 1 1 1 1 0 1 1 3 10 24 116 1 1 0 0 0 0 0 1 1 0 2 7 24 82 −2 1 -1 0 0 0 0 −2 1 -1 2 5 17 58 −3 1 -2 1 0 0 0 −3 1 -2 3 3 12 41 −4 1 -3 3 -1 0 0 −4 1 -3 5 0 9 29 − − 0 1 2 3 4 5 n 0 1 2 3 4 n Pascal’srecursion continued to Pascal’srecursion with n 1 negative integers pn (1 n) = − pi (n 2i) − i=0 − P We will begin the study of polynomials with a discussion of bases, and then use those bases to define linear operators, from polynomials to polynomials. Be- cause all our operators are linear, we will drop the word “linear” in the future. There is a special subset of these operators that maps polynomials to coeffi cients, which can be viewed as a polynomial of degree 0. Such operators are called linear functionals. Again, we drop the word “linear”. The vector space of polynomials with coeffi cients in F (a field containing Z) will be denoted by F [x], where we think of x as a formal variable. This way it is clear how F [x] is embedded in the much larger space k [[x]], the formal power series over k F. Remember that a basis must have the properties that (1) ⊇ every element p k [x] can be written in a unique way as the finite sum p (x) = n ∈ akpk (x) with coeffi cients ak F, and (2) the elements pn of the basis are k=0 ∈ linearly independent. The dimension of F [x] is infinite; the set of polynomials P pn : n = 0, 1,... can serve as a basis as long as deg pn = n, and p0 = 0. In the{ following, we} will always assume that these last two properties define6 a basis pn of F [x]. This condition requires F to be a field and not just some integral domain.{ } The polynomials solving Pascal’srecursion above just have coeffi cients in the integral domain Z; embedding them in a vector space requires F = Q. In case of an infinite dimensional space the term Hamel basis is sometimes used; we will use “basis”for finite as well as infinite bases. Note that the power series k [[t]] do not have a basis - however, it is clear that tn : n = 0, 1,... acts like a basis: (1) Every element f k [[t]] can be written in { } k ∈ a unique way as the sum f (x) = n∞=0 akt with coeffi cients ak k, and (2) the elements tn of the basis are linearly independent. The problem∈ is that the sum k P f (x) = n∞=0 akt is not finite in general. This type of “basis” is often called a pseudobasis. If (φn) is a sequence in k [[t]], where k is an integral domain, such P n that ord (φn) = n and [t ] φn is a unit in k, then (φn) is a pseudobasis (Exercise 2.1.2). 2.1. Polynomials, Operators, and Functionals 25
2.1.2 Standard Bases and Linear Operators
A basis pn : n = 0, 1,... of F [x] can also be seen as the sequence p0, p1,... . We write{(f ) or (f ) }for sequences, and if the set p is a basis, we call the n n N n n ∈ { } sequence (pn) a basis. Remember that we required deg pn = n . Every such basis (pn) uniquely determines a linear operator Q from F [x] onto itself by defining
Qpn = pn 1 − for all n 1, and Qp0 = 0. By linear extension every polynomial p gets the image ≥ n n
Qp = αkQpk = αkpk 1 − k=0 k=1 X X n if p (x) = k=0 αkpk (x). The set of such operators Q has the properties that Q reduces the degree of polynomials by 1 and Q maps constants polynomials into P 0. Thus Q has a kernel consisting of all constants (which we can identify with F), ker Q = F. We denote the set of all such operators by Ω. They are also called Gel‘fond-Leontiev operators (or generalized “difference-tial”operator in [56]).
Lemma 2.1.2. For every operator Q Ω exists a basis (pn) such that Qpn = pn 1 ∈ − and Qp0 = 0.
Proof. Let p0 (x) = 1. Because of the second property we have Qp0 = 0. From Qx = c, say, we see that c = 0 (otherwise x ker Q), and we define p1 (x) = x/c. 6 ∈ Assume that p0, p1, . . . , pn are already defined, Qpi = pi 1 for i = 1, . . . , n. The n+1 n+1 n − polynomial Qx has degree n, thus Qx = k=0 akpk, with an = 0. It follows n+1 n 1 n 1 6 that Qx /a = p + − a p /a = p + − a (Qp ) /a . Hence p = n n k=0 k k n n Pk=0 k k+1 n n+1 n+1 n 1 x − akpk+1 /an. By induction, (pn) is defined. The polynomials pn are − k=0 P P of degree n, and p0 = 0, thus (pn) is a basis. P 6 The proof of the Lemma shows a unique basis corresponding to every Q. However, there are more bases than the one constructed - every nonzero multiple, for example. It is easy to check that if (rn) and (sn) both satisfy the condition Qrn = rn 1 and Qsn = sn 1, and if deg (arn + bsn) = n for all n 0, ar0 + bs0 = − − ≥ 6 0, then (arn + bsn) is also a basis corresponding to Q. We need to “standardize” the bases to bring them into a one-to-one correspondence with Ω, by requiring specific initial values. The proof above shows that we can ask that they evaluate to 0 at 0 for all n 1, and are identically 1 for n = 0. Such a basis we call a standard basis. ≥
Lemma 2.1.3. Let Q Ω. There exists a unique standard basis (qn) such that ∈
Qqn = qn 1 for all n 1, and Qq0 = 0 − ≥ qn (0) = 0 for all n 1, and q0 (x) = 1. ≥ 26 Chapter 2. Finite Operator Calculus in One Variable
Proof. See the construction in the proof to the previous Lemma. Functionals are special operators mapping F [x] F, which can be seen as a vector space over itself, containing all polynomials of→ degree 0. For functionals L we have two notations: Lp and L p . In the special case where L stands for the n h | i n coeffi cient of x in p (x), we also write [x ] p (x) and [p]n, as in the case of formal power series.
Remark 2.1.4. If F F we can think of F [x] being embedded in F [[x]]; then x ⊆ must be a formal variable. The coeffi cient functional [p]n can be extended to F [[x]]. However, polynomials differ from formal power series in that they can be evaluated at any element a F. Such an evaluation functional we denote by Evala : p (x) ∈ 7→ p (a). Let p¯ be the polynomial p (a), a F. If F would be a finite finite field then ∈ degp ¯ can be smaller than deg p. We avoid this complication by assuming that F has charactericstic 0. Note that only when a = 0 the evaluation can be extended to F [[x]].
2.1.3 Exercises
2.1.1. Apply Theorem 2.1.1 to show that the recursion pn (x) = pn (x 1) + n 1 − pn 1 (x) for all x > 1 n with initial values pn (1 n) = i=0− pi (n 2i) for − − − − n 1, and p0 (1) = 1, has a solution that can be extended to a polynomial. ≥ P 2.1.2. Suppose F is an integral domain. Show that any sequence (φn) in F [[t]] is n a pseudobasis of F [[t]] if ord (φn) = n and [t ] φn is a unit in F.
2.1.3. Let dn (m) = dn (m 1) + dn 1 (m) dn 2 (m 2)for all m > n + 1, with − − − − − initial values at dn (n + 1) ,and d0 (m) = 1 for all m 1. Show that the solution to this problem can be extended to a polynomial sequence.≥ 2.1.4. Show that the following operators are in Ω: The forward difference opera- tor ∆ : p (x) p (x + 1) p (x), the backwards difference operator : p (x) p (x) p (x 1)7→, the division− operator χ : p (x) (p (x) p (0)) /x, and∇ the deriv-7→ − − 7→ − ative operator : p (x) p0 (x). Find the standard basis (Lemma 2.1.3) for each operator. D 7→ n a 2.1.5. Prove that [x ], Evala and 0 are functionals on F [x]. R 2.2. Finite Operators 27
2.2 Finite Operators
Take any operator that is degree reducing, deg (T xn) = n k for all n k and some fixed k 1, and T xi = 0 for all i = 0, . . . , k 1.− Special examples≥ are the Gel‘fond-Leontiev≥ operators defined in section 2.1.− The powers T,T 2,T 3,... 1 n are defined (and are degree reducing), and therefore a0 + a1T + + anT p is n ··· i defined, where T p = 0 if deg p nk. Hence the infinite sum i 0 aiT is defined ≤ ≥ on any polynomial p F [x], if ai F for all i 0. Linear operators of this form, i ∈ ∈ ≥ P i 0 aiT where T is degree reducing , are called finite operators, because on ≥ n/k i anyP polynomial of degree n they act like the finite sum ib=0 c aiT . The integral domain of all such finite operators for fixed T is denoted by ΣT . Any operator Q P in ΣT is given by T and the sequence of coeffi cients (a0, a1,... ). We saw that the formal power series are a way of “storing”such a sequence of coeffi cients. Hence we i i have a bijection between i 0 aiT and i 0 ait . The additive structure on ΣT ≥ ≥ and F [[t]] are preserved under this bijection. Is there more structure preserved? To answer this we have toP look at the compositionP of two linear operators R and Q, (RQ) p := R (Qp) i j If R and Q are both in ΣT , Q = i 0 aiT and R = j 0 bjT , then ≥ ≥ i P j i P j i R (Qp) = R ai T p = bjT ai T p = bj aiT T p i 0 j 0 i 0 j 0 i 0 X≥ X≥ X≥ X≥ X≥ i i+j i i = bj aiT p = bj ai jT p = bjai j T p − − j 0 i 0 j 0 i j i 0 j=0 X≥ X≥ X≥ X≥ X≥ X Therefore, the coeffi cients in the composition RQ are the same as in the Cauchy product in the corresponding power series. The integral domains ΣT and F [[t]] are i j isomorphic. We can write Q = α (T ) = i 0 aiT and R = β (T ) = j 0 bjT , and we have α (T ) a ti =: α (t),≥β (T ) b tj =: β (t≥). From i 0 i P j 0 j P α(t)β (t) RQ and 'α (t) β≥(t) = β (t) α (t) follows'RQ =≥ QR; any two opera- ' P P tors in ΣT commute! Three remarks about the isomorphism between formal power series and ΣT : Composition of operators corresponds to multiplication of power series. This is due to an unfortunate choice of the word “composition”for what should be called the product of operators. This gets even more confusing later, when we need the composition of power series. The second remark concerns the choice of letters. We say that ΣT F [[t]], ' but ΣT F [[s]], or any other formal variable, as long as F remains the same, because' the name of the formal variable does not matter. But it has to be a formal variable! For any degree reducing operator A we could write F [[A]] = ΣA, but we will not do this, because we believe that F [[t]] ΣA is a “cleaner”notation. The process of replacing the formal variable by A is' called evaluation. 28 Chapter 2. Finite Operator Calculus in One Variable
Example 2.2.1. Suppose the basis (pn) follows Pascal’srecursion pn (x) pn (x 1) = − − pn 1 (x). The operator : f (x) f (x) f (x 1) is called the backwards differ- ence− operator. Taylor’sTheorem∇ 7→ tells us that− f (−x 1) = ( 1)n nf (x) /n!, − n 0 − D where = d/dx. Hence = I ( 1)n n/n! = I ≥ e . We find in n 0 P −D Σ . TheD backwards differeence∇ operator− ≥ is− also degreeD reducing.− ∇ D P If we choose the coeffi cients (a0, a1,... ) from some integral domain F F, i ⊇ then the finite operators i 0 aiT may map F [x] to some different space. How- ≥ i ever, if S is a degree reducing operator on F [x] and F = ΣS, then i 0 AiT , P ≥ with Ai ΣS, maps F [x] into itself. Such operator coeffi cients will not give us any new operators,∈ but they simplify the description of operators. For example,P we can see the operator as an element of Σ with coeffi cients (0, 0, 1, 1/2!, 1/3!,... ) ∇D D − in F, Σ F [[t]] as before. But we can also see as an element of Σ with coeffi cientsD '(0, , 0,... ) in Σ , Σ Σ [[t]], or as∇D an element of Σ withD coeffi - cients (0, , 0,...∇ ) in Σ , Σ ∇ ΣD '[[t]].∇ We will do this in section 2.4∇ on transfer D D ∇ ' D theorems, but it will be clearly stated. In general, the coeffi cients in ΣS will come from the same field F as in F [x], the polynomials.
2.2.1 Translation Operators Besides the degree reducing operators there is another important set of operators, the translation operators Ea : p (x) p (x + a). The translation operators are 7→ defined for all a F, and they form a group, EaEb = Ea+b. They are degree ∈ preserving. Because x is a formal variable, and a F, we should say what we n ∈n n n k n k mean by the polynomial (x + a) . Of course, (x + a) := k=0 k a x − . Note that on the basis (xn) we have P n n k a n n n k n k k n E x = (x + a) = a x − = a D x (2.3) k k! k=0 k=0 X X where k = dk/dxk is the k-th power of the derivative operator. Hence D k a k a E = a D = e D k! k 0 X≥ This shows that Ea Σ , and it follows that Ea commutes with every other ∈ D operator in Σ . We call a linear operator T on F [x] translation invariant, if a aD TE p (x) = E T p (x) for all a F and p F [x]. The operators in ΣD are exactly the translation invariant operators,∈ as we∈ show in the following Lemma (the “First Expansion Theorem”in [83, p. 691]).
Lemma 2.2.2. A linear operator T on F [x] is translation invariant iff T Σ . In i i ∈ D that case, T = i 0 Eval0 T x /i! ≥ | D P 2.2. Finite Operators 29
Proof. All we have left to show is that every operator T is in Σ if it satisfies a n a n n i D TE x = E T x for all n 0 and a F. Let T x = i 0 cn,ix , where for given ≥ ∈ ≥ n the ring elements cn,i are eventually 0 for large enough i. Thus P
i i i + k EaT xn = Ea c xi = c ai kxk = xk aic n,i n,i k − k n,i+k i 0 i 0 k=0 k 0 i 0 X≥ X≥ X X≥ X≥ and
n n n a n n n i i n n i k k n i TE x = T a − x = a − ci,kx = x a cn i,k i i i − i=0 i=0 k 0 k 0 i=0 X X X≥ X≥ X Both expressions are equal iff the coeffi cients of xk agree,
n i + k i n i a cn,i+k = a cn i,k k i − i 0 i=0 X≥ X for all k 0 and for all a F. Both sides are a polynomial in a F. The right hand side is of≥ degree at most∈n, hence the left hand side is likewise,∈ and we conclude that 0 i n also holds on the left side. The coeffi cients of ai must be equal for all 0 ≤i ≤n, hence ≤ ≤
(i + k)!cn,i+k/n! = k!cn i,k/ (n i)! for all 0 i n and 0 k n i. − − ≤ ≤ ≤ ≤ −
If k = 0, then i!cn,i/n! = cn i,0/ (n i)!,thus − − n n n i ci,0 n! n i ci,0 i n T x = cn,ix = x − = x . i! (n i)! i! D i=0 i=0 i 0 X X − X≥ and therefore T Σ . ∈ D Remark 2.2.3. The degree-by-one reducing operators in Ω are of general interest. Let a (x) be a polynomial and define the multiplication operator M (a): p (x) 7→ a (x) p (x) for all p (x) F [x]. If a1 (x) and a2 (x) define M (a1 + a2) = M (a1) + ∈ M (a2), and M (a1a2) = M (a1) M (a2) = M (a2) M (a1). For any linear oper- ator T on F [x] there exists a sequence of polynomials (an (x)) such that T = n n 0 M (an) R for some given R Ω (the polynomials an (x) are not neces- sarily≥ of deg n). See Exercise 2.2.2.∈ We saw in Lemma 2.2.2 that for R = the Ptranslation invariant operators are obtained by the choice of constant polynomialsD n cn (x) = Eval0 T x /n!. This and most of the other results in this section were already knownh at| thei end of the 19th century (see Pincherle [74]). 30 Chapter 2. Finite Operator Calculus in One Variable
2.2.2 Basic Sequences and Delta Operators Even the set of standard bases is too large for our purposes. We will focus on a “tiny” subset, often called the sequences of binomial type. These sequences (bn) follow the binomial theorem (see also section 2.3),
n
bn (x + y) = bi (y) bn i (x) − i=0 X x for all n 0. Examples are bn (x) = n (Vandermonde convolution) and, if Q F, ≥ n ⊂ bn (x) = x /n! (giving the original binomial theorem). Standard sequences that are of binomial type are called basic sequences ; not a great name, but from here on, the only bases in F [x] we will consider in this section, are basic sequences. The binomial theorem is a convolution identity. It says that for the generating n function b (x, t) = n 0 bn (x) t holds b (x + y, t) = b (x, t) b (y, t), where b (0, t) = b (x, 0) = 1. If b (x, t)≥is of the form f (t)x, where f (0) = 1, then the binomial P theorem holds. In this case, f (t) = eβ(t) for some delta series β. In other words, the logarithm of f (t) = 1+ higher order terms in t exists, and log f (t) = β (t). We will investigate that approach in this chapter and the next. 1 Because β (t) is a delta series, the compositional inverse β− (t) of β (t) exists and is also a delta series (section 1.2). If γ (t) is any delta series, we call the operator 1 γ ( ) a delta operator, hence β− ( ) and β ( ) are both delta operators. We will D 1 D D study the delta operator β− ( ) and and see what it does to (bn). The meaning of β ( ) has to wait until sectionD 2.3.2. D
Transforms of Operators on F [x] [[t]] The investigation follows the ideas of J. M. Freeman [35] in his “Transform of Operators”, opening the door to a unify- ing theory beyond the Finite Operator Calculus. First we note that b (x, t) is in F [x] [[t]], the formal power series with coeffi cients in the ring of polynomials F [x]. We worked already with a power series in F [x] [[t]] in Example 1.2.3. An element from F [x] [[t]] can be understood as an infinite matrix, whose rows stand for the coeffi cients of t, the n-th row containing the coeffi cients of a polynomial. The matrix is triangular, if the polynomials are a basis for F [x]. An operator A on F [x] is extended to an operator on F [x] [[t]] by defin- i i ing A i 0 pi (x) t := i 0 (Api (x)) t . Such a t-linear extension is called an ≥ ≥ x-operator on F [x] [[t]]. For example, ext Q [x] [[t]], and P P ∈ next = tnext. (2.4) D In the same way, there are t-operators on F [x] [[t]], extended by x-linearity. Substituting a delta series α (t) is a linear operator on F [[t]] (see again section 1.2), and by x-linearity C(α)f (x, t) = f (x, α (t)) for all f (x, t) F [x] [[t]]. Note that x-operators commute with t-operators! ∈ 1 xβ(t) Now we are ready to see what the x-operator β− ( ) does to b (x, t) = e , D 2.2. Finite Operators 31
1 1 xβ(t) 1 xt 1 xt β− ( ) b (x, t) = β− ( ) e = β− ( ) C(β)e = C(β)β− ( ) e D D D D 1 n because x- and t-operators commute. Suppose β− ( ) = n 1 γn . Then D ≥ D 1 xt n xt 1 xtP β− ( ) e = γnt e = β− (t) e , D n 1 X≥ 1 xt as in (2.4), showing that the action of the x-operator β− ( ) on e is the same 1 D 1 as multiplication by the formal power series β− (t). We say that β− ( ) and 1 xt D M β− are transforms of each other, with respect to e . Hence