arXiv:0911.5278v1 [math-ph] 27 Nov 2009 aclto fresultant of Calculation 3 Objects Basic 2 Introduction 1 Contents practical be resul could these which of resultants, overview for brief formulas a explicit present We lo field. a this in studied: known well rather course of are resultants n research, with or forms non-quadratic with equations, (polynomial) Conclusion 6 applications and discriminants Integral 5 discriminant of Calculation 4 ∗ † . omlso yvse ye...... type Sylvester of Formulas 3.1 eutnsaegtigicesnl motn nmdr th modern in important increasingly getting are Resultants . vlaino ra ...... series . . . hypergeometric . as . . . . . numbers . . . Algebraic ...... areas . 5.4 . . of . . . Evaluation ...... 5.3 ...... Action-independence . identities . . . . Ward . . and . . . 5.2 . integrals . . . . Non-Gaussian ...... 5.1 ...... case . . . . . symmetric . . . The . . . . polytopes . . Newton . . 4.3 . . through . . Discriminant ...... invariants 4.2 . . through . Discriminant ...... 4.1 ...... type . Schur of . type Formulas Bezout . + . Sylvester 3.4 of Formulas type Bezout 3.3 of Formulas 3.2 TP ocw Russia Moscow, ITEP, [email protected] Russia; Moscow, ITEP, .. aalysfrua...... complex formula Koszul . Macaulay’s of . determinant a matrix 3.1.3 as Sylvester Resultant of determinant a 3.1.2 as Resultant 3.1.1 e n l eut nRslatTheory Resultant in Results Old and New and IT ogpun,Russia Dolgoprudny, MIPT, A.Morozov yueu naftr hsc research. physics future a in useful ly fepii omls euiu rprisadintriguin and properties beautiful formulas, explicit of t ABSTRACT nGusa nerl.Bigasbeto oeta three-h than more of subject a Being integrals. on-Gaussian ∗ s nldn ohrcn n led lsia.Epai i Emphasis classical. already and recent both including ts, n Sh.Shakirov and oeia hsc:te perweee n el ihnon- with deals one whenever appear they physics: eoretical 1 [email protected] ; 42 ...... 36 ...... † 6 ...... 33 ...... 27 ...... 8 ...... 19 ...... 6 ...... 38 ...... 39 ...... 22 . . . . . 15 ...... 33 ...... 14 ...... ITEP/TH-68/09 eainhp are relationships g undred-year aeon made s linear 48 36 27 6 4 2 1 Introduction
This paper is devoted to resultants – natural and far-going generalization of determinants from conventional linear to generic non-linear systems of algebraic equations
f1(x1,...,xn)=0 f2(x1,...,xn)=0 . . . f (x ,...,x )=0 n 1 n 1 where f1,...,fn are homogeneous polynomials of degrees r1,...,rn. This system of n polynomial equations in n variables is over-defined: in general position it does not have non-zero solutions (x1,...,xn) at all. A non-vanishing solution exists if and only if a single algebraic constraint is satisfied,
R f1,f2,...,fn =0 (1) where R is an irreducible polynomial function called resultant, depending on coefficients of the non-linear system under consideration. Clearly, for linear equations resultant is nothing but the determinant – a familiar object of linear algebra. For non-linear equations, resultant provides a natural generalization of the determinant and therefore can be considered as a central object of non-linear algebra [1]. Historically, the foundations of resultant theory were laid in the 19th century by Cayley, Sylvester and Bezout [2]. In this period mostly the case of n = 2 variables was considered, with some rare exceptions. The theory of resultants for n > 2 and closely related objects, namely discriminants and hyperdeterminants, was significantly developed by Gelfand, Kapranov and Zelevinsky in their seminal book [3]. Let us emphasize, that determinant has a lot of applications in physics. Resultant is expected to have at least as many, though only a few are established yet. This is partly because today’s physics is dominated by methods and ideas of linear algebra. These methods are not always successful in dealing with problems, which require an essentially non-linear treatment. We expect that resultants and related algebraic objects will cure this problem and become an important part of the future non-linear physics and for entire string theory program in the sence of [4]. To illustrate this idea, let us briefly describe the relevance of resultants to non-Gaussian integrals – a topic of undisputable value in physics. It is commonly accepted, that quantum physics can be formulated in the language of functional integrals. It is hard to question the importance of this approach: since the early works of Wigner, Dirac and Feynman, functional integrals have been a source of new insights and puzzles. Originated in quantum mechanics, they have spread fastly in quantum field theory and finally developed into a widely used tool of modern theoretical physics. From the point of view of functional integration, any quantum theory deals with integrals of a form
f(x) = dx f(x) e−S(x) (2) 1In this paper we assume homogeneity of the polynomial equations to simplify our considerations; actually, the homogeneity condition can be relaxed and the whole theory can be formulated for non-homogeneous polynomials as well.
2 called correlators. The integration variable x belongs to a linear space (which is often infinite-dimensional), and the choice of particular action S(x) specifies a particular theory or a model. In the simplest models, S(x) is merely a quadratic form, which in some particular basis can be written as
S(x)= Sij xixj (3) i,j
If this is the case, the integral (2) is called Gaussian. In a sence, Gaussian integrals constitute a foundation of quantum theory: they describe free (non-interacting) particles and fields. These systems are relatively simple, and so are the Gaussian integrals. For example, the simplest two-dimensional Gaussian integral can be evaluated, say, by diagonalisation, and equals
2 2 const dx dy e−(ax +bxy+cy ) = (4) √b2 4ac − where the expression b2 4ac can be recognized as a determinant of the quadratic form ax2 + bxy + cy2. Other − Gaussian integrals, with more variables, with non-homogeneous action or with non-trivial insertion f(φ), are equally easy and similarly related to determinants – central objects of linear algebra, which describe consistency of linear systems of equations. Remarkably, the study of low-dimensional non-Gaussian integrals of degree three
S(x)= Sijkxixj xk (5) i,j,k and higher, reveals their relation to resultants: for example, a two-dimensional non-Gaussian integral equals
3 2 2 3 const dx dy e−(ax +bx y+cxy +dy ) = (6) √6 27a2d2 b2c2 18abcd +4ac3 +4b3d − − where the expression 27a2d2 b2c2 18abcd +4ac3 +4b3d can be recognized as a discriminant of the cubic − − form ax3 + bx2y + cxy2 + dy3, given by resultant of its derivatives:
R 3ax2 +2bxy + cy2, bx2 +2cxy +3dy2 = 27a2d2 b2c2 18abcd +4ac3 +4b3d (7) − −
In this way, resultants naturally appear in the study of low-dimensional non-Gaussian integrals. This connection between functional integration and non-linear algebra can be traced much further: section 5 of present paper provides a brief summary of this interesting direction of research. See also [5] for more details. In this paper we present a selection of topics in resultant theory in a more-or-less explicit manner, which is more convenient for a reader who needs to perform a practical calculation. This also makes our paper accessible to non-experts in the field. Most of statements are explained rather than proved, for some proofs see [3].
3 2 Basic Objects Homogeneous polynomial
Basic objects of study in non-linear algebra are generic finite-dimensional tensors [1]. However, especially important in applications are symmetric tensors or, what is the same, homogeneous polynomials of degree r in n variables. In present paper, we restrict consideration to this case. It should be emphasized, that restriction to symmetric tensors is by no means an oversimplification: most of the essential features of non-linear algebra are present at this level. There are two different notations for homogeneous polynomials, which can be useful under different circumstances: tensor notation
n
S(x1, x2,...,xn)= Si1,i2,...,ir xi1 xi2 ...xir i ,i ,...,i =1 1 2 r and monomial notation
a1 a2 an S(x1, x2,...,xn)= sa1,a2...an x1 x2 ...xn a +a +...+a =r 1 2 n To distinguish between these notations, we denote coefficients by capital and small letters, respectively. A r homogeneous polynomial of degree r in n variables has Cn+r−1 independent coefficients. Resultant
A system of n polynomial equations, homogeneous of degrees r1,...,rn in variables x1,...,xn
f1(x1,...,xn)=0 f2(x1,...,xn)=0 . . .
fn(x1,...,xn)=0 has non-zero solutions, if and only if its coefficients satisfy one algebraic constraint:
Rr ,...,r f ,...,fn =0 1 n { 1 }
The left hand side of this algebraic constraint is called a resultant of this system of equations. Resultant always exists and is unique up to rescaling. Resultant is a homogeneous polynomial of degree
1 1 deg R = r ...r + . . . + r1,...,rn 1 n r r 1 n in the coefficients of the system. Moreover, it is homogeneous in coefficients of each equation fi separately, and has degree r1 ...ri−1ri+1 ...rn in these coefficients. When all equations are linear, i.e, when ri = 1, resultant is equal to the determinant of this system: R ... f ,f ,...,fn = f f . . . fn. Resultant is defined 11 1{ 1 2 } 1 ∧ 2 ∧ ∧ only when the number of homogeneous equations matches the number of variables: when these numbers differ, various more complicated quantities should be used to describe the existence and/or degeneracy of solutions. Some of these quantities are known as higher resultants or subresultants [6], some as complanarts [7].
4 Discriminant
A single homogeneous polynomial S(x1,...,xn) of degree r in n variables gives rise to a system of equations
∂S(x ,...,x ) 1 n =0 ∂x1 . . . ∂S(x ,...,x ) 1 n =0 ∂x n which are all homogeneous of degree r 1 in variables x ,...,xn. Resultant of this system is called a discriminant − 1 of S and is denoted as Dn|r(S). Discriminant is a homogeneous polynomial of degree
deg D (S)= n(r 1)n−1 n|r − in the coefficients of S, what follows from the analogous formula for the generic resultant. When S is a quadratic form, i.e, when r = 2, discriminant is equal to its determinant: Dn|2(S) = det S. Invariant
A Lie group of linear transformations
x1 G11 G12 ... G1n x1
. . . . . . . . . → xn Gn1 G12 ... Gnn xn with det G = 0 is called GL(n). Its subgroup with det G = 1 is called SL(n). In tensor notations, they act
xi Gij xj →
Thus they act on coefficients of homogeneous polynomials S(x1,...,xn) of degree r in n variables by the rule
Si ,...,i Gi j ...Gi j Sj ,...,j 1 r → 1 1 r r 1 r
Polynomials I(S) which map to theirselves under SL(n) transformations are called SL(n) invariants of S. In particular, discriminant Dn|r(S) is a SL(n) invariant of S. However, there are many more invariants which have no obvious relation to the discriminant. The number of functionally independent invariants of S is equal to the number of independent coefficients of S minus the dimension of SL(n), that is
# functionally independent invariants of S = Cr n2 +1 n+r−1 −
Therefore, invariants of given degree form a finite-dimensional linear space. All invariants form a ring.
5 3 Calculation of resultant 3.1 Formulas of Sylvester type
To find the resultant of n polynomial equations, homogeneous of degree r in variables x1,...,xn
f1(x1,...,xn)=0 f2(x1,...,xn)=0 . . . fn(x1,...,xn)=0 it is possible to consider another system, multiplied by various monomials:
s1 sn consider x1 ...xn fi = 0 instead of just fi =0
In the monomial basis, coefficients of this system can be viewed as a rectangular nCq Cr+q matrix, q+n−1 × r+q+n−1 where q = s1 + . . . + sn is the multiplier’s degree. This matrix is called the (generalized) Sylvester matrix.
3.1.1 Resultant as a determinant of Sylvester matrix
q r+q In some cases it is possible to choose q in such a way, that Sylvester matrix is square, i.e, nCq+n−1 = Cr+q+n−1. In these cases, resultant is equal to the determinant of Sylvester matrix. Looking closer at this numeric relation, it is easy to see that the only cases when it is possible are n = any, r = 1 and n = 2, r = any. In the first case, it suffices to choose q = 0 and Sylvester matrix equals the matrix of the system. This is the case of linear algebra. The second case is less trivial: it is necessary to choose q = r 1 to make the Sylvester matrix square. − For example, in the case n =2, r = 2 the system of equations has a form
2 2 f11x1 + f12x1x2 + f22x2 =0
2 2 g11x1 + g12x1x2 + g22x2 =0 Multiplying the equations of this system by all monomials of degree q = 1, that is, by x1 and x2, we obtain
3 2 2 3 f11x1 + f12x1x2 + f22x1x2 +0x2 =0
3 2 2 3 0x1 + f11x1x2 + f12x1x2 + f22x2 =0 3 2 2 3 g11x1 + g12x1x2 + g22x1x2 +0x2 =0 3 2 2 3 0x1 + g11x1x2 + g12x1x2 + g22x2 =0 These four polynomials are linear combinations of four monomials, i.e, give rise to a square Sylvester matrix:
f11 f12 f22 0
0 f11 f12 f22 R2|2 f,g = det (8) { } g11 g12 g22 0 0 g g g 11 12 22
6 Formulas, which express resultant as a determinant of some auxillary matrix, are usually called determinantal. In the next case n =2, r = 3 the system of equations has a form
3 2 2 3 f111x1 + f112x1x2 + f122x1x2 + f222x2 =0
3 2 2 3 g111x1 + g112x1x2 + g122x1x2 + g222x2 =0
2 2 Multiplying the equations of this system by all monomials of degree q = 2, that is, by x1, x1x2 and x2, we get
5 4 3 2 2 3 4 5 f111x1 + f112x1x2 + f122x1x2 + f222x1x2 +0x1x2 +0x2 =0
5 4 3 2 2 3 4 5 0x1 + f111x1x2 + f112x1x2 + f122x1x2 + f222x1x2 +0x2 =0 5 4 3 2 2 3 4 5 0x1 +0x1x2 + f111x1x2 + f112x1x2 + f122x1x2 + f222x2 =0 5 4 3 2 2 3 4 5 g111x1 + g112x1x2 + g122x1x2 + g222x1x2 +0x1x2 +0x2 =0 5 4 3 2 2 3 4 5 0x1 + g111x1x2 + g112x1x2 + g122x1x2 + g222x1x2 +0x2 =0 5 4 3 2 2 3 4 5 0x1 +0x1x2 + g111x1x2 + g112x1x2 + g122x1x2 + g222x2 =0 These four polynomials are linear combinations of four monomials, i.e, give rise to a square Sylvester matrix:
f111 f112 f122 f222 0 0
0 f111 f112 f122 f222 0
0 0 f111 f112 f122 f222 R2|3 f,g = det (9) { } g111 g112 g122 g222 0 0 0 g111 g112 g122 g222 0 0 0 g111 g112 g122 g222 Note, that the matrix has a typical ladder form. The beauty and simplicity made this formula and corresponding resultant R2|r widely known. In our days, Sylvester method is the main apparatus used to calculate two- dimensional resultants and it is the only piece of resultant theory included in undergraduate text books and computer programs such as Mathematica and Maple. However, its generalization to n> 2 is not straightforward, since for n> 2 and r 2 Sylvester matrix is not square. ≥ It turns out that, despite Sylvester matrix in higher dimensions is rectangular, it is still related closely to the resultant. Namely, top-dimensional minors of the Sylvester matrix are always divisible by the resultant, if their size is bigger, than resultant’s degree (i.e, for big enough q). Various generalizations of the Sylvester method explicitly describe the other, non-resultantal, factor. Therefore, such generalized Sylvester formulas always express the resultant as a ratio of two polynomial quantities. This is certainly a drawback, at least from the computational point of view. Sylvester’s method posesses at least two different generalizations for n > 2: one homological, which leads to the theory of so-called Koszul complexes, another more combinatorial, which leads to the construction of so-called Macaulay matrices.
7 3.1.2 Resultant as a determinant of Koszul complex
In construction of the Koszul complex, one complements the commuting variables
x , x ,...,xn : xixj xj xi =0 1 2 − with anti-commuting variables
θ1,θ2,...,θn : θiθj + θj θi =0
and considers polynomials, depending both on x1,...,xn and θ1,...,θn. Denote through Ω(p, q) the space of such polynomials of degree p in x-variables and degree q in θ-variables. The degree q can not be greater than n, since θ are anti-commuting variables. Dimensions of these spaces are
p q dim Ω(p, q)= Cp+n−1Cn (10)
Koszul differential, built from f, is a linear operator which acts on these spaces by the rule
∂ ∂ dˆ= f1(x1,...,xn) + . . . + fn(x1,...,xn) (11) ∂θ1 ∂θn and is automatically nilpotent:
∂ ∂ ∂ ∂ ∂ ∂ dˆdˆ= f (x)f (x) = f (x)f (x) + =0 j k ∂θ ∂θ j k ∂θ ∂θ ∂θ ∂θ j k j k k j
It sends
dˆ: Ω(p, q) Ω(p + r, q 1) → − giving rise to the following Koszul complex:
dˆ dˆ dˆ dˆ Ω(p, q) Ω(p + r, q 1) Ω(p +2r, q 2) ... Ω(0, R) → − → − → →
Thus, for one and the same operator dˆthere are many different Koszul complexes, depending on a single integer parameter R – the x-degree of the rightmost space. Using the formula (10) for dimensions, it is easy to write down all of them. For example, the tower of Koszul complexes for the case 3 2 is |
8 R Spaces Dimensions Euler characteristic 0 Ω(0, 0) 1 1 1 Ω(1, 0) 3 3 2 Ω(0, 1) Ω(2, 0) 3 6 3 3 Ω(1, 1) → Ω(3, 0) 9 →10 1 3 2 → → | 4 Ω(0, 2) Ω(2, 1) Ω(4, 0) 3 18 15 0 5 Ω(1, 2) → Ω(3, 1) → Ω(5, 0) 9 → 30 → 21 0 6 Ω(0, 3) Ω(2→, 2) Ω(4→, 1) Ω(6, 0) 1 18→ 45→ 28 0 7 Ω(1, 3) → Ω(3, 2) → Ω(5, 1) → Ω(7, 0) 3 → 30 → 63 → 36 0 ...→ ...→ → → → ...→ ... where Euler characteristic is the alternating combination of dimensions. The rightmost differential
dˆ dˆ: Ω(R r, 1) Ω(R, 0) − → acts on the corresponding linear spaces as
i1 i2 in i1 i2 in x x ...x θj x x ...x fj x ,...,xn 1 2 n → 1 2 n 1 and is represented by nCR−r CR matrix. It is easy to see, that this is exactly the rectangular R−r+n−1 × R+n−1 Sylvester matrix. Therefore, Koszul complex in a sence complements the Sylvester matrix with many other matrices, which carry additional information. Resultant is equal to certain characteristic of this complex
R f ,...,fn = DET (d , d ,...,dp ) (12) n|r{ 1 } 1 2 −1
which is also called determinant, not to be confused with ordinary determinant of a linear map. Here di are the rectangular matrices, representing particular terms of the complex. The ”determinantal” terminology was introduced by Cayley, who first studied DET (d1, d2,...,dp−1), and is rather unusual for modern literature, where DET (d1, d2,...,dp−1) is called a (Reidemeister) torsion of the complex d1,...,dp−1. Since the quantity itself is a rather natural generalisation of determinants from linear maps to complexes, we prefer to use the historical name ”determinant of a complex”. Calculation of determinants of complexes is described in some detail in [3, 8]. Given a complex
d1 d2 dp−1 L L ... Lp (13) 1 −→ 2 −→ −→ with vanishing Euler characteristic
p χ = dim L dim L + dim L . . . + ( 1) dim Lp =0 1 − 2 3 − −
one needs to select a basis in each linear space Li and label the basis vectors by elements of a set
9 1, 2,...,li
After the basis is chosen, the complex is represented by a sequence of rectangular matrices. The li li matrix × +1 di has li rows and li+1 coloums. To calculate the determinant of this complex, one needs to select arbitrary subsets σi 1, 2,...,li which consist of ⊂ i−1 i+k+1 σi = ( 1) lk − k=1 elements, i.e,
σ =0, σ = l , σ = l l , σ = l l + l , ... 1 2 1 3 2 − 1 4 3 − 2 1 and define conjugate subsets σi = 1, 2,...,li /σi which consist of σi = li σi = σi − +1 elements. Finally, one needs to calculate minors
Mi = the minor of di, which occupies rows σi and coloumns σi+1