Computing Syzygies in Finite Dimension Using Fast Linear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

Computing Syzygies in Finite Dimension Using Fast Linear Algebra Computing syzygies in finite dimension using fast linear algebra Vincent Neiger Univ. Limoges, CNRS, XLIM, UMR 7252, F-87000 Limoges, France Eric´ Schost University of Waterloo, Waterloo ON, Canada Abstract We consider the computation of syzygies of multivariate polynomials in a finite-dimensional setting: for a K[X1;:::; Xr]-module M of finite dimension D as a K-vector space, and given elements f1;:::; fm in M, the problem is to compute syzygies between the fi’s, that is, poly- m nomials (p1;:::; pm) in K[X1;:::; Xr] such that p1 f1 + ··· + pm fm = 0 in M. Assuming that the multiplication matrices of the r variables with respect to some basis of M are known, we give an algorithm which computes the reduced Grobner¨ basis of the module of these syzygies, − for any monomial order, using O(mD! 1 + rD! log(D)) operations in the base field K, where ! is the exponent of matrix multiplication. Furthermore, assuming that M is itself given as n M = K[X1;:::; Xr] =N, under some assumptions on N we show that these multiplication matri- ces can be computed from a Grobner¨ basis of N within the same complexity bound. In particular, taking n = 1, m = 1 and f1 = 1 in M, this yields a change of monomial order algorithm along the lines of the FGLM algorithm with a complexity bound which is sub-cubic in D. Keywords: Grobner¨ basis, syzygies, complexity, fast linear algebra. 1. Introduction In what follows, K is a field and we consider the polynomial ring K[X] = K[X1;:::; Xr]. The set of m × n matrices over a ring R is denoted by Rm×n; when orientation matters, a vector in Rn is considered as being in R1×n (row vector) or in Rn×1 (column vector). We are interested in the efficient computation of relations, known as syzygies, between elements of a K[X]-module M. Let us write the K[X]-action on M as (p; f ) 2 K[X] × M 7! p · f , and let f1;:::; fm be in M. Then, for a given monomial order ≺ on K[X]m, we want to compute the Grobner¨ basis of the kernel of the homomorphism arXiv:1912.01848v2 [cs.SC] 19 Jun 2020 K[X]m !M (p1;:::; pm) 7! p1 · f1 + ··· + pm · fm: This kernel is called the module of syzygies of ( f1;:::; fm) and written SyzM( f1;:::; fm). In this paper, we focus on the case where M has finite dimension D as a K-vector space; as a m result, the quotient K[X] = SyzM( f1;:::; fm) has dimension at most D as a K-vector space. Then one may adopt a linear algebra viewpoint detailed in the next paragraph, where the elements of M are seen as row vectors of length D and the multiplication by the variables is represented by so-called multiplication matrices. This representation was used and studied in [2, 37,1, 27], mainly in the context where M is a quotient K[X]=I for some ideal I (thus zero-dimensional of degree D) and more generally a quotient K[X]n=N for some submodule N ⊆ K[X]n with n 2 N>0 (see [1, Sec. 4.4 and 6]). This representation with multiplication matrices allows one to perform computations in such a quotient via linear algebra operations. Assume we are given a K-vector space basis F of M. For i in f1;:::; rg, the matrix of the structure morphism f 7! Xi · f with respect to this basis is denoted by Mi; this means that for f in 1×D M represented by the vector f 2 K of its coefficients on F , the coefficients of Xi · f 2 M on F are f Mi. We call M1;:::; Mr multiplication matrices; note that they are pairwise commuting. The data formed by these matrices defines the module M up to isomorphism; we use it as a × representation of M. For p in K[X] and for f in M represented by the vector f 2 K1 D of its coefficients on F , the coefficients of p · f 2 M on F are f p(M1;:::; Mr); hereafter this vector is written p · f. From this point of view, syzygy modules can be described as follows. Definition 1.1. For m and D in N>0, let M = (M1;:::; Mr) be pairwise commuting matrices in D×D m×D m K , and let F 2 K . Denoting by f1;:::; fm the rows of F, for p = (p1;:::; pm) 2 K[X] we write 1×D p · F = p1 · f1 + ··· + pm · fm = f1 p1(M) + ··· + fm pm(M) 2 K : The syzygy module SyzM(F), whose elements are called syzygies for F, is defined as m SyzM(F) = fp 2 K[X] j p · F = 0g; m as noted above, K[X] = SyzM(F) has dimension at most D as a K-vector space. In particular, if in the above context F is the matrix of the coefficients of f1;:::; fm 2 M on the basis F , then SyzM(F) = SyzM( f1;:::; fm). Our main goal in this paper is to give a fast algorithm to solve the following problem (for the notions of monomial orders and Grobner¨ basis for modules, we refer to [13] and the overview in Section2). Problem 1 – Gr¨obnerbasis of syzygies Input: • a monomial order ≺ on K[X]m, D×D • pairwise commuting matrices M = (M1;:::; Mr) in K , × • a matrix F 2 Km D. Output: the reduced ≺-Grobner¨ basis of SyzM(F). Example 1.2. An important class of examples has r = 1; in this case, we are working with D univariate polynomials. Restricting further, consider the case M = K[X1]=hX1 i endowed with the canonical K[X1]-module structure; then M1 is the upper shift matrix, whose entries are all 0 m except those on the superdiagonal which are 1. Given f1;:::; fm in M,(p1;:::; pm) 2 K[X1] is a D syzygy for f1;:::; fm if p1 f1 +···+pm fm = 0 mod X1 . Such syzygies are known as Hermite-Pad´e D approximants of ( f1;:::; fm)[22, 40]. Using moduli other than X1 leads one to generalizations such as M-Pad´eapproximants or rational interpolants (corresponding to a modulus that splits into linear factors) [33,4, 49]. For r = 1, SyzM(F) is known to be free of rank m. Bases of such K[X1]-modules are often described by means of their so-called Popov form [43, 26]. In commutative algebra terms, this 2 is a term over position Grobner¨ basis. Another common choice is the Hermite form, which is a position over term Grobner¨ basis [29]. Example 1.3. For arbitrary r, let I be a zero-dimensional ideal in K[X] and let M = K[X]=I with the canonical K[X]-module structure. Then, taking m = 1 and f1 = 1 2 M, we have SyzM( f1) = fp 2 K[X] j p f1 = 0g = fp 2 K[X] j p 2 Ig = I: Suppose we know a Grobner¨ basis of I for some monomial order ≺1, together with the cor- responding monomial basis of M, and the multiplication matrices of X1;:::; Xr in M. Then solving Problem1 amounts to computing the Gr obner¨ basis of I for the new order ≺. More generally for a given f1 = f 2 M, the case m = 1 corresponds to the computation of f g the annihilator of f in K[X], often denoted by AnnK[X]( f ). Indeed, the latter set is defined as fp 2 K[X] j p f = 0g, which is precisely SyzM( f ). r Example 1.4. For an arbitrary r, let α1;:::; αD be pairwise distinct points in K , with αi = (αi;1; : : : ; αi;r) for all i. Let I be the vanishing ideal of fα1;:::; αDg and M = K[X]=I. As above, take m = 1 and f1 = 1, so that SyzM(1) = I. The Chinese Remainder Theorem gives an explicit isomorphism M! KD that amounts to D evaluation at α1;:::; αD. The multiplication matrices induced by this module structure on K are 1×D diagonal, with M j having diagonal (α1; j; : : : ; αD; j) for 1 6 j 6 r. Taking F = [1 ··· 1] 2 K , solving Problem1 allows us to compute the Gr obner¨ basis of the vanishing ideal I for any given order ≺. This problem was introduced by Moller¨ and Buchberger [36]; it may be extended to cases where vanishing multiplicities are prescribed [35]. Example 1.5. Now we consider an extension of the Moller-Buchberger¨ problem due to Kehrein, Kreuzer and Robbiano [27]. Given r pairwise commuting d × d matrices N1;:::; Nr, we look for their ideal of syzygies, that is, the ideal I of polynomials p 2 K[X] such that p(N1;:::; Nr) = 0. When r = 1, this ideal is generated by the minimal polynomial of N1. × One may see this problem in our framework by considering M = Kd d endowed with the d×d K[X]-module structure given by Xk · A = ANk for all 1 6 k 6 r and A 2 K . The ideal I defined above is the module of syzygies SyzM( f ) of the identity matrix f = Id 2 M, so we have m = d and D = d2 here. To form the input of Problem1, we choose as a basis of M the list of d×d elementary matrices F = (c1;1;:::; c1;d;:::; cd;1;:::; cd;d) where ci; j is the matrix in K whose only nonzero entry is a 1 at index (i; j).
Recommended publications
  • 10 7 General Vector Spaces Proposition & Definition 7.14
    10 7 General vector spaces Proposition & Definition 7.14. Monomial basis of n(R) P Letn N 0. The particular polynomialsm 0,m 1,...,m n n(R) defined by ∈ ∈P n 1 n m0(x) = 1,m 1(x) = x, . ,m n 1(x) =x − ,m n(x) =x for allx R − ∈ are called monomials. The family =(m 0,m 1,...,m n) forms a basis of n(R) B P and is called the monomial basis. Hence dim( n(R)) =n+1. P Corollary 7.15. The method of equating the coefficients Letp andq be two real polynomials with degreen N, which means ∈ n n p(x) =a nx +...+a 1x+a 0 andq(x) =b nx +...+b 1x+b 0 for some coefficientsa n, . , a1, a0, bn, . , b1, b0 R. ∈ If we have the equalityp=q, which means n n anx +...+a 1x+a 0 =b nx +...+b 1x+b 0, (7.3) for allx R, then we can concludea n =b n, . , a1 =b 1 anda 0 =b 0. ∈ VL18 ↓ Remark: Since dim( n(R)) =n+1 and we have the inclusions P 0(R) 1(R) 2(R) (R) (R), P ⊂P ⊂P ⊂···⊂P ⊂F we conclude that dim( (R)) and dim( (R)) cannot befinite natural numbers. Sym- P F bolically, we write dim( (R)) = in such a case. P ∞ 7.4 Coordinates with respect to a basis 11 7.4 Coordinates with respect to a basis 7.4.1 Basis implies coordinates Again, we deal with the caseF=R andF=C simultaneously. Therefore, letV be an F-vector space with the two operations+ and .
    [Show full text]
  • Selecting a Monomial Basis for Sums of Squares Programming Over a Quotient Ring
    Selecting a Monomial Basis for Sums of Squares Programming over a Quotient Ring Frank Permenter1 and Pablo A. Parrilo2 Abstract— In this paper we describe a method for choosing for λi(x) and s(x), this feasibility problem is equivalent to a “good” monomial basis for a sums of squares (SOS) program an SDP. This follows because the sum of squares constraint formulated over a quotient ring. It is known that the monomial is equivalent to a semidefinite constraint on the coefficients basis need only include standard monomials with respect to a Groebner basis. We show that in many cases it is possible of s(x), and the equality constraint is equivalent to linear to use a reduced subset of standard monomials by combining equations on the coefficients of s(x) and λi(x). Groebner basis techniques with the well-known Newton poly- Naturally, the complexity of this sums of squares program tope reduction. This reduced subset of standard monomials grows with the complexity of the underlying variety, as yields a smaller semidefinite program for obtaining a certificate does the size of the corresponding SDP. It is therefore of non-negativity of a polynomial on an algebraic variety. natural to explore how algebraic structure can be exploited I. INTRODUCTION to simplify this sums of squares program. Consider the Many practical engineering problems require demonstrat- following reformulation [10], which is feasible if and only ing non-negativity of a multivariate polynomial over an if (2) is feasible: algebraic variety, i.e. over the solution set of polynomial Find s(x) equations.
    [Show full text]
  • 1 Expressing Vectors in Coordinates
    Math 416 - Abstract Linear Algebra Fall 2011, section E1 Working in coordinates In these notes, we explain the idea of working \in coordinates" or coordinate-free, and how the two are related. 1 Expressing vectors in coordinates Let V be an n-dimensional vector space. Recall that a choice of basis fv1; : : : ; vng of V is n the same data as an isomorphism ': V ' R , which sends the basis fv1; : : : ; vng of V to the n standard basis fe1; : : : ; eng of R . In other words, we have ' n ': V −! R vi 7! ei 2 3 c1 6 . 7 v = c1v1 + ::: + cnvn 7! 4 . 5 : cn This allows us to manipulate abstract vectors v = c v + ::: + c v simply as lists of numbers, 2 3 1 1 n n c1 6 . 7 n the coordinate vectors 4 . 5 2 R with respect to the basis fv1; : : : ; vng. Note that the cn coordinates of v 2 V depend on the choice of basis. 2 3 c1 Notation: Write [v] := 6 . 7 2 n for the coordinates of v 2 V with respect to the basis fvig 4 . 5 R cn fv1; : : : ; vng. For shorthand notation, let us name the basis A := fv1; : : : ; vng and then write [v]A for the coordinates of v with respect to the basis A. 2 2 Example: Using the monomial basis f1; x; x g of P2 = fa0 + a1x + a2x j ai 2 Rg, we obtain an isomorphism ' 3 ': P2 −! R 2 3 a0 2 a0 + a1x + a2x 7! 4a15 : a2 2 3 a0 2 In the notation above, we have [a0 + a1x + a2x ]fxig = 4a15.
    [Show full text]
  • Tight Monomials and the Monomial Basis Property
    PDF Manuscript TIGHT MONOMIALS AND MONOMIAL BASIS PROPERTY BANGMING DENG AND JIE DU Abstract. We generalize a criterion for tight monomials of quantum enveloping algebras associated with symmetric generalized Cartan matrices and a monomial basis property of those associated with symmetric (classical) Cartan matrices to their respective sym- metrizable case. We then link the two by establishing that a tight monomial is necessarily a monomial defined by a weakly distinguished word. As an application, we develop an algorithm to compute all tight monomials in the rank 2 Dynkin case. The existence of Hall polynomials for Dynkin or cyclic quivers not only gives rise to a simple realization of the ±-part of the corresponding quantum enveloping algebras, but also results in interesting applications. For example, by specializing q to 0, degenerate quantum enveloping algebras have been investigated in the context of generic extensions ([20], [8]), while through a certain non-triviality property of Hall polynomials, the authors [4, 5] have established a monomial basis property for quantum enveloping algebras associated with Dynkin and cyclic quivers. The monomial basis property describes a systematic construction of many monomial/integral monomial bases some of which have already been studied in the context of elementary algebraic constructions of canonical bases; see, e.g., [15, 27, 21, 5] in the simply-laced Dynkin case and [3, 18], [9, Ch. 11] in general. This property in the cyclic quiver case has also been used in [10] to obtain an elementary construction of PBW-type bases, and hence, of canonical bases for quantum affine sln. In this paper, we will complete this program by proving this property for all finite types.
    [Show full text]
  • Polynomial Interpolation CPSC 303: Numerical Approximation and Discretization
    Lecture Notes 2: Polynomial Interpolation CPSC 303: Numerical Approximation and Discretization Ian M. Mitchell [email protected] http://www.cs.ubc.ca/~mitchell University of British Columbia Department of Computer Science Winter Term Two 2012{2013 Copyright 2012{2013 by Ian M. Mitchell This work is made available under the terms of the Creative Commons Attribution 2.5 Canada license http://creativecommons.org/licenses/by/2.5/ca/ Outline • Background • Problem statement and motivation • Formulation: The linear system and its conditioning • Polynomial bases • Monomial • Lagrange • Newton • Uniqueness of polynomial interpolant • Divided differences • Divided difference tables and the Newton basis interpolant • Divided difference connection to derivatives • Osculating interpolation: interpolating derivatives • Error analysis for polynomial interpolation • Reducing the error using the Chebyshev points as abscissae CPSC 303 Notes 2 Ian M. Mitchell | UBC Computer Science 2/ 53 Interpolation Motivation n We are given a collection of data samples f(xi; yi)gi=0 n • The fxigi=0 are called the abscissae (singular: abscissa), n the fyigi=0 are called the data values • Want to find a function p(x) which can be used to estimate y(x) for x 6= xi • Why? We often get discrete data from sensors or computation, but we want information as if the function were not discretely sampled • If possible, p(x) should be inexpensive to evaluate for a given x CPSC 303 Notes 2 Ian M. Mitchell | UBC Computer Science 3/ 53 Interpolation Formulation There are lots of ways to define a function p(x) to approximate n f(xi; yi)gi=0 • Interpolation means p(xi) = yi (and we will only evaluate p(x) for mini xi ≤ x ≤ maxi xi) • Most interpolants (and even general data fitting) is done with a linear combination of (usually nonlinear) basis functions fφj(x)g n X p(x) = pn(x) = cjφj(x) j=0 where cj are the interpolation coefficients or interpolation weights CPSC 303 Notes 2 Ian M.
    [Show full text]
  • Arxiv:1805.04488V5 [Math.NA]
    GENERALIZED STANDARD TRIPLES FOR ALGEBRAIC LINEARIZATIONS OF MATRIX POLYNOMIALS∗ EUNICE Y. S. CHAN†, ROBERT M. CORLESS‡, AND LEILI RAFIEE SEVYERI§ Abstract. We define generalized standard triples X, Y , and L(z) = zC1 − C0, where L(z) is a linearization of a regular n×n −1 −1 matrix polynomial P (z) ∈ C [z], in order to use the representation X(zC1 − C0) Y = P (z) which holds except when z is an eigenvalue of P . This representation can be used in constructing so-called algebraic linearizations for matrix polynomials of the form H(z) = zA(z)B(z)+ C ∈ Cn×n[z] from generalized standard triples of A(z) and B(z). This can be done even if A(z) and B(z) are expressed in differing polynomial bases. Our main theorem is that X can be expressed using ℓ the coefficients of the expression 1 = Pk=0 ekφk(z) in terms of the relevant polynomial basis. For convenience, we tabulate generalized standard triples for orthogonal polynomial bases, the monomial basis, and Newton interpolational bases; for the Bernstein basis; for Lagrange interpolational bases; and for Hermite interpolational bases. We account for the possibility of common similarity transformations. Key words. Standard triple, regular matrix polynomial, polynomial bases, companion matrix, colleague matrix, comrade matrix, algebraic linearization, linearization of matrix polynomials. AMS subject classifications. 65F15, 15A22, 65D05 1. Introduction. A matrix polynomial P (z) ∈ Fm×n[z] is a polynomial in the variable z with coef- ficients that are m by n matrices with entries from the field F. We will use F = C, the field of complex k numbers, in this paper.
    [Show full text]
  • Groebner Bases
    1 Monomial Orders. In the polynomial algebra over F a …eld in one variable x; F [x], we can do long division (sometimes incorrectly called the Euclidean algorithm). If f(x) = n a0 + a1x + ::: + anx F [x] with an = 0 then we write deg f(x) = n and 2 6 LC(f(x)) = am. If g(x) is another element of F [x] then we have f(x) = h(x)g(x) + r(x) with h(x); r(x) F [x] and deg r[x] < m. This expression is unique. This result can be vari…ed2 by long division. As with all divisions we assume g(x) = 0. Which we will recall as a pseudo code (such code terminates on a return)6 the input being f; g and the output h and r: f0 = f; h0 = 0; k = 0; m = deg g; Repeat: If deg(fk(x) < deg(g(x) return hk(x); fk(x);n = deg fk; LC(fk(x)) n m fk+1(x) = fk(x) x g(x) LC(g(x)) LC(fk(x)) n m hk+1(x) = hk(x) + LC(g(x)) x ; k = k + 1; Continue; We note that this code terminates since at each step when there is no return then the new value of the degree of fx(x) has strictly decreased. The theory of Gröbner bases is based on a generalization of this algorithm to more than one variable. Unfortunately there is an immediate di¢ culty. The degree of a polynomial does not determine the highest degree part of the poly- nomial up to scalar multiple.
    [Show full text]
  • KRULL DIMENSION and MONOMIAL ORDERS Introduction Let R Be An
    KRULL DIMENSION AND MONOMIAL ORDERS GREGOR KEMPER AND NGO VIET TRUNG Abstract. We introduce the notion of independent sequences with respect to a mono- mial order by using the least terms of polynomials vanishing at the sequence. Our main result shows that the Krull dimension of a Noetherian ring is equal to the supremum of the length of independent sequences. The proof has led to other notions of indepen- dent sequences, which have interesting applications. For example, we can show that dim R=0 : J 1 is the maximum number of analytically independent elements in an arbi- trary ideal J of a local ring R and that dim B ≤ dim A if B ⊂ A are (not necessarily finitely generated) subalgebras of a finitely generated algebra over a Noetherian Jacobson ring. Introduction Let R be an arbitrary Noetherian ring, where a ring is always assumed to be commu- tative with identity. The aim of this paper is to characterize the Krull dimension dim R by means of a monomial order on polynomial rings over R. We are inspired of a result of Lombardi in [13] (see also Coquand and Lombardi [4], [5]) which says that for a positive integer s, dim R < s if and only if for every sequence of elements a1; : : : ; as in R, there exist nonnegative integers m1; : : : ; ms and elements c1; : : : ; cs 2 R such that m1 ms m1+1 m1 m2+1 m1 ms−1 ms+1 a1 ··· as + c1a1 + c2a1 a2 + ··· + csa1 ··· as−1 as = 0: This result has helped to develop a constructive theory for the Krull dimension [6], [7], [8].
    [Show full text]
  • An Algebraic Approach to Harmonic Polynomials on S3
    AN ALGEBRAIC APPROACH TO HARMONIC POLYNOMIALS ON S3 Kevin Mandira Limanta A thesis in fulfillment of the requirements for the degree of Master of Science (by Research) SCHOOL OF MATHEMATICS AND STATISTICS FACULTY OF SCIENCE UNIVERSITY OF NEW SOUTH WALES June 2017 ORIGINALITY STATEMENT 'I hereby declare that this submission is my own work and to the best of my knowledge it contains no materials previously published or written by another person, or substantial proportions of material which have been accepted for the award of any other degree or diploma at UNSW or any other educational institution, except where due acknowledgement is made in the thesis. Any contribution made to the research by others, with whom I have worked at UNSW or elsewhere, is explicitly acknowledged in the thesis. I also declare that the intellectual content of this thesis is the product of my own work, except to the extent that assistance from others in the project's design and conception or in style, presentation and linguistic expression is acknowledged.' Signed .... Date Show me your ways, Lord, teach me your paths. Guide me in your truth and teach me, for you are God my Savior, and my hope is in you all day long. { Psalm 25:4-5 { i This is dedicated for you, Papa. ii Acknowledgement This thesis is the the result of my two years research in the School of Mathematics and Statistics, University of New South Wales. I learned quite a number of life lessons throughout my entire study here, for which I am very grateful of.
    [Show full text]
  • Interpolation Polynomial Interpolation Piecewise Polynomial Interpolation Outline
    Interpolation Polynomial Interpolation Piecewise Polynomial Interpolation Outline 1 Interpolation 2 Polynomial Interpolation 3 Piecewise Polynomial Interpolation Michael T. Heath Scientific Computing 2 / 56 Chapter 7: Interpolation ❑ Topics: ❑ Examples ❑ Polynomial Interpolation – bases, error, Chebyshev, piecewise ❑ Orthogonal Polynomials ❑ Splines – error, end conditions ❑ Parametric interpolation ❑ Multivariate interpolation: f(x,y) Interpolation Motivation Polynomial Interpolation Choosing Interpolant Piecewise Polynomial Interpolation Existence and Uniqueness Interpolation Basic interpolation problem: for given data (t ,y ), (t ,y ),...(t ,y ) with t <t < <t 1 1 2 2 m m 1 2 ··· m determine function f : R R such that ! f(ti)=yi,i=1,...,m f is interpolating function, or interpolant, for given data Additional data might be prescribed, such as slope of interpolant at given points Additional constraints might be imposed, such as smoothness, monotonicity, or convexity of interpolant f could be function of more than one variable, but we will consider only one-dimensional case Michael T. Heath Scientific Computing 3 / 56 Interpolation Motivation Polynomial Interpolation Choosing Interpolant Piecewise Polynomial Interpolation Existence and Uniqueness Purposes for Interpolation Plotting smooth curve through discrete data points Reading between lines of table Differentiating or integrating tabular data Quick and easy evaluation of mathematical function Replacing complicated function by simple one Michael T. Heath Scientific Computing 4 / 56 Interpolation Motivation Polynomial Interpolation Choosing Interpolant Piecewise Polynomial Interpolation Existence and Uniqueness Interpolation vs Approximation By definition, interpolating function fits given data points exactly Interpolation is inappropriate if data points subject to significant errors It is usually preferable to smooth noisy data, for example by least squares approximation Approximation is also more appropriate for special function libraries Michael T.
    [Show full text]
  • 1. Fields Fields and Vector Spaces. Typical Vector Spaces: R, Q, C. for Infinite Dimensional Vector Spaces, See Notes by Karen Smith
    1. Fields Fields and vector spaces. Typical vector spaces: R; Q; C. For infinite dimensional vector spaces, see notes by Karen Smith. Important to consider a field as a vector space over a sub-field. Also have: algebraic closure of Q. Galois fields: GF (pa). Don't limit what field you work over. 2. Polynomial rings over a field Notation for a polynomial ring: K[x1; : : : ; xn]. α1 α2 αn Monomial: x1 x2 ··· xn n Set α = (α1; : : : ; αn) 2 N . α α1 α2 αn Write x for x1 x2 ··· xn . α A term is a monomial multiplied by a field element: cαx . A polynomial is a finite K-linear combination of monomials: X α f = cαx ; α so a polynomial is a finite sum of terms. The support of f are the monomials that appear (with non-zero coefficients) in the polynomial f. If α = (α1; : : : ; αn), put jαj = α1 + ··· + αn. α If f 2 K[x1; : : : ; xn], deg(f) = maxfjαj : x is in the support of fg. Example 2.1. f = 7x3y2z + 11xyz2 deg(f) = maxf6; 4g = 6. 7x3y2z is a term. x3y2z is a monomial. n Given f 2 K[x1; : : : ; xn], evaluation is the map Ff : K ! K given by (c1; : : : ; cn) ! f(c1; : : : ; cn). When is Ff the zero map? Example 2.2. If K is a finite field, Ff can be the zero map without f being the zero polynomial. For instance take the field with two elements, K = Z=2Z, and consider the polynomial f = x2 + x = x(x + 1). Then f is not zero in the ring K[x], however f(c) = 0 for all c 2 K (there are only two to check!).
    [Show full text]
  • Gröbner Bases Tutorial
    Gröbner Bases Tutorial David A. Cox Gröbner Basics Gröbner Bases Tutorial Notation and Definitions Gröbner Bases Part I: Gröbner Bases and the Geometry of Elimination The Consistency and Finiteness Theorems Elimination Theory The Elimination Theorem David A. Cox The Extension and Closure Theorems Department of Mathematics and Computer Science Prove Extension and Amherst College Closure ¡ ¢ £ ¢ ¤ ¥ ¡ ¦ § ¨ © ¤ ¥ ¨ Theorems The Extension Theorem ISSAC 2007 Tutorial The Closure Theorem An Example Constructible Sets References Outline Gröbner Bases Tutorial 1 Gröbner Basics David A. Cox Notation and Definitions Gröbner Gröbner Bases Basics Notation and The Consistency and Finiteness Theorems Definitions Gröbner Bases The Consistency and 2 Finiteness Theorems Elimination Theory Elimination The Elimination Theorem Theory The Elimination The Extension and Closure Theorems Theorem The Extension and Closure Theorems 3 Prove Extension and Closure Theorems Prove The Extension Theorem Extension and Closure The Closure Theorem Theorems The Extension Theorem An Example The Closure Theorem Constructible Sets An Example Constructible Sets 4 References References Begin Gröbner Basics Gröbner Bases Tutorial David A. Cox k – field (often algebraically closed) Gröbner α α α Basics x = x 1 x n – monomial in x ,...,x Notation and 1 n 1 n Definitions α ··· Gröbner Bases c x , c k – term in x1,...,xn The Consistency and Finiteness Theorems ∈ k[x]= k[x1,...,xn] – polynomial ring in n variables Elimination Theory An = An(k) – n-dimensional affine space over k The Elimination Theorem n The Extension and V(I)= V(f1,...,fs) A – variety of I = f1,...,fs Closure Theorems ⊆ nh i Prove I(V ) k[x] – ideal of the variety V A Extension and ⊆ ⊆ Closure √I = f k[x] m f m I – the radical of I Theorems { ∈ |∃ ∈ } The Extension Theorem The Closure Theorem Recall that I is a radical ideal if I = √I.
    [Show full text]