Optimization of the Determinant of the Vandermonde Matrix and Related Matrices

Total Page:16

File Type:pdf, Size:1020Kb

Optimization of the Determinant of the Vandermonde Matrix and Related Matrices Methodol Comput Appl Probab (2018) 20:1417–1428 https://doi.org/10.1007/s11009-017-9595-y Optimization of the Determinant of the Vandermonde Matrix and Related Matrices Karl Lundengard˚ 1 · Jonas Osterberg¨ 1 · Sergei Silvestrov1 Received: 30 October 2015 / Revised: 1 September 2017 / Accepted: 28 September 2017 / Published online: 14 November 2017 © The Author(s) 2017. This article is an open access publication Abstract The value of the Vandermonde determinant is optimized over various surfaces, including the sphere, ellipsoid and torus. Lagrange multipliers are used to find a system of polynomial equations which give the local extreme points in its solutions. Using Grobner¨ basis and other techniques the extreme points are given either explicitly or as roots of poly- nomials in one variable. The behavior of the Vandermonde determinant is also presented visually in some interesting cases. Keywords Vandermonde determinant · Optimization · Grobner¨ basis · Orthogonal polynomials · Ellipsoid · Optimal experiment design · Homogeneous polynomials Mathematics Subject Classification (2010) 33C45 · 11C20 · 15B99 · 08B99 1 Introduction In this paper we will consider the extreme points of the Vandermonde determinant on various surfaces. The examination is primarily motivated by mathematical curiosity but the techniques used here are likely to be extensible to some problems related to optimal experiment design for polynomial regression, see Section 3. Karl Lundengard˚ [email protected] Jonas Osterberg¨ [email protected] Sergei Silvestrov [email protected] 1 Division of Applied Mathematics, UKK, Malardalen¨ University, Hogskoleplan¨ 1, Box 883, 721 23 Vaster¨ as,˚ Sweden 1418 Methodol Comput Appl Probab (2018) 20:1417–1428 To the authors knowledge this problem has previously been examined on cubes (see Sec- tion 3) and spheres, see Szego(˝ 1939). Here we will consider some techniques to extend some of these results to other surfaces such as ellipsoids, cylinders, p-norm spheres and other surfaces defined by homogeneous polynomials. Our examination will mostly be restricted to three dimensions but many of the techniques can in principle be extended to higher dimensions. 2 The Vandermonde Matrix A rectangular Vandermonde matrix of size m×n is determined by n values x = (x1, ··· ,xn) and is defined by ⎡ ⎤ 11··· 1 ⎢ ··· ⎥ ⎢ x1 x2 xn ⎥ = i−1 = ⎢ ⎥ Vmn(x) xj . (1) mn ⎣ . .. ⎦ m−1 m−1 ··· m−1 x1 x2 xn Note that some authors use the transpose of this as the definition and possibly also let indices run from 0. All entries in the first row of Vandermonde matrices are ones and by considering 0 0 = 1 this is true even when some xj is zero. In this paper we will primarily consider square Vandermonde matrices and for convenience we will use the notation Vn(x) = Vnn(x). The determinant of the Vandermonde matrix is well known. Theorem 1 The determinant of square Vandermonde matrices has the form det Vn(x) ≡ vn(x) = (xj − xi). (2) 1≤i<j≤n This determinant is also simply referred to as the Vandermonde determinant or Vander- monde polynomial or Vandermondian (Vein and Dale 1999). In this paper we will use the method of Lagrange multipliers to optimize the Vandermonde determinant over a surface. For this purpose the following properties will be useful. n(n−1) Lemma 1 The Vandermonde determinant is a homogeneous polynomial of degree 2 . n n(n − 1) Proof Considering (2) the numbers of factor of v (x) is i − 1 = . Thus n 2 i=1 n(n−1) vn(cx) = c 2 vn(x). (3) 3 Application to D-optimal Experiment Designs for Polynomial Regression with a Cost-function Suppose an experiment is conducted where m data points from some compact interval, X ⊂ R, i = 1, 2,...,m, are used to create a polynomial regression model of degree Methodol Comput Appl Probab (2018) 20:1417–1428 1419 m n − 1. A vector containing the data points, xm = (x1,x2,...,xm) ∈ X , is called a design m and a design is said to be D-optimal if det(Mn(xm)) ≥ det(Mn(ym)) for all y ∈ X where ⎡ ⎤ m m − ⎢ m x ... xn 1 ⎥ ⎢ i i ⎥ ⎢ i=1 i=1 ⎥ ⎢ m m m ⎥ ⎢ ⎥ ⎢ x x2 ... xn ⎥ ⎢ i i i ⎥ Mm(x) = ⎢ = = = ⎥ ⎢ i 1 i 1 i 1 ⎥ ⎢ . ⎥ ⎢ . .. ⎥ ⎢ m m m ⎥ ⎣ n−1 n 2n−2 ⎦ xi xi ... xi i=1 i=1 i=1 is the Fischer information matrix, see Kiefer (1959) and Gaffke and Krafft (1982). Optimal experiment design is often used to find the minimum number of points needed for a certain model. If we let m = n we get a interpolation problem defined by a square Van- dermonde matrix and the Fischer information matrix is Mn(x) = Vn(x) Vn(x) and since 2 Vn(x) is an n × n matrix det(Mn(x)) = det(Vn(x) ) det(Vn(x)) = det(Vn(x)) . Thus the maximization of the determinant of the Fischer information matrix is equivalent to find- ing the extreme points of the determinant of a square Vandermonde matrix in some volume given by the set of possible designs, see Gaffke and Krafft (1982). Optimal designs for various kinds of polynomial regression models are known, see Dette and Trampisch (2010) for an overview. The typical set of possible design is given by con- straining each parameter of the model to be in a certain interval. Usually these intervals are also normalized such that the vector of parameters x can be found in the n-dimensional cube x ∈[−1, 1]n. When consider certain other sets of possible designs the results presented in this paper might be useful. Suppose there is a cost-function associated with the data such that the total cost of the experiment being below some threshold value, g(x) ≤ 1, defines some compact set, G ={x ∈ Rm|g(x) ≤ 1}, such that G ⊂ X m. Since the Vandermonde determinant is a homogeneous polynomial for any c>1 |vn(x)| > |vn(cx)| the extreme points will be on the surface of the compact set and thus it is enough to consider the set of points defined by g(x) = 1. 4 Optimization using Grobner¨ Bases Grobner¨ bases together with algorithms to find them, and algorithms for solving a polyno- mial equation is an important tool that arises in many applications. One such application is the optimization of polynomials over affine varieties through the method of Lagrange mul- tipliers. We will here give some main points and informal discussion on these methods as an introduction and to fix some notation. Definition 1 (Cox et al. 1997)Letf1, ··· ,fm be polynomials in R[x1, ··· ,xn].Theaffine n variety V(f1, ··· ,fm) defined by f1, ··· ,fm is the set of all points (x1, ··· ,xn) ∈ R such that fi(x1, ··· ,xn) = 0forall1≤ i ≤ m. 1420 Methodol Comput Appl Probab (2018) 20:1417–1428 When n = 3 we will sometimes use the variables x,y,z instead of x1,x2,x3. Affine vari- eties are the common zeros of a set of multivariate polynomials. Such sets of polynomials will generate a greater set of polynomials (Cox et al. 1997)by m f1, ··· ,fm ≡ hifi : h1, ··· ,hm ∈ R[x1, ··· ,xn] , i=1 and this larger set will define the same variety. But it will also define an ideal (a set of polynomials that contains the zero-polynomial and is closed under addition, and absorbs multiplication by any other polynomial) by I(f1, ··· ,fm) =f1, ··· ,fm .AGrobner¨ basis for this ideal is then a finite set of polynomials {g1, ··· ,gk} such that the ideal gener- ated by the leading terms of the polynomials g1, ··· ,gk is the same ideal as that generated by all the leading terms of polynomials in I =f1, ··· ,fm . In this paper we consider the optimization of the Vandermonde determinant vn(x) over surfaces defined by a polynomial equation on the form n p sn(x1, ··· ,xn ; p; a1, ··· ,an) ≡ ai|xi| = 1, (4) i=1 where we will select the constants ai and p to get ellipsoids in three dimensions, cylinders in three dimensions, and spheres under the p-norm in n dimensions. The case of the ellipsoid is suitable for solution by Grobner¨ basis methods, but due to the existing symmetries the spheres are more suitable for other methods, as provided in Section 8. From (3) and the convexity of the interior of the sets defined by (4), under a suitable choice of the constant p and non-negative ai, it is easy to see that the optimal value of n | |p ≤ n | |p = vn on i=1 ai xi 1 will be attained on i=1 ai xi 1. And so, by the method of Lagrange multipliers we have that the minimal/maximal values of vn(x1, ··· ,xn) on sn(x1, ··· ,xn) ≤ 1 will be attained at points such that ∂vn/∂xi − λ∂sn/∂xi = 0for1≤ i ≤ n and some constant λ and sn(x1, ··· ,xn) − 1 = 0, Tyrrell Rockafellar (1993). For p = 2 the resulting set of equations will form a set of polynomials in λ, x1, ··· ,xn. These polynomials will define an ideal over R[λ, x1, ··· ,xn], and by finding a Grbner basis for this ideal we can use the especially nice properties of Grbner bases to find analytical solutions to these problems, that is, to find roots for the polynomials in the computed basis. 5 Extreme Points on the Ellipsoid in Three Dimensions In this section we will find the extreme points of the Vandermonde determinant on the three dimensional ellipsoid given by ax2 + by2 + cz2 = 1, (x,y,z) ∈ R3 (5) where a>0, b>0, c>0. Using the method of Lagrange multipliers together with (5) and some rewriting gives that all stationary points of the Vandermonde determinant lie in the variety V = V ax2 + by2 + cz2 − 1,ax+ by + cz, ax(z − x)(y − x) − by(z − y)(y − x) + cz(z − y)(z − x)) .
Recommended publications
  • 1111: Linear Algebra I
    1111: Linear Algebra I Dr. Vladimir Dotsenko (Vlad) Lecture 11 Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 1 / 13 Previously on. Theorem. Let A be an n × n-matrix, and b a vector with n entries. The following statements are equivalent: (a) the homogeneous system Ax = 0 has only the trivial solution x = 0; (b) the reduced row echelon form of A is In; (c) det(A) 6= 0; (d) the matrix A is invertible; (e) the system Ax = b has exactly one solution. A very important consequence (finite dimensional Fredholm alternative): For an n × n-matrix A, the system Ax = b either has exactly one solution for every b, or has infinitely many solutions for some choices of b and no solutions for some other choices. In particular, to prove that Ax = b has solutions for every b, it is enough to prove that Ax = 0 has only the trivial solution. Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 2 / 13 An example for the Fredholm alternative Let us consider the following question: Given some numbers in the first row, the last row, the first column, and the last column of an n × n-matrix, is it possible to fill the numbers in all the remaining slots in a way that each of them is the average of its 4 neighbours? This is the \discrete Dirichlet problem", a finite grid approximation to many foundational questions of mathematical physics. Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 3 / 13 An example for the Fredholm alternative For instance, for n = 4 we may face the following problem: find a; b; c; d to put in the matrix 0 4 3 0 1:51 B 1 a b -1C B C @0:5 c d 2 A 2:1 4 2 1 so that 1 a = 4 (3 + 1 + b + c); 1 8b = 4 (a + 0 - 1 + d); >c = 1 (a + 0:5 + d + 4); > 4 < 1 d = 4 (b + c + 2 + 2): > > This is a system with 4:> equations and 4 unknowns.
    [Show full text]
  • Determinants
    12 PREFACECHAPTER I DETERMINANTS The notion of a determinant appeared at the end of 17th century in works of Leibniz (1646–1716) and a Japanese mathematician, Seki Kova, also known as Takakazu (1642–1708). Leibniz did not publish the results of his studies related with determinants. The best known is his letter to l’Hospital (1693) in which Leibniz writes down the determinant condition of compatibility for a system of three linear equations in two unknowns. Leibniz particularly emphasized the usefulness of two indices when expressing the coefficients of the equations. In modern terms he actually wrote about the indices i, j in the expression xi = j aijyj. Seki arrived at the notion of a determinant while solving the problem of finding common roots of algebraic equations. In Europe, the search for common roots of algebraic equations soon also became the main trend associated with determinants. Newton, Bezout, and Euler studied this problem. Seki did not have the general notion of the derivative at his disposal, but he actually got an algebraic expression equivalent to the derivative of a polynomial. He searched for multiple roots of a polynomial f(x) as common roots of f(x) and f (x). To find common roots of polynomials f(x) and g(x) (for f and g of small degrees) Seki got determinant expressions. The main treatise by Seki was published in 1674; there applications of the method are published, rather than the method itself. He kept the main method in secret confiding only in his closest pupils. In Europe, the first publication related to determinants, due to Cramer, ap- peared in 1750.
    [Show full text]
  • Moments of Logarithmic Derivative SO
    Alvarez, E. , & Snaith, N. C. (2020). Moments of the logarithmic derivative of characteristic polynomials from SO(2N) and USp(2N). Journal of Mathematical Physics, 61, [103506 (2020)]. https://doi.org/10.1063/5.0008423, https://doi.org/10.1063/5.0008423 Peer reviewed version Link to published version (if available): 10.1063/5.0008423 10.1063/5.0008423 Link to publication record in Explore Bristol Research PDF-document This is the author accepted manuscript (AAM). The final published version (version of record) is available online via American Institute of Physics at https://aip.scitation.org/doi/10.1063/5.0008423 . Please refer to any applicable terms of use of the publisher. University of Bristol - Explore Bristol Research General rights This document is made available in accordance with publisher policies. Please cite only the published version using the reference above. Full terms of use are available: http://www.bristol.ac.uk/red/research-policy/pure/user-guides/ebr-terms/ Moments of the logarithmic derivative of characteristic polynomials from SO(N) and USp(2N) E. Alvarez1, a) and N.C. Snaith1, b) School of Mathematics, University of Bristol, UK We study moments of the logarithmic derivative of characteristic polynomials of orthogonal and symplectic random matrices. In particular, we compute the asymptotics for large matrix size, N, of these moments evaluated at points which are approaching 1. This follows work of Bailey, Bettin, Blower, Conrey, Prokhorov, Rubinstein and Snaith where they compute these asymptotics in the case of unitary random matrices. a)Electronic mail: [email protected] b)Electronic mail (Corresponding author): [email protected] 1 I.
    [Show full text]
  • 1 1.1. the DFT Matrix
    FFT January 20, 2016 1 1.1. The DFT matrix. The DFT matrix. By definition, the sequence f(τ)(τ = 0; 1; 2;:::;N − 1), posesses a discrete Fourier transform F (ν)(ν = 0; 1; 2;:::;N − 1), given by − 1 NX1 F (ν) = f(τ)e−i2π(ν=N)τ : (1.1) N τ=0 Of course, this definition can be immediately rewritten in the matrix form as follows 2 3 2 3 F (1) f(1) 6 7 6 7 6 7 6 7 6 F (2) 7 1 6 f(2) 7 6 . 7 = p F 6 . 7 ; (1.2) 4 . 5 N 4 . 5 F (N − 1) f(N − 1) where the DFT (i.e., the discrete Fourier transform) matrix is defined by 2 3 1 1 1 1 · 1 6 − 7 6 1 w w2 w3 ··· wN 1 7 6 7 h i 6 1 w2 w4 w6 ··· w2(N−1) 7 1 − − 1 6 7 F = p w(k 1)(j 1) = p 6 3 6 9 ··· 3(N−1) 7 N 1≤k;j≤N N 6 1 w w w w 7 6 . 7 4 . 5 1 wN−1 w2(N−1) w3(N−1) ··· w(N−1)(N−1) (1.3) 2πi with w = e N being the primitive N-th root of unity. 1.2. The IDFT matrix. To recover N values of the function from its discrete Fourier transform we simply have to invert the DFT matrix to obtain 2 3 2 3 f(1) F (1) 6 7 6 7 6 f(2) 7 p 6 F (2) 7 6 7 −1 6 7 6 .
    [Show full text]
  • Numerically Stable Coded Matrix Computations Via Circulant and Rotation Matrix Embeddings
    1 Numerically stable coded matrix computations via circulant and rotation matrix embeddings Aditya Ramamoorthy and Li Tang Department of Electrical and Computer Engineering Iowa State University, Ames, IA 50011 fadityar,[email protected] Abstract Polynomial based methods have recently been used in several works for mitigating the effect of stragglers (slow or failed nodes) in distributed matrix computations. For a system with n worker nodes where s can be stragglers, these approaches allow for an optimal recovery threshold, whereby the intended result can be decoded as long as any (n − s) worker nodes complete their tasks. However, they suffer from serious numerical issues owing to the condition number of the corresponding real Vandermonde-structured recovery matrices; this condition number grows exponentially in n. We present a novel approach that leverages the properties of circulant permutation matrices and rotation matrices for coded matrix computation. In addition to having an optimal recovery threshold, we demonstrate an upper bound on the worst-case condition number of our recovery matrices which grows as ≈ O(ns+5:5); in the practical scenario where s is a constant, this grows polynomially in n. Our schemes leverage the well-behaved conditioning of complex Vandermonde matrices with parameters on the complex unit circle, while still working with computation over the reals. Exhaustive experimental results demonstrate that our proposed method has condition numbers that are orders of magnitude lower than prior work. I. INTRODUCTION Present day computing needs necessitate the usage of large computation clusters that regularly process huge amounts of data on a regular basis. In several of the relevant application domains such as machine learning, arXiv:1910.06515v4 [cs.IT] 9 Jun 2021 datasets are often so large that they cannot even be stored in the disk of a single server.
    [Show full text]
  • Some Results on Vandermonde Matrices with an Application to Time Series Analysis∗
    SIAM J. MATRIX ANAL. APPL. c 2003 Society for Industrial and Applied Mathematics Vol. 25, No. 1, pp. 213–223 SOME RESULTS ON VANDERMONDE MATRICES WITH AN APPLICATION TO TIME SERIES ANALYSIS∗ ANDRE´ KLEIN† AND PETER SPREIJ‡ Abstract. In this paper we study Stein equations in which the coefficient matrices are in companion form. Solutions to such equations are relatively easy to compute as soon as one knows how to invert a Vandermonde matrix (in the generic case where all eigenvalues have multiplicity one) or a confluent Vandermonde matrix (in the general case). As an application we present a way to compute the Fisher information matrix of an autoregressive moving average (ARMA) process. The computation is based on the fact that this matrix can be decomposed into blocks where each block satisfies a certain Stein equation. Key words. ARMA process, Fisher information matrix, Stein’s equation, Vandermonde matrix, confluent Vandermonde matrix AMS subject classifications. 15A09, 15A24, 62F10, 62M10, 65F05, 65F10, 93B25 PII. S0895479802402892 1. Introduction. In this paper we investigate some properties of (confluent) Vandermonde and related matrices aimed at and motivated by their application to a problem in time series analysis. Specifically, we show how to apply results on these matrices to obtain a simpler representation of the (asymptotic) Fisher information matrixof an autoregressive moving average (ARMA) process. The Fisher informa- tion matrixis prominently featured in the asymptotic analysis of estimators and in asymptotic testing theory, e.g., in the classical Cram´er–Rao bound on the variance of unbiased estimators. See [10] for general results and see [2] for time series models.
    [Show full text]
  • Zero-Sum Triangles for Involutory, Idempotent, Nilpotent and Unipotent Matrices
    Zero-Sum Triangles for Involutory, Idempotent, Nilpotent and Unipotent Matrices Pengwei Hao1, Chao Zhang2, Huahan Hao3 Abstract: In some matrix formations, factorizations and transformations, we need special matrices with some properties and we wish that such matrices should be easily and simply generated and of integers. In this paper, we propose a zero-sum rule for the recurrence relations to construct integer triangles as triangular matrices with involutory, idempotent, nilpotent and unipotent properties, especially nilpotent and unipotent matrices of index 2. With the zero-sum rule we also give the conditions for the special matrices and the generic methods for the generation of those special matrices. The generated integer triangles are mostly newly discovered, and more combinatorial identities can be found with them. Keywords: zero-sum rule, triangles of numbers, generation of special matrices, involutory, idempotent, nilpotent and unipotent matrices 1. Introduction Zero-sum was coined in early 1940s in the field of game theory and economic theory, and the term “zero-sum game” is used to describe a situation where the sum of gains made by one person or group is lost in equal amounts by another person or group, so that the net outcome is neutral, or the sum of losses and gains is zero [1]. It was later also frequently used in psychology and political science. In this work, we apply the zero-sum rule to three adjacent cells to construct integer triangles. Pascal's triangle is named after the French mathematician Blaise Pascal, but the history can be traced back in almost 2,000 years and is referred to as the Staircase of Mount Meru in India, the Khayyam triangle in Persia (Iran), Yang Hui's triangle in China, Apianus's Triangle in Germany, and Tartaglia's triangle in Italy [2].
    [Show full text]
  • Analytical Solution of Linear Ordinary Differential Equations by Differential
    ANALYTICAL SOLUTION OF LINEAR ORDINARY DIFFERENTIAL EQUATIONS BY DIFFERENTIAL TRANSFER MATRIX METHOD SINA KHORASANI AND ALI ADIBI Abstract. We report a new analytical method for exact solution of homoge- neous linear ordinary differential equations with arbitrary order and variable coefficients. The method is based on the definition of jump transfer matrices and their extension into limiting differential form. The approach reduces the nth-order differential equation to a system of n linear differential equations with unity order. The full analytical solution is then found by the pertur- bation technique. The important feature of the presented method is that it deals with the evolution of independent solutions, rather than its derivatives. We prove the validity of method by direct substitution of the solution in the original differential equation. We discuss the general properties of differential transfer matrices and present several analytical examples, showing the appli- cability of the method. We show that the Abel-Liouville-Ostogradski theorem can be easily recovered through this approach. 1. Introduction Solution of ordinary differential equations (ODEs) is in general possible by dif- ferent methods [1]. Lie group analysis [2-6] can be effectively used in cases when a symmetry algebra exists. Factorization methods are reported for reduction of ODEs into linear autonomous forms [7,8] with constant coefficients, which can be readily solved. But these methods require some symbolic transformations and are complicated for general higher order ODEs, and thus not suitable for many com- putational purposes where the form of ODE is not available in a simple closed form. arXiv:math-ph/0301010v7 12 Mar 2003 General second-order linear ODEs can be solved in terms of special function [9] and analytic continuation [10].
    [Show full text]
  • Fast Algorithms to Compute Matrix Vector Products for Pascal Matrices
    FAST ALGORITHMS TO COMPUTE MATRIX-VECTOR PRODUCTS FOR PASCAL MATRICES∗ ZHIHUI TANG† , RAMANI DURAISWAMI‡, AND NAIL GUMEROV§ Abstract. The Pascal matrix arises in a number of applications. We present a few ways to decompose the Pascal matrices of size n n into products of matrices with structure. Based on these decompositions, we propose fast algorithms× to compute the product of a Pascal matrix and a vector with complexity O(n log n). We also present a strategy to stabilize the proposed algorithms. Finally, we also present some interesting properties of the Pascal matrices that help us to compute fast the product of the inverse of a Pascal matrix and a vector, and fast algorithms for generalized Pascal Matrices. Key words. Matrix-vector product, Pascal matrices, matrix decomposition, structured matri- ces, Toeplitz matrices, Hankel matrices, Vandermonde matrices Introduction. If we arrange the binomial coefficients r! Cn = ,r=0, 1, 2, ··· ,n=0, 1, ··· ,r (0.1) r n!(r − n)! in staggered rows, 1 11 121 1331 , (0.2) 14641 1 5 10 10 5 1 1 6 15 20 15 6 1 we obtain the famous Pascal triangle. However, Pascal was not the first to discover this triangle, it had been described about 500 years earlier by Chinese mathematician Yang Hui. Therefore in China, it is known as “Yang Hui’s triangle" . It also has been discovered in India, Persia and Europe. For a fascinating introduction, see [Edwards02]. The Pascal matrix arises in many applications such as in binomial expansion, probability, combinatorics, linear algebra [Edelman03], electrical engineering [Biolkova99] and order statistics [Aggarwala2001].
    [Show full text]
  • On the Vandermonde Matrix
    On the Vandermonde Matrix Tom Copeland ([email protected]) May 2014 Part I The n-simplices and the Vandermonde matrix There are several denitions of the Vandermonde matrix Vn. Here 2 1 1 ··· 1 3 6 x1 x2 ··· xn 7 6 . 7 Vn = Vn(x1; x2; :::; xn) = 6 . .. 7 : 4 . 5 n−1 n−1 n−1 x1 x2 ··· xn with the determinant Y jVnj = jVn(x1; : : : ; xn)j = (xj − xi): 1≤i<j≤n The coordinates of the n vertices Cn−1(x1);Cn−1(x2); :::; Cn−1(xn) of the 2 n−1 (n − 1)-simplex along the unique moment curve Cn−1(x) = (x; x ; :::; x ) (ra- tional normal curve) that passes through the vertices comprise the columns of Vn (except for the rst row of ones). With one of the vertices located at the origin, say vertex m with xm = 0, the determinant gives the (possibly signed) hyper-volume of the hyper-parallelepiped framed by the position vectors from the origin to the other vertices, which bound the facet opposing the vertex at the origin (see the V 3 example below). Note that the xj are unit-less parameters, i.e., have no physical units or measures attached to them and in that sense are dimensionless. In other words, k should not be directly interpreted as a quantity with dimension a length, xj k an area, a volume, etc. It is a dimensionless parameter that must be interpreted in context. Ignoring this leads to confusion, such as believing that jV4j with six factors in its product formula must somehow have dimension six.
    [Show full text]
  • Extreme Points of the Vandermonde Determinant on the Sphere and Some Limits Involving the Generalized Vandermonde Determinant
    Extreme points of the Vandermonde determinant on the sphere and some limits involving the generalized Vandermonde determinant Karl Lundengård, Jonas Österberg, Sergei Silvestrov Mälardalen University December 24, 2013 Abstract The values of the determinant of Vandermonde matrices with real elements are analyzed both visually and analytically over the unit sphere in various di- mensions. For three dimensions some generalized Vandermonde matrices are analyzed visually. The extreme points of the ordinary Vandermonde deter- minant on finite-dimensional unit spheres are given as the roots of rescaled Hermite polynomials and a recursion relation is provided for the polynomial coefficients. Analytical expressions for these roots are also given for dimension three to seven. A transformation of the optimization problem is provided and some relations between the ordinary and generalized Vandermonde matrices involving limits are discussed. Keywords: Vandermonde matrix, Determinants, Extreme points, Unit sphere, Generalized Vandermonde matrix. arXiv:1312.6193v1 [math.CA] 21 Dec 2013 1 Contents 1 Introduction 3 1.1 Visual exploration in 3D . .4 2 Optimizing the Vandermonde determinant over the unit sphere 8 2.1 The extreme points given by roots of a polynomial . .8 2.2 Extreme points of the Vandermonde determinant on the three dimen- sional unit sphere . 15 2.3 A transformation of the optimization problem . 16 2.4 Further visual exploration . 18 2.4.1 Four dimensions . 19 2.4.2 Five dimensions . 20 2.4.3 Six dimensions . 21 2.4.4 Seven dimensions . 23 3 Some limit theorems involving the generalized Vandermonde matrix 24 4 Conclusions 27 2 1 Introduction Here a generalized Vandermonde matrix is determined by two vectors ~xn = (x1; ··· ; xn) 2 n m K and ~am = (a1; ··· ; am) 2 K , where K is usually the real or complex field, and is defined as G (~x ;~a ) = xai : (1) mn n m j mn For square matrices only one index is given, Gn ≡ Gnn.
    [Show full text]
  • Unified Canonical Forms for Matrices Over a Differential Ring
    Unified Canonical Forms for Matrices Over a Differential Ring J. Zhu and C. D. Johnson Electrical and Computer Engineering Department University of Alabama in Huntsville Huntsville, Alabama 35899 Submitted by Russell Menis ABSTRACT Linear differential equations with variable coefficients of the vector form (i) % = A(t)x and the scalar form (ii) y(“) + cy,(t)y(“-‘) + . * . + a2(thj + a,(t)y = 0 can be studied as operators on a differential module over a differential ring. Using this differential algebraic structure and a classical result on differential operator factoriza- tions developed by Cauchy and Floquet, a new (variable) eigenvalue theory and an associated set of matrix canonical forms are developed in this paper for matrices over a differential ring. In particular, eight new spectral and modal matrix canonical forms are developed that include as special cases the classical companion, Jordan (diagonal), and (generalized) Vandermonde canonical forms, as traditionally used for matrices over a number ring (field). Therefore, these new matrix canonical forms can be viewed as unifications of those classical ones. Two additional canonical matrices that perform order reductions of a Cauchy-Floquet factorization of a linear differential operator are also introduced. General and explicit formulas for obtaining the new canonical forms are derived in terms of the new (variable) eigenvalues. The results obtained here have immediate applications in diverse engineering and scientific fields, such as physics, control systems, communications, and electrical networks, in which linear differential equations (i) and (ii) are used as mathematical models. 1. INTRODUCTION This paper is concerned with the class of finite-dimensional, homoge- neous linear differential equations of the vector form S = A( t)x, x(h) = x0 (1.1) LlNEAR ALGEBRA AND ITS APPLlCATlONS 147:201-248 (1991) 201 0 Elsevier Science Publishing Co., Inc., 1991 655 Avenue of the Americas, New York, NY 10010 0024-3795/91/$3.50 202 J.
    [Show full text]