Tridiagonal-Diagonal Reduction of Symmetric Indefinite Pairs∗

Total Page:16

File Type:pdf, Size:1020Kb

Tridiagonal-Diagonal Reduction of Symmetric Indefinite Pairs∗ TRIDIAGONAL-DIAGONAL REDUCTION OF SYMMETRIC INDEFINITE PAIRS¤ FRANC» OISE TISSEURy Abstract. We consider the reduction of a symmetric inde¯nite matrix pair (A; B), with B non- singular, to tridiagonal-diagonal form by congruence transformations. This is an important reduction in solving polynomial eigenvalue problems with symmetric coe±cient matrices and in frequency re- sponse computations. First the pair is reduced to symmetric-diagonal form. Then we describe three methods reducing the symmetric-diagonal pair to tridiagonal-diagonal form. Two of them employ more stable versions of Brebner and Grad's pseudosymmetric Givens and pseudosymmetric House- holder reductions, while the third is new and based on a combination of Householder reflectors and hyperbolic rotations. We prove an optimality condition for the transformations used in the third reduction. We present numerical experiments that compare the di®erent approaches and show improvements over Brebner and Grad's reductions. Key words. symmetric inde¯nite generalized eigenvalue problem, tridiagonalization, hyperbolic rotation, uni¯ed rotation, hyperbolic Householder reflector. AMS subject classi¯cations. 65F15 1. Introduction. Motivation for this work comes from the symmetric polyno- mial eigenvalue problem (PEP) m m¡1 (1.1) (¸ A + ¸ A 1 + + A0)u = 0; m m¡ ¢ ¢ ¢ where the Ai, i = 0: m, are n n symmetric matrices. ¸ is called an eigenvalue and u = 0 is the corresponding righ£t eigenvector. The standard way of dealing with the PEP6 in practice is to reformulate it as a generalized eigenvalue problem (GEP) (1.2) Ax = ¸Bx; of size mn. This process is called linearization, as the GEP is linear in ¸. Symmetry in the problem is maintained with an appropriate choice of linearization. For example, we can take 0 0 A0 0 0 A0 0 ¢ ¢ ¢ ¢ ¢ ¢ ¢ ¢ ¢ . 2 . A0 A1 3 2 . A0 A1 . 3 A = . ; B = . 6 . 7 6 0 A0 . 7 6 7 6 7 6 0 A0 Am¡2 7 6 A0 A1 Am¡2 0 7 6 7 6 ¢ ¢ ¢ 7 6 A0 A1 Am¡2 Am¡1 7 6 0 0 Am 7 4 ¢ ¢ ¢ 5 4 ¢ ¢ ¢ ¢ ¢ ¢ ¡ 5 and x = [uT ; ¸uT ; : : : ; ¸m¡1uT ]T . The resulting A and B are symmetric but not de¯nite, and in general the pair (A; B) is inde¯nite. The ¯rst step in most eigensystem computations is the reduction, in a ¯nite number of operations, to a simple form. Only then is an iterative procedure applied. A symmetric inde¯nite pair (A; B) can be reduced to Hessenberg-triangular form and the ¤Numerical Analysis Report 409, Manchester Centre for Computational Mathematics, September 2002. yDepartment of Mathematics, University of Manchester, Manchester, M13 9PL, England ([email protected], http://www.ma.man.ac.uk/~ftisseur/). This work was supported by En- gineering and Physical Sciences Research Council grant GR/L76532 and Nu±eld Foundation grant NAL/00216/G. 1 resulting generalized eigenvalue problem solved by the QZ algorithm. This approach is numerically stable, but unfortunately the reduction to Hessenberg-triangular form destroys the symmetry. Moreover, in ¯nite precision arithmetic there is no guarantee that the set of left and right eigenvectors computed via the QZ algorithm will coincide, a property possessed by GEPs with real symmetric matrices. Also, by preserving symmetry, storage and computational costs can be reduced. The tridiagonal-diagonal reduction of a pair (A; B) is the most compact form we can obtain in a ¯nite number of steps. Such reductions have been proposed by Brebner and Grad [5] and by ZurmuhlÄ and Falk [26] for nonsingular B. They re- quire non-orthogonal transformations and can be unstable. Once (A; B) is reduced to tridiagonal-diagonal form the eigenvalues and eigenvectors can be obtained by ap- plying, for example, an HR iteration or associated iterations [5], [6], [16], [25], Uhlig's DQR algorithm [24], or, if one is interested in the eigenvalues only, Aberth's method can be used in an e±cient way [1]. A robust tridiagonal-diagonal reduction is therefore of prime importance before one can consider using any of the methods cited above. We note that Garvey et al. [8] have considered a less compact form that allows the second matrix to be in tridiagonal form. One feature of their approach is that no assumption is made on the nonsingularity of the two matrices. The simultaneous tridiagonalization is convenient if one needs to solve linear systems of the form (A !B)x = b ¡ for many values of !, as is required in frequency response computations [8], but it is less attractive than the tridiagonal-diagonal form for eigenvalue computations. Three di®erent tridiagonal-diagonal reductions for inde¯nite pairs (A; B) with B nonsingular are described in this paper. They all consist of two stages. The ¯rst, common to all, is the reduction of the symmetric inde¯nite pair (A; B) to symmetric- diagonal form (C; J) with the aid of a block LDLT factorization of B. During the second stage, C is tridiagonalized using a sequence of congruence transformations that preserve the diagonal form of the second matrix J. Each of the three reductions proposed in this paper uses di®erent types of transformations. These transformations are not necessarily orthogonal, so they may be unstable in ¯nite precision arithmetic. We descrive several techniques that can be used to make them more robust and to improve stability during the reduction process: in particular, pivoting and zeroing strategies in order to minimize the condition numbers of the transformations, and mixed application of hyperbolic rotations. The paper is organized as follows. Section 2 sets up notations and de¯nitions. It is shown that if the tridiagonal-diagonal reduction exists, it is determined up to signs by the ¯rst column of the transformation matrix. Section 3 describes the ¯rst stage of the reduction, that is, the reduction of (A; B) to symmetric-diagonal form (C; J). The description is accompanied by an error analysis. The second stage of the reduc- tion is described in section 4. Three algorithms are proposed. The ¯rst two are an improvement over Brebner and Grad's pseudosymmetric Givens and pseudosymmet- ric Householder methods [5]. The third algorithm is based on some transformations used to compute hyperbolic QR factorizations in inde¯nite least square problems [3]. Numerical comparisons of these algorithms and comparisons to Brebner and Grad's reductions are given in the last section. 2. Background material. Unless otherwise speci¯ed, denotes the 2-norm. We denote by diagn( 1) the set of all n n diagonal matrices withk ¢ k q diagonal elements q § £ 2 n equal to 1 and n q equal to 1. A matrix J diagq ( 1) for some q is called a signature matrix. ¡ ¡ 2 § n Let J, J diag ( 1). A matrix H Rn£n is said to be (J; J)-orthogonal 2 q § 2 if HT JH = J. Note that (J; J)-orthogonal matrices are sometimes called (J; J)- hyperexchangee or (J; J)-hypernormal matrices in the signal processing literaturee [17]. We recall ethat a tridiagonalematrix is unreduced if none of its next-to-diagonale elements (that is, theeelements on the ¯rst subdiagonal and the ¯rst superdiagonal) is zero. The following result, is related to the implicit Q theorem [11]. A more general form can be found in [18, Th.2.2]. Theorem 2.1. If C Rn£n admits a representation of the form 2 (2.1) QT CQ = T; where T is unreduced tridiagonal and Q is (J; J)-orthogonal then the columns of Q and the next-to-diagonal elements of T are determined up to signs by the ¯rst (or last) column of Q. e We give the proof since we need to refer to it later in the text. This is a construc- tive proof that describes a Lanczos process. Proof. Let J = diag(σ~ ), σ~ = 1, i = 1: n, and i i § e ®1 ¯2 ¯2 ®2 ¯3 2 .. .. 3 T = ¯3 . : 6 7 6 .. 7 6 . ®n¡1 ¯n 7 6 7 6 ¯n ®n 7 4 5 T We assume that q1 is given and normalized such that σ~1 = q1 Jq1. This yields T ®1 = q1 Cq1: Using the (J; J)-orthogonality of Q, equation (2.1) can be rewritten as (2.2) e JCQ = QJT: Equating the ¯rst column on each side of (2.2)egives p1 JCq1 ®1σ~1q1 = ¯2σ~2q2: ´ ¡ ¡2 T From the (J; J)-orthogonality of Q we get σ~2 = ¯2 p1 Jp1, which implies T T e σ~2 = sign(p Jp1); ¯2 = p Jp1 ; 1 § j 1 j q so that ¡1 q2 = σ~2¯2 p1 is determined up to the sign chosen for ¯2. The second diagonal element of T is uniquely determined by T ®2 = q2 Cq2: 3 Hence, the construction of q2, ®2, ¯2 and σ~2 requires just the knowledge of p1. Now suppose that the ¯rst j < n columns of Q and the leading j j principal submatrices of T and J are known. Then by equating the jth columns£on each side of (2.2) we obtain e p JCq σ~ ® q σ~ 1¯ q 1 = σ~ +1¯ +1q +1: j ´ j ¡ j j j ¡ j¡ j j¡ j j j Using once again the (J; J)-orthogonality of Q we have T T (2.3) σ~ +1e= sign(p Jp ); ¯ +1 = p Jp : j j j j § j j j j q Hence ¡1 T (2.4) qj+1 = σ~j+1¯j+1pj; ®j+1 = qj+1Cqj+1: Again, ¯j+1 and qj+1 are determined up to a sign. By induction on j all columns of Q and all next-to-diagonal elements of T are determined, up to a sign by q1.
Recommended publications
  • Eigenvalues and Eigenvectors of Tridiagonal Matrices∗
    Electronic Journal of Linear Algebra ISSN 1081-3810 A publication of the International Linear Algebra Society Volume 15, pp. 115-133, April 2006 ELA http://math.technion.ac.il/iic/ela EIGENVALUES AND EIGENVECTORS OF TRIDIAGONAL MATRICES∗ SAID KOUACHI† Abstract. This paper is continuation of previous work by the present author, where explicit formulas for the eigenvalues associated with several tridiagonal matrices were given. In this paper the associated eigenvectors are calculated explicitly. As a consequence, a result obtained by Wen- Chyuan Yueh and independently by S. Kouachi, concerning the eigenvalues and in particular the corresponding eigenvectors of tridiagonal matrices, is generalized. Expressions for the eigenvectors are obtained that differ completely from those obtained by Yueh. The techniques used herein are based on theory of recurrent sequences. The entries situated on each of the secondary diagonals are not necessary equal as was the case considered by Yueh. Key words. Eigenvectors, Tridiagonal matrices. AMS subject classifications. 15A18. 1. Introduction. The subject of this paper is diagonalization of tridiagonal matrices. We generalize a result obtained in [5] concerning the eigenvalues and the corresponding eigenvectors of several tridiagonal matrices. We consider tridiagonal matrices of the form −α + bc1 00 ... 0 a1 bc2 0 ... 0 .. .. 0 a2 b . An = , (1) .. .. .. 00. 0 . .. .. .. . cn−1 0 ... ... 0 an−1 −β + b n−1 n−1 ∞ where {aj}j=1 and {cj}j=1 are two finite subsequences of the sequences {aj}j=1 and ∞ {cj}j=1 of the field of complex numbers C, respectively, and α, β and b are complex numbers. We suppose that 2 d1, if j is odd ajcj = 2 j =1, 2, ..., (2) d2, if j is even where d 1 and d2 are complex numbers.
    [Show full text]
  • Parametrizations of K-Nonnegative Matrices
    Parametrizations of k-Nonnegative Matrices Anna Brosowsky, Neeraja Kulkarni, Alex Mason, Joe Suk, Ewin Tang∗ October 2, 2017 Abstract Totally nonnegative (positive) matrices are matrices whose minors are all nonnegative (positive). We generalize the notion of total nonnegativity, as follows. A k-nonnegative (resp. k-positive) matrix has all minors of size k or less nonnegative (resp. positive). We give a generating set for the semigroup of k-nonnegative matrices, as well as relations for certain special cases, i.e. the k = n − 1 and k = n − 2 unitriangular cases. In the above two cases, we find that the set of k-nonnegative matrices can be partitioned into cells, analogous to the Bruhat cells of totally nonnegative matrices, based on their factorizations into generators. We will show that these cells, like the Bruhat cells, are homeomorphic to open balls, and we prove some results about the topological structure of the closure of these cells, and in fact, in the latter case, the cells form a Bruhat-like CW complex. We also give a family of minimal k-positivity tests which form sub-cluster algebras of the total positivity test cluster algebra. We describe ways to jump between these tests, and give an alternate description of some tests as double wiring diagrams. 1 Introduction A totally nonnegative (respectively totally positive) matrix is a matrix whose minors are all nonnegative (respectively positive). Total positivity and nonnegativity are well-studied phenomena and arise in areas such as planar networks, combinatorics, dynamics, statistics and probability. The study of total positivity and total nonnegativity admit many varied applications, some of which are explored in “Totally Nonnegative Matrices” by Fallat and Johnson [5].
    [Show full text]
  • Package 'Igraphmatch'
    Package ‘iGraphMatch’ January 27, 2021 Type Package Title Tools Graph Matching Version 1.0.1 Description Graph matching methods and analysis. The package works for both 'igraph' objects and matrix objects. You provide the adjacency matrices of two graphs and some other information you might know, choose the graph matching method, and it returns the graph matching results. 'iGraphMatch' also provides a bunch of useful functions you might need related to graph matching. Depends R (>= 3.3.1) Imports clue (>= 0.3-54), Matrix (>= 1.2-11), igraph (>= 1.1.2), irlba, methods, Rcpp Suggests dplyr (>= 0.5.0), testthat (>= 2.0.0), knitr, rmarkdown VignetteBuilder knitr License GPL (>= 2) LazyData TRUE RoxygenNote 7.1.1 Language en-US Encoding UTF-8 LinkingTo Rcpp NeedsCompilation yes Author Daniel Sussman [aut, cre], Zihuan Qiao [aut], Joshua Agterberg [ctb], Lujia Wang [ctb], Vince Lyzinski [ctb] Maintainer Daniel Sussman <[email protected]> Repository CRAN Date/Publication 2021-01-27 16:00:02 UTC 1 2 bari_start R topics documented: bari_start . .2 best_matches . .4 C.Elegans . .5 center_graph . .6 check_seeds . .7 do_lap . .8 Enron . .9 get_perm . .9 graph_match_convex . 10 graph_match_ExpandWhenStuck . 12 graph_match_IsoRank . 13 graph_match_Umeyama . 15 init_start . 16 innerproduct . 17 lapjv . 18 lapmod . 18 largest_common_cc . 19 matched_adjs . 20 match_plot_igraph . 20 match_report . 22 pad.............................................. 23 row_cor . 24 rperm . 25 sample_correlated_gnp_pair . 25 sample_correlated_ieg_pair . 26 sample_correlated_sbm_pair . 28 split_igraph . 29 splrMatrix-class . 30 splr_sparse_plus_constant . 31 splr_to_sparse . 31 Index 32 bari_start Start matrix initialization Description initialize the start matrix for graph matching iteration. bari_start 3 Usage bari_start(nns, ns = 0, soft_seeds = NULL) rds_sinkhorn_start(nns, ns = 0, soft_seeds = NULL, distribution = "runif") rds_perm_bari_start(nns, ns = 0, soft_seeds = NULL, g = 1, is_splr = TRUE) rds_from_sim_start(nns, ns = 0, soft_seeds = NULL, sim) Arguments nns An integer.
    [Show full text]
  • (Hessenberg) Eigenvalue-Eigenmatrix Relations∗
    (HESSENBERG) EIGENVALUE-EIGENMATRIX RELATIONS∗ JENS-PETER M. ZEMKE† Abstract. Explicit relations between eigenvalues, eigenmatrix entries and matrix elements are derived. First, a general, theoretical result based on the Taylor expansion of the adjugate of zI − A on the one hand and explicit knowledge of the Jordan decomposition on the other hand is proven. This result forms the basis for several, more practical and enlightening results tailored to non-derogatory, diagonalizable and normal matrices, respectively. Finally, inherent properties of (upper) Hessenberg, resp. tridiagonal matrix structure are utilized to construct computable relations between eigenvalues, eigenvector components, eigenvalues of principal submatrices and products of lower diagonal elements. Key words. Algebraic eigenvalue problem, eigenvalue-eigenmatrix relations, Jordan normal form, adjugate, principal submatrices, Hessenberg matrices, eigenvector components AMS subject classifications. 15A18 (primary), 15A24, 15A15, 15A57 1. Introduction. Eigenvalues and eigenvectors are defined using the relations Av = vλ and V −1AV = J. (1.1) We speak of a partial eigenvalue problem, when for a given matrix A ∈ Cn×n we seek scalar λ ∈ C and a corresponding nonzero vector v ∈ Cn. The scalar λ is called the eigenvalue and the corresponding vector v is called the eigenvector. We speak of the full or algebraic eigenvalue problem, when for a given matrix A ∈ Cn×n we seek its Jordan normal form J ∈ Cn×n and a corresponding (not necessarily unique) eigenmatrix V ∈ Cn×n. Apart from these constitutional relations, for some classes of structured matrices several more intriguing relations between components of eigenvectors, matrix entries and eigenvalues are known. For example, consider the so-called Jacobi matrices.
    [Show full text]
  • Explicit Inverse of a Tridiagonal (P, R)–Toeplitz Matrix
    Explicit inverse of a tridiagonal (p; r){Toeplitz matrix A.M. Encinas, M.J. Jim´enez Departament de Matemtiques Universitat Politcnica de Catalunya Abstract Tridiagonal matrices appears in many contexts in pure and applied mathematics, so the study of the inverse of these matrices becomes of specific interest. In recent years the invertibility of nonsingular tridiagonal matrices has been quite investigated in different fields, not only from the theoretical point of view (either in the framework of linear algebra or in the ambit of numerical analysis), but also due to applications, for instance in the study of sound propagation problems or certain quantum oscillators. However, explicit inverses are known only in a few cases, in particular when the tridiagonal matrix has constant diagonals or the coefficients of these diagonals are subjected to some restrictions like the tridiagonal p{Toeplitz matrices [7], such that their three diagonals are formed by p{periodic sequences. The recent formulae for the inversion of tridiagonal p{Toeplitz matrices are based, more o less directly, on the solution of second order linear difference equations, although most of them use a cumbersome formulation, that in fact don not take into account the periodicity of the coefficients. This contribution presents the explicit inverse of a tridiagonal matrix (p; r){Toeplitz, which diagonal coefficients are in a more general class of sequences than periodic ones, that we have called quasi{periodic sequences. A tridiagonal matrix A = (aij) of order n + 2 is called (p; r){Toeplitz if there exists m 2 N0 such that n + 2 = mp and ai+p;j+p = raij; i; j = 0;:::; (m − 1)p: Equivalently, A is a (p; r){Toeplitz matrix iff k ai+kp;j+kp = r aij; i; j = 0; : : : ; p; k = 0; : : : ; m − 1: We have developed a technique that reduces any linear second order difference equation with periodic or quasi-periodic coefficients to a difference equation of the same kind but with constant coefficients [3].
    [Show full text]
  • On Reciprocal Systems and Controllability
    On reciprocal systems and controllability Timothy H. Hughes a aDepartment of Mathematics, University of Exeter, Penryn Campus, Penryn, Cornwall, TR10 9EZ, UK Abstract In this paper, we extend classical results on (i) signature symmetric realizations, and (ii) signature symmetric and passive realizations, to systems which need not be controllable. These results are motivated in part by the existence of important electrical networks, such as the famous Bott-Duffin networks, which possess signature symmetric and passive realizations that are uncontrollable. In this regard, we provide necessary and sufficient algebraic conditions for a behavior to be realized as the driving-point behavior of an electrical network comprising resistors, inductors, capacitors and transformers. Key words: Reciprocity; Passive system; Linear system; Controllability; Observability; Behaviors; Electrical networks. 1 Introduction Practical motivation for developing a theory of reci- procity that does not assume controllability arises from This paper is concerned with reciprocal systems (see, electrical networks. Notably, the driving-point behavior e.g., Casimir, 1963; Willems, 1972; Anderson and Vong- of an electrical network comprising resistors, inductors, panitlerd, 2006; Newcomb, 1966; van der Schaft, 2011). capacitors and transformers (an RLCT network) is nec- Reciprocity is an important form of symmetry in physi- essarily reciprocal, and also passive, 2 but it need not be cal systems which arises in acoustics (Rayleigh-Carson controllable (see C¸amlibel et al., 2003; Willems, 2004; reciprocity); elasticity (the Maxwell-Betti reciprocal Hughes, 2017d). Indeed, as noted by C¸amlibel et al. work theorem); electrostatics (Green's reciprocity); and (2003), it is not known what (uncontrollable) behav- electromagnetics (Lorentz reciprocity), where it follows iors can be realized as the driving-point behavior of an as a result of Maxwell's laws (Newcomb, 1966, p.
    [Show full text]
  • Emergent Spacetimes
    Emergent spacetimes Silke Christine Edith Weinfurtner Department of Mathematics, Statistics and Computer Science — Te Kura Tatau arXiv:0711.4416v1 [gr-qc] 28 Nov 2007 Thesis submitted for the degree of Doctor of Philosophy at the Victoria University of Wellington. In Memory of Johann Weinfurtner Abstract In this thesis we discuss the possibility that spacetime geometry may be an emergent phenomenon. This idea has been motivated by the Analogue Gravity programme. An “effective gravitational field” dominates the kinematics of small perturbations in an Analogue Model. In these models there is no obvious connection between the “gravitational” field tensor and the Einstein equations, as the emergent spacetime geometry arises as a consequence of linearising around some classical field. After a brief survey of the most relevant literature on this topic, we present our contributions to the field. First, we show that the spacetime geometry on the equatorial slice through a rotating Kerr black hole is formally equivalent to the geometry felt by phonons entrained in a rotating fluid vortex. The most general acoustic geometry is compatible with the fluid dynamic equations in a collapsing/ ex- panding perfect-fluid line vortex. We demonstrate that there is a suitable choice of coordinates on the equatorial slice through a Kerr black hole that puts it into this vortex form; though it is not possible to put the entire Kerr spacetime into perfect-fluid “acoustic” form. We then discuss an analogue spacetime based on the propagation of excitations in a 2-component Bose–Einstein condensate. This analogue spacetime has a very rich and complex structure, which permits us to provide a mass-generating mechanism for the quasi-particle excitations.
    [Show full text]
  • Determinant Formulas and Cofactors
    Determinant formulas and cofactors Now that we know the properties of the determinant, it’s time to learn some (rather messy) formulas for computing it. Formula for the determinant We know that the determinant has the following three properties: 1. det I = 1 2. Exchanging rows reverses the sign of the determinant. 3. The determinant is linear in each row separately. Last class we listed seven consequences of these properties. We can use these ten properties to find a formula for the determinant of a 2 by 2 matrix: � � � � � � � a b � � a 0 � � 0 b � � � = � � + � � � c d � � c d � � c d � � � � � � � � � � a 0 � � a 0 � � 0 b � � 0 b � = � � + � � + � � + � � � c 0 � � 0 d � � c 0 � � 0 d � = 0 + ad + (−cb) + 0 = ad − bc. By applying property 3 to separate the individual entries of each row we could get a formula for any other square matrix. However, for a 3 by 3 matrix we’ll have to add the determinants of twenty seven different matrices! Many of those determinants are zero. The non-zero pieces are: � � � � � � � � � a a a � � a 0 0 � � a 0 0 � � 0 a 0 � � 11 12 13 � � 11 � � 11 � � 12 � � a21 a22 a23 � = � 0 a22 0 � + � 0 0 a23 � + � a21 0 0 � � � � � � � � � � a31 a32 a33 � � 0 0 a33 � � 0 a32 0 � � 0 0 a33 � � � � � � � � 0 a 0 � � 0 0 a � � 0 0 a � � 12 � � 13 � � 13 � + � 0 0 a23 � + � a21 0 0 � + � 0 a22 0 � � � � � � � � a31 0 0 � � 0 a32 0 � � a31 0 0 � = a11 a22a33 − a11a23 a33 − a12a21a33 +a12a23a31 + a13 a21a32 − a13a22a31. Each of the non-zero pieces has one entry from each row in each column, as in a permutation matrix.
    [Show full text]
  • The Unsymmetric Eigenvalue Problem
    Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Supplemental Notes The Unsymmetric Eigenvalue Problem Properties and Decompositions Let A be an n × n matrix. A nonzero vector x is called an eigenvector of A if there exists a scalar λ such that Ax = λx: The scalar λ is called an eigenvalue of A, and we say that x is an eigenvector of A corresponding to λ. We see that an eigenvector of A is a vector for which matrix-vector multiplication with A is equivalent to scalar multiplication by λ. We say that a nonzero vector y is a left eigenvector of A if there exists a scalar λ such that λyH = yH A: The superscript H refers to the Hermitian transpose, which includes transposition and complex conjugation. That is, for any matrix A, AH = AT . An eigenvector of A, as defined above, is sometimes called a right eigenvector of A, to distinguish from a left eigenvector. It can be seen that if y is a left eigenvector of A with eigenvalue λ, then y is also a right eigenvector of AH , with eigenvalue λ. Because x is nonzero, it follows that if x is an eigenvector of A, then the matrix A − λI is singular, where λ is the corresponding eigenvalue. Therefore, λ satisfies the equation det(A − λI) = 0: The expression det(A−λI) is a polynomial of degree n in λ, and therefore is called the characteristic polynomial of A (eigenvalues are sometimes called characteristic values). It follows from the fact that the eigenvalues of A are the roots of the characteristic polynomial that A has n eigenvalues, which can repeat, and can also be complex, even if A is real.
    [Show full text]
  • Sparse Linear Systems Section 4.2 – Banded Matrices
    Band Systems Tridiagonal Systems Cyclic Reduction Parallel Numerical Algorithms Chapter 4 – Sparse Linear Systems Section 4.2 – Banded Matrices Michael T. Heath and Edgar Solomonik Department of Computer Science University of Illinois at Urbana-Champaign CS 554 / CSE 512 Michael T. Heath and Edgar Solomonik Parallel Numerical Algorithms 1 / 28 Band Systems Tridiagonal Systems Cyclic Reduction Outline 1 Band Systems 2 Tridiagonal Systems 3 Cyclic Reduction Michael T. Heath and Edgar Solomonik Parallel Numerical Algorithms 2 / 28 Band Systems Tridiagonal Systems Cyclic Reduction Banded Linear Systems Bandwidth (or semibandwidth) of n × n matrix A is smallest value w such that aij = 0 for all ji − jj > w Matrix is banded if w n If w p, then minor modifications of parallel algorithms for dense LU or Cholesky factorization are reasonably efficient for solving banded linear system Ax = b If w / p, then standard parallel algorithms for LU or Cholesky factorization utilize few processors and are very inefficient Michael T. Heath and Edgar Solomonik Parallel Numerical Algorithms 3 / 28 Band Systems Tridiagonal Systems Cyclic Reduction Narrow Banded Linear Systems More efficient parallel algorithms for narrow banded linear systems are based on divide-and-conquer approach in which band is partitioned into multiple pieces that are processed simultaneously Reordering matrix by nested dissection is one example of this approach Because of fill, such methods generally require more total work than best serial algorithm for system with dense band We will illustrate for tridiagonal linear systems, for which w = 1, and will assume pivoting is not needed for stability (e.g., matrix is diagonally dominant or symmetric positive definite) Michael T.
    [Show full text]
  • Eigenvalues of a Special Tridiagonal Matrix
    Eigenvalues of a Special Tridiagonal Matrix Alexander De Serre Rothney∗ October 10, 2013 Abstract In this paper we consider a special tridiagonal test matrix. We prove that its eigenvalues are the even integers 2;:::; 2n and show its relationship with the famous Kac-Sylvester tridiagonal matrix. 1 Introduction We begin with a quick overview of the theory of symmetric tridiagonal matrices, that is, we detail a few basic facts about tridiagonal matrices. In particular, we describe the symmetrization process of a tridiagonal matrix as well as the orthogonal polynomials that arise from the characteristic polynomials of said matrices. Definition 1.1. A tridiagonal matrix, Tn, is of the form: 2a1 b1 0 ::: 0 3 6 :: :: : 7 6c1 a2 : : : 7 6 7 6 :: :: :: 7 Tn = 6 0 : : : 0 7 ; (1.1) 6 7 6 : :: :: 7 4 : : : an−1 bn−15 0 ::: 0 cn−1 an where entries below the subdiagonal and above the superdiagonal are zero. If bi 6= 0 for i = 1; : : : ; n − 1 and ci 6= 0 for i = 1; : : : ; n − 1, Tn is called a Jacobi matrix. In this paper we will use a more compact notation and only describe the subdiagonal, diagonal, and superdiagonal (where appropriate). For example, Tn can be rewritten as: 0 1 b1 : : : bn−1 Tn = @ a1 a2 : : : an−1 an A : (1.2) c1 : : : cn−1 ∗Bishop's University, Sherbrooke, Quebec, Canada 1 Note that the study of symmetric tridiagonal matrices is sufficient for our purpose as any Jacobi matrix with bici > 0 8i can be symmetrized through a similarity transformation: 0 p p 1 b1c1 ::: bn−1cn−1 −1 An = Dn TnDn = @ a1 a2 : : : an−1 an A ; (1.3) p p b1c1 ::: bn−1cn−1 r cici+1 ··· cn−1 where Dn = diag(γ1; : : : ; γn) and γi = .
    [Show full text]
  • Institute of Computer Science Efficient Tridiagonal Preconditioner for The
    Institute of Computer Science Academy of Sciences of the Czech Republic Efficient tridiagonal preconditioner for the matrix-free truncated Newton method Ladislav Lukˇsan,Jan Vlˇcek Technical report No. 1177 January 2013 Pod Vod´arenskou vˇeˇz´ı2, 182 07 Prague 8 phone: +420 2 688 42 44, fax: +420 2 858 57 89, e-mail:e-mail:[email protected] Institute of Computer Science Academy of Sciences of the Czech Republic Efficient tridiagonal preconditioner for the matrix-free truncated Newton method Ladislav Lukˇsan,Jan Vlˇcek 1 Technical report No. 1177 January 2013 Abstract: In this paper, we study an efficient tridiagonal preconditioner, based on numerical differentiation, applied to the matrix-free truncated Newton method for unconstrained optimization. It is proved that this preconditioner is positive definite for many practical problems. The efficiency of the resulting matrix-free truncated Newton method is demonstrated by results of extensive numerical experiments. Keywords: 1This work was supported by the Institute of Computer Science of the AS CR (RVO:67985807) 1 Introduction We consider the unconstrained minimization problem x∗ = arg min F (x); x2Rn where function F : D(F ) ⊂ Rn ! R is twice continuously differentiable and n is large. We use the notation g(x) = rF (x);G(x) = r2F (x) and the assumption that kG(x)k ≤ G, 8x 2 D(F ). Numerical methods for unconstrained minimization are usually iterative and their iteration step has the form xk+1 = xk + αkdk; k 2 N; where dk is a direction vector and αk is a step-length. In this paper, we will deal with the Newton method, which uses the quadratic model 1 F (x + d) ≈ Q(x + d) = F (x ) + gT (x )d + dT G(x )d (1) k k k k 2 k for direction determination in such a way that dk = arg min Q(xk + d): (2) d2Mk There are two basic possibilities for direction determination: the line-search method, where n n Mk = R ; and the trust-region method, where Mk = fd 2 R : kdk ≤ ∆kg (here ∆k > 0 is the trust region radius).
    [Show full text]