Positive Semidefinite Matrices
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Math 221: LINEAR ALGEBRA
Math 221: LINEAR ALGEBRA Chapter 8. Orthogonality §8-3. Positive Definite Matrices Le Chen1 Emory University, 2020 Fall (last updated on 11/10/2020) Creative Commons License (CC BY-NC-SA) 1 Slides are adapted from those by Karen Seyffarth from University of Calgary. Positive Definite Matrices Cholesky factorization – Square Root of a Matrix Positive Definite Matrices Definition An n × n matrix A is positive definite if it is symmetric and has positive eigenvalues, i.e., if λ is a eigenvalue of A, then λ > 0. Theorem If A is a positive definite matrix, then det(A) > 0 and A is invertible. Proof. Let λ1; λ2; : : : ; λn denote the (not necessarily distinct) eigenvalues of A. Since A is symmetric, A is orthogonally diagonalizable. In particular, A ∼ D, where D = diag(λ1; λ2; : : : ; λn). Similar matrices have the same determinant, so det(A) = det(D) = λ1λ2 ··· λn: Since A is positive definite, λi > 0 for all i, 1 ≤ i ≤ n; it follows that det(A) > 0, and therefore A is invertible. Theorem A symmetric matrix A is positive definite if and only if ~xTA~x > 0 for all n ~x 2 R , ~x 6= ~0. Proof. Since A is symmetric, there exists an orthogonal matrix P so that T P AP = diag(λ1; λ2; : : : ; λn) = D; where λ1; λ2; : : : ; λn are the (not necessarily distinct) eigenvalues of A. Let n T ~x 2 R , ~x 6= ~0, and define ~y = P ~x. Then ~xTA~x = ~xT(PDPT)~x = (~xTP)D(PT~x) = (PT~x)TD(PT~x) = ~yTD~y: T Writing ~y = y1 y2 ··· yn , 2 y1 3 6 y2 7 ~xTA~x = y y ··· y diag(λ ; λ ; : : : ; λ ) 6 7 1 2 n 1 2 n 6 . -
Orthogonal Reduction 1 the Row Echelon Form -.: Mathematical
MATH 5330: Computational Methods of Linear Algebra Lecture 9: Orthogonal Reduction Xianyi Zeng Department of Mathematical Sciences, UTEP 1 The Row Echelon Form Our target is to solve the normal equation: AtAx = Atb ; (1.1) m×n where A 2 R is arbitrary; we have shown previously that this is equivalent to the least squares problem: min jjAx−bjj : (1.2) x2Rn t n×n As A A2R is symmetric positive semi-definite, we can try to compute the Cholesky decom- t t n×n position such that A A = L L for some lower-triangular matrix L 2 R . One problem with this approach is that we're not fully exploring our information, particularly in Cholesky decomposition we treat AtA as a single entity in ignorance of the information about A itself. t m×m In particular, the structure A A motivates us to study a factorization A=QE, where Q2R m×n is orthogonal and E 2 R is to be determined. Then we may transform the normal equation to: EtEx = EtQtb ; (1.3) t m×m where the identity Q Q = Im (the identity matrix in R ) is used. This normal equation is equivalent to the least squares problem with E: t min Ex−Q b : (1.4) x2Rn Because orthogonal transformation preserves the L2-norm, (1.2) and (1.4) are equivalent to each n other. Indeed, for any x 2 R : jjAx−bjj2 = (b−Ax)t(b−Ax) = (b−QEx)t(b−QEx) = [Q(Qtb−Ex)]t[Q(Qtb−Ex)] t t t t t t t t 2 = (Q b−Ex) Q Q(Q b−Ex) = (Q b−Ex) (Q b−Ex) = Ex−Q b : Hence the target is to find an E such that (1.3) is easier to solve. -
Romani Syntactic Typology Evangelia Adamou, Yaron Matras
Romani Syntactic Typology Evangelia Adamou, Yaron Matras To cite this version: Evangelia Adamou, Yaron Matras. Romani Syntactic Typology. Yaron Matras; Anton Tenser. The Palgrave Handbook of Romani Language and Linguistics, Springer, pp.187-227, 2020, 978-3-030-28104- 5. 10.1007/978-3-030-28105-2_7. halshs-02965238 HAL Id: halshs-02965238 https://halshs.archives-ouvertes.fr/halshs-02965238 Submitted on 13 Oct 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Romani syntactic typology Evangelia Adamou and Yaron Matras 1. State of the art This chapter presents an overview of the principal syntactic-typological features of Romani dialects. It draws on the discussion in Matras (2002, chapter 7) while taking into consideration more recent studies. In particular, we draw on the wealth of morpho- syntactic data that have since become available via the Romani Morpho-Syntax (RMS) database.1 The RMS data are based on responses to the Romani Morpho-Syntax questionnaire recorded from Romani speaking communities across Europe and beyond. We try to take into account a representative sample. We also take into consideration data from free-speech recordings available in the RMS database and the Pangloss Collection. -
Handout 9 More Matrix Properties; the Transpose
Handout 9 More matrix properties; the transpose Square matrix properties These properties only apply to a square matrix, i.e. n £ n. ² The leading diagonal is the diagonal line consisting of the entries a11, a22, a33, . ann. ² A diagonal matrix has zeros everywhere except the leading diagonal. ² The identity matrix I has zeros o® the leading diagonal, and 1 for each entry on the diagonal. It is a special case of a diagonal matrix, and A I = I A = A for any n £ n matrix A. ² An upper triangular matrix has all its non-zero entries on or above the leading diagonal. ² A lower triangular matrix has all its non-zero entries on or below the leading diagonal. ² A symmetric matrix has the same entries below and above the diagonal: aij = aji for any values of i and j between 1 and n. ² An antisymmetric or skew-symmetric matrix has the opposite entries below and above the diagonal: aij = ¡aji for any values of i and j between 1 and n. This automatically means the digaonal entries must all be zero. Transpose To transpose a matrix, we reect it across the line given by the leading diagonal a11, a22 etc. In general the result is a di®erent shape to the original matrix: a11 a21 a11 a12 a13 > > A = A = 0 a12 a22 1 [A ]ij = A : µ a21 a22 a23 ¶ ji a13 a23 @ A > ² If A is m £ n then A is n £ m. > ² The transpose of a symmetric matrix is itself: A = A (recalling that only square matrices can be symmetric). -
On the Eigenvalues of Euclidean Distance Matrices
“main” — 2008/10/13 — 23:12 — page 237 — #1 Volume 27, N. 3, pp. 237–250, 2008 Copyright © 2008 SBMAC ISSN 0101-8205 www.scielo.br/cam On the eigenvalues of Euclidean distance matrices A.Y. ALFAKIH∗ Department of Mathematics and Statistics University of Windsor, Windsor, Ontario N9B 3P4, Canada E-mail: [email protected] Abstract. In this paper, the notion of equitable partitions (EP) is used to study the eigenvalues of Euclidean distance matrices (EDMs). In particular, EP is used to obtain the characteristic poly- nomials of regular EDMs and non-spherical centrally symmetric EDMs. The paper also presents methods for constructing cospectral EDMs and EDMs with exactly three distinct eigenvalues. Mathematical subject classification: 51K05, 15A18, 05C50. Key words: Euclidean distance matrices, eigenvalues, equitable partitions, characteristic poly- nomial. 1 Introduction ( ) An n ×n nonzero matrix D = di j is called a Euclidean distance matrix (EDM) 1, 2,..., n r if there exist points p p p in some Euclidean space < such that i j 2 , ,..., , di j = ||p − p || for all i j = 1 n where || || denotes the Euclidean norm. i , ,..., Let p , i ∈ N = {1 2 n}, be the set of points that generate an EDM π π ( , ,..., ) D. An m-partition of D is an ordered sequence = N1 N2 Nm of ,..., nonempty disjoint subsets of N whose union is N. The subsets N1 Nm are called the cells of the partition. The n-partition of D where each cell consists #760/08. Received: 07/IV/08. Accepted: 17/VI/08. ∗Research supported by the Natural Sciences and Engineering Research Council of Canada and MITACS. -
Serial Verb Constructions Revisited: a Case Study from Koro
Serial Verb Constructions Revisited: A Case Study from Koro By Jessica Cleary-Kemp A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Linguistics in the Graduate Division of the University of California, Berkeley Committee in charge: Associate Professor Lev D. Michael, Chair Assistant Professor Peter S. Jenks Professor William F. Hanks Summer 2015 © Copyright by Jessica Cleary-Kemp All Rights Reserved Abstract Serial Verb Constructions Revisited: A Case Study from Koro by Jessica Cleary-Kemp Doctor of Philosophy in Linguistics University of California, Berkeley Associate Professor Lev D. Michael, Chair In this dissertation a methodology for identifying and analyzing serial verb constructions (SVCs) is developed, and its application is exemplified through an analysis of SVCs in Koro, an Oceanic language of Papua New Guinea. SVCs involve two main verbs that form a single predicate and share at least one of their arguments. In addition, they have shared values for tense, aspect, and mood, and they denote a single event. The unique syntactic and semantic properties of SVCs present a number of theoretical challenges, and thus they have invited great interest from syntacticians and typologists alike. But characterizing the nature of SVCs and making generalizations about the typology of serializing languages has proven difficult. There is still debate about both the surface properties of SVCs and their underlying syntactic structure. The current work addresses some of these issues by approaching serialization from two angles: the typological and the language-specific. On the typological front, it refines the definition of ‘SVC’ and develops a principled set of cross-linguistically applicable diagnostics. -
QUADRATIC FORMS and DEFINITE MATRICES 1.1. Definition of A
QUADRATIC FORMS AND DEFINITE MATRICES 1. DEFINITION AND CLASSIFICATION OF QUADRATIC FORMS 1.1. Definition of a quadratic form. Let A denote an n x n symmetric matrix with real entries and let x denote an n x 1 column vector. Then Q = x’Ax is said to be a quadratic form. Note that a11 ··· a1n . x1 Q = x´Ax =(x1...xn) . xn an1 ··· ann P a1ixi . =(x1,x2, ··· ,xn) . P anixi 2 (1) = a11x1 + a12x1x2 + ... + a1nx1xn 2 + a21x2x1 + a22x2 + ... + a2nx2xn + ... + ... + ... 2 + an1xnx1 + an2xnx2 + ... + annxn = Pi ≤ j aij xi xj For example, consider the matrix 12 A = 21 and the vector x. Q is given by 0 12x1 Q = x Ax =[x1 x2] 21 x2 x1 =[x1 +2x2 2 x1 + x2 ] x2 2 2 = x1 +2x1 x2 +2x1 x2 + x2 2 2 = x1 +4x1 x2 + x2 1.2. Classification of the quadratic form Q = x0Ax: A quadratic form is said to be: a: negative definite: Q<0 when x =06 b: negative semidefinite: Q ≤ 0 for all x and Q =0for some x =06 c: positive definite: Q>0 when x =06 d: positive semidefinite: Q ≥ 0 for all x and Q = 0 for some x =06 e: indefinite: Q>0 for some x and Q<0 for some other x Date: September 14, 2004. 1 2 QUADRATIC FORMS AND DEFINITE MATRICES Consider as an example the 3x3 diagonal matrix D below and a general 3 element vector x. 100 D = 020 004 The general quadratic form is given by 100 x1 0 Q = x Ax =[x1 x2 x3] 020 x2 004 x3 x1 =[x 2 x 4 x ] x2 1 2 3 x3 2 2 2 = x1 +2x2 +4x3 Note that for any real vector x =06 , that Q will be positive, because the square of any number is positive, the coefficients of the squared terms are positive and the sum of positive numbers is always positive. -
Tagalog Pala: an Unsurprising Case of Mirativity
Tagalog pala: an unsurprising case of mirativity Scott AnderBois Brown University Similar to many descriptions of miratives cross-linguistically, Schachter & Otanes(1972)’s clas- sic descriptive grammar of Tagalog describes the second position particle pala as “expressing mild surprise at new information, or an unexpected event or situation.” Drawing on recent work on mi- rativity in other languages, however, we show that this characterization needs to be refined in two ways. First, we show that while pala can be used in cases of surprise, pala itself merely encodes the speaker’s sudden revelation with the counterexpectational nature of surprise arising pragmatically or from other aspects of the sentence such as other particles and focus. Second, we present data from imperatives and interrogatives, arguing that this revelation need not concern ‘information’ per se, but rather the illocutionay update the sentence encodes. Finally, we explore the interactions between pala and other elements which express mirativity in some way and/or interact with the mirativity pala expresses. 1. Introduction Like many languages of the Philippines, Tagalog has a prominent set of discourse particles which express a variety of different evidential, attitudinal, illocutionary, and discourse-related meanings. Morphosyntactically, these particles have long been known to be second-position clitics, with a number of authors having explored fine-grained details of their distribution, rela- tive order, and the interaction of this with different types of sentences (e.g. Schachter & Otanes (1972), Billings & Konopasky(2003) Anderson(2005), Billings(2005) Kaufman(2010)). With a few recent exceptions, however, comparatively little has been said about the semantics/prag- matics of these different elements beyond Schachter & Otanes(1972)’s pioneering work (which is quite detailed given their broad scope of their work). -
4.1 RANK of a MATRIX Rank List Given Matrix M, the Following Are Equal
page 1 of Section 4.1 CHAPTER 4 MATRICES CONTINUED SECTION 4.1 RANK OF A MATRIX rank list Given matrix M, the following are equal: (1) maximal number of ind cols (i.e., dim of the col space of M) (2) maximal number of ind rows (i.e., dim of the row space of M) (3) number of cols with pivots in the echelon form of M (4) number of nonzero rows in the echelon form of M You know that (1) = (3) and (2) = (4) from Section 3.1. To see that (3) = (4) just stare at some echelon forms. This one number (that all four things equal) is called the rank of M. As a special case, a zero matrix is said to have rank 0. how row ops affect rank Row ops don't change the rank because they don't change the max number of ind cols or rows. example 1 12-10 24-20 has rank 1 (maximally one ind col, by inspection) 48-40 000 []000 has rank 0 example 2 2540 0001 LetM= -2 1 -1 0 21170 To find the rank of M, use row ops R3=R1+R3 R4=-R1+R4 R2ØR3 R4=-R2+R4 2540 0630 to get the unreduced echelon form 0001 0000 Cols 1,2,4 have pivots. So the rank of M is 3. how the rank is limited by the size of the matrix IfAis7≈4then its rank is either 0 (if it's the zero matrix), 1, 2, 3 or 4. The rank can't be 5 or larger because there can't be 5 ind cols when there are only 4 cols to begin with. -
8.3 Positive Definite Matrices
8.3. Positive Definite Matrices 433 Exercise 8.2.25 Show that every 2 2 orthog- [Hint: If a2 + b2 = 1, then a = cos θ and b = sinθ for × cos θ sinθ some angle θ.] onal matrix has the form − or sinθ cosθ cos θ sin θ Exercise 8.2.26 Use Theorem 8.2.5 to show that every for some angle θ. sinθ cosθ symmetric matrix is orthogonally diagonalizable. − 8.3 Positive Definite Matrices All the eigenvalues of any symmetric matrix are real; this section is about the case in which the eigenvalues are positive. These matrices, which arise whenever optimization (maximum and minimum) problems are encountered, have countless applications throughout science and engineering. They also arise in statistics (for example, in factor analysis used in the social sciences) and in geometry (see Section 8.9). We will encounter them again in Chapter 10 when describing all inner products in Rn. Definition 8.5 Positive Definite Matrices A square matrix is called positive definite if it is symmetric and all its eigenvalues λ are positive, that is λ > 0. Because these matrices are symmetric, the principal axes theorem plays a central role in the theory. Theorem 8.3.1 If A is positive definite, then it is invertible and det A > 0. Proof. If A is n n and the eigenvalues are λ1, λ2, ..., λn, then det A = λ1λ2 λn > 0 by the principal axes theorem (or× the corollary to Theorem 8.2.5). ··· If x is a column in Rn and A is any real n n matrix, we view the 1 1 matrix xT Ax as a real number. -
A Cross-Linguistic Study of Grammatical Organization
Complement Clauses and Complementation Systems: A Cross-Linguistic Study of Grammatical Organization Dissertation zur Erlangung des akademischen Grades eines Doctor philosophiae (Dr. phil.) vorgelegt dem Rat der Philosophischen Fakultät der Friedrich-Schiller-Universität Jena von Karsten Schmidtke-Bode, M.A. geb. am 26.06.1981 in Eisenach Gutachter: 1. Prof. Dr. Holger Diessel (Friedrich-Schiller-Universität Jena) 2. Prof. Dr. Volker Gast (Friedrich-Schiller-Universität Jena) 3. Prof. Dr. Martin Haspelmath (MPI für Evolutionäre Anthropologie Leipzig) Tag der mündlichen Prüfung: 16.12.2014 Contents Abbreviations and notational conventions iii 1 Introduction 1 2 The phenomenon of complementation 7 2.1 Introduction 7 2.2 Argument status 9 2.2.1 Complement clauses and argument-structure typology 10 2.2.2 On the notion of ‘argument’ 21 2.3 On the notion of ‘clause’ 26 2.3.1 Complementation constructions as biclausal units 27 2.3.2 The internal structure of clauses 31 2.4 The semantic content of complement clauses 34 2.5 Environments of complementation 36 2.5.1 Predicate classes as environments of complementation 37 2.5.2 Environments studied in the present work 39 3 Data and methods 48 3.1 Sampling and sources of information 48 3.2 Selection and nature of the data points 53 3.3 Storage and analysis of the data 59 4 The internal structure of complementation patterns 62 4.1 Introduction 62 4.2 The morphological status of the predicate 64 4.2.1 Nominalization 65 4.2.2 Converbs 68 4.2.3 Participles 70 4.2.4 Bare verb stems 71 4.2.5 Other dependent -
Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation