Cofactor Expansions the Determinant Is a Scalar Characteristic of a Square

Total Page:16

File Type:pdf, Size:1020Kb

Cofactor Expansions the Determinant Is a Scalar Characteristic of a Square Linear Algebra Grinshpan Cofactor expansions The determinant is a scalar characteristic of a square matrix. The determinant of A is denoted by det A or by jAj: For 1 × 1; 2 × 2; and 3 × 3 matrices the formulas are: [ ] det a = a; [ ] a b det = ad − bc; c d 2 3 a b c det 4d e f5 = aei + bfg + cdh − ceg − bdi − afh: g h i What is the pattern? One way to describe it is as follows. i+j Let A be n × n: To each entry aij of A; associate the sign (−1) and the submatrix Aij obtained from A by deleting the i-th row and the j-th column. i+j The number (−1) jAijj is the cofactor of A corresponding to the position (i; j): For every row index i; i+1 i+2 i+n jAj = ai1(−1) jAi1j + ai2(−1) jAi2j + ::: + ain(−1) jAinj: This is the Laplace or cofactor expansion of jAj along the i-th row. For every column index j; 1+j 2+j n+j jAj = a1j(−1) jA1jj + a2j(−1) jA2jj + ::: + anj(−1) jAnjj: This is the cofactor expansion of jAj along the j-th column. The above expansions give jAj in terms of determinants of one order less, so any of them may be used for calculating jAj: The reduction may take several steps. 1 2 3 2 3 1 3 1 2 4 5 6 = −4 + 5 − 6 = −4(18 − 24) + 5(9 − 21) − 6(8 − 14) = 0 8 9 7 9 7 8 7 8 9 1 2 3 5 6 2 3 2 3 4 5 6 = +1 − 4 + 7 = (45 − 48) − 4(18 − 24) + 7(12 − 15) = 0 8 9 8 9 5 6 7 8 9 For every 2 × 2 matrix, the determinant is plus or minus the area of the parallelogram bounded by its columns (rows). For every 3 × 3 matrix, the determinant is plus or minus the volume of the parallelepiped bounded by its columns (rows). For instance, the columns of the preceding determinant are coplanar, so enclose zero volume. Properties of the determinant The determinant may be defined by the following four properties. (1) Normalization: the determinant of an identity matrix is 1 [ ] 1 0 det = 1 0 1 (2) Antisymmetry: a column interchange changes the sign of the determinant 2 3 2 3 a ∗ d d ∗ a det 4b ∗ e5 = − det 4e ∗ b5 c ∗ f f ∗ c (3) Column additivity [ ] [ ] [ ] ∗ a + b ∗ a ∗ b det = det + det ∗ c + d ∗ c ∗ d (4) Column homogeneity 2 3 2 3 ∗ 3a ∗ ∗ a ∗ det 4∗ 3b ∗5 = 3 det 4∗ b ∗5 ∗ 3c ∗ ∗ c ∗ Immediate consequences: A matrix with two identical columns has zero determinant. Adding a multiple of one column to another does not change the determinant. A matrix with a zero column has zero determinant. The determinant of a permutation matrix is 1: The above set of rules allows us to calculate any determinant. For instance, using additivity and homogeneity in each column, we have 2 5 2 5 0 5 = + 3 7 0 7 3 7 2 5 2 0 0 5 0 0 = + + + 0 0 0 7 3 0 3 7 1 1 1 0 0 1 0 0 = 2 · 5 + 2 · 7 + 3 · 5 + 3 · 7 0 0 0 1 1 0 1 1 = 2 · 5 · 0 + 2 · 7 · (+1) + 3 · 5 · (−1) + 3 · 7 · 0 = −1: Similarly, 1 2 3 1 0 0 1 0 0 0 2 0 0 0 3 0 0 3 0 2 0 4 5 6 = 0 5 0 + 0 0 6 + 4 0 0 + 4 0 0 + 0 5 0 + 0 0 6 7 8 9 0 0 9 0 8 0 0 0 9 0 8 0 7 0 0 7 0 0 =1 · 5 · 9 · (+1) + 1 · 6 · 8 · (−1) + 2 · 4 · 9 · (−1) + 3 · 4 · 8 · (+1) + 3 · 5 · 7 · (−1) + 2 · 6 · 7 · (+1) = 0: We see that the determinant is a sum of signed products. In each product the entries are taken one per column and one per row. The sign is given by the determinant of the corresponding permutation matrix. For an n × n matrix, there are n! summands. For any permutation matrix P; we have det P = det P >: Therefore, transposition does not change the determinant, det A = det A>; and all of the above properties equally well apply to rows. A row interchange changes the sign of the determinant [ ] [ ] 0 1 1 0 det = − det 1 0 0 1 The determinant is row additive [ ] [ ] [ ] 1 2 0 1 1 1 det = det + det a b a b a b The determinant is row homogeneous [ ] [ ] 1 3 1 3 det = 2 det 2 4 1 2 A matrix with two identical rows has zero determinant. Adding a multiple of one row to another does not change the determinant. A matrix with a zero row has zero determinant. Here are some more practical observations. The determinant of a diagonal matrix is the product of its diagonal entries. The same is true for triangular matrices. 2 3 2 3 2 3 a 0 0 a ∗ ∗ a 0 0 det 40 b 05 = det 40 b ∗5 = det 4∗ b 05 = abc 0 0 c 0 0 c ∗ ∗ c The determinant and row operations We now know what happens to the determinant under elementary row operations. Row interchange: the determinant changes sign 2 3 2 3 1 2 3 4 5 6 det 44 5 65 = − det 41 2 35 ∗ ∗ ∗ ∗ ∗ ∗ Scale a row: scale the determinant 2 3 2 3 ∗ ∗ ∗ ∗ ∗ ∗ det 43a 3b 3c5 = 3 det 4a b c5 ∗ ∗ ∗ ∗ ∗ ∗ Add a multiple of one row to another: no change 2 3 2 3 a b c a b c det 4d e f5 = det 4d + 2a e + 2b f + 2c5 ∗ ∗ ∗ ∗ ∗ ∗ Based on these properties are several useful conclusions. If A is reduced to an echelon form U using k row interchanges and no scalings, then det A = (−1)k det U = (−1)k × (product of diagonal entries of U): So a matrix is invertible if and only if its determinant is nonzero. A matrix with linearly dependent rows and columns has zero determinant. 2 3 1 2 3 det 44 5 65 = 0 7 8 9 The determinant is multiplicative: det(AB) = (det A)(det B): If A is invertible, then det(A−1) = 1= det A: Exercises Calculate the determinants 3 1 1 1 1 1 1 1 2 0 0 3 0 1 1 x 1 1 1 2 4 1 3 1 1 1 −1 1 −1 0 2 3 0 1 0 1 ; ; 1 x 1 ; ; 1 3 9 ; : 1 1 3 1 1 1 −1 −1 0 3 2 0 1 1 0 1 1 x 1 5 25 1 1 1 3 1 −1 −1 1 3 0 0 2 .
Recommended publications
  • Determinant Notes
    68 UNIT FIVE DETERMINANTS 5.1 INTRODUCTION In unit one the determinant of a 2× 2 matrix was introduced and used in the evaluation of a cross product. In this chapter we extend the definition of a determinant to any size square matrix. The determinant has a variety of applications. The value of the determinant of a square matrix A can be used to determine whether A is invertible or noninvertible. An explicit formula for A–1 exists that involves the determinant of A. Some systems of linear equations have solutions that can be expressed in terms of determinants. 5.2 DEFINITION OF THE DETERMINANT a11 a12 Recall that in chapter one the determinant of the 2× 2 matrix A = was a21 a22 defined to be the number a11a22 − a12 a21 and that the notation det (A) or A was used to represent the determinant of A. For any given n × n matrix A = a , the notation A [ ij ]n×n ij will be used to denote the (n −1)× (n −1) submatrix obtained from A by deleting the ith row and the jth column of A. The determinant of any size square matrix A = a is [ ij ]n×n defined recursively as follows. Definition of the Determinant Let A = a be an n × n matrix. [ ij ]n×n (1) If n = 1, that is A = [a11], then we define det (A) = a11 . n 1k+ (2) If na>=1, we define det(A) ∑(-1)11kk det(A ) k=1 Example If A = []5 , then by part (1) of the definition of the determinant, det (A) = 5.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • MATH 304 Linear Algebra Lecture 24: Scalar Product. Vectors: Geometric Approach
    MATH 304 Linear Algebra Lecture 24: Scalar product. Vectors: geometric approach B A B′ A′ A vector is represented by a directed segment. • Directed segment is drawn as an arrow. • Different arrows represent the same vector if • they are of the same length and direction. Vectors: geometric approach v B A v B′ A′ −→AB denotes the vector represented by the arrow with tip at B and tail at A. −→AA is called the zero vector and denoted 0. Vectors: geometric approach v B − A v B′ A′ If v = −→AB then −→BA is called the negative vector of v and denoted v. − Vector addition Given vectors a and b, their sum a + b is defined by the rule −→AB + −→BC = −→AC. That is, choose points A, B, C so that −→AB = a and −→BC = b. Then a + b = −→AC. B b a C a b + B′ b A a C ′ a + b A′ The difference of the two vectors is defined as a b = a + ( b). − − b a b − a Properties of vector addition: a + b = b + a (commutative law) (a + b) + c = a + (b + c) (associative law) a + 0 = 0 + a = a a + ( a) = ( a) + a = 0 − − Let −→AB = a. Then a + 0 = −→AB + −→BB = −→AB = a, a + ( a) = −→AB + −→BA = −→AA = 0. − Let −→AB = a, −→BC = b, and −→CD = c. Then (a + b) + c = (−→AB + −→BC) + −→CD = −→AC + −→CD = −→AD, a + (b + c) = −→AB + (−→BC + −→CD) = −→AB + −→BD = −→AD. Parallelogram rule Let −→AB = a, −→BC = b, −−→AB′ = b, and −−→B′C ′ = a. Then a + b = −→AC, b + a = −−→AC ′.
    [Show full text]
  • Inner Product Spaces
    CHAPTER 6 Woman teaching geometry, from a fourteenth-century edition of Euclid’s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition and scalar multiplication) of R2 and R3. We ignored other important features, such as the notions of length and angle. These ideas are embedded in the concept we now investigate, inner products. Our standing assumptions are as follows: 6.1 Notation F, V F denotes R or C. V denotes a vector space over F. LEARNING OBJECTIVES FOR THIS CHAPTER Cauchy–Schwarz Inequality Gram–Schmidt Procedure linear functionals on inner product spaces calculating minimum distance to a subspace Linear Algebra Done Right, third edition, by Sheldon Axler 164 CHAPTER 6 Inner Product Spaces 6.A Inner Products and Norms Inner Products To motivate the concept of inner prod- 2 3 x1 , x 2 uct, think of vectors in R and R as x arrows with initial point at the origin. x R2 R3 H L The length of a vector in or is called the norm of x, denoted x . 2 k k Thus for x .x1; x2/ R , we have The length of this vector x is p D2 2 2 x x1 x2 . p 2 2 x1 x2 . k k D C 3 C Similarly, if x .x1; x2; x3/ R , p 2D 2 2 2 then x x1 x2 x3 . k k D C C Even though we cannot draw pictures in higher dimensions, the gener- n n alization to R is obvious: we define the norm of x .x1; : : : ; xn/ R D 2 by p 2 2 x x1 xn : k k D C C The norm is not linear on Rn.
    [Show full text]
  • Properties of Transpose
    3.2, 3.3 Inverting Matrices P. Danziger Properties of Transpose Transpose has higher precedence than multiplica- tion and addition, so T T T T AB = A B and A + B = A + B As opposed to the bracketed expressions (AB)T and (A + B)T Example 1 1 2 1 1 0 1 Let A = ! and B = !. 2 5 2 1 1 0 Find ABT , and (AB)T . T 1 1 1 2 1 1 0 1 1 2 1 0 1 ABT = ! ! = ! 0 1 2 5 2 1 1 0 2 5 2 B C @ 1 0 A 2 3 = ! 4 7 Whereas (AB)T is undefined. 1 3.2, 3.3 Inverting Matrices P. Danziger Theorem 2 (Properties of Transpose) Given ma- trices A and B so that the operations can be pre- formed 1. (AT )T = A 2. (A + B)T = AT + BT and (A B)T = AT BT − − 3. (kA)T = kAT 4. (AB)T = BT AT 2 3.2, 3.3 Inverting Matrices P. Danziger Matrix Algebra Theorem 3 (Algebraic Properties of Matrix Multiplication) 1. (k + `)A = kA + `A (Distributivity of scalar multiplication I) 2. k(A + B) = kA + kB (Distributivity of scalar multiplication II) 3. A(B + C) = AB + AC (Distributivity of matrix multiplication) 4. A(BC) = (AB)C (Associativity of matrix mul- tiplication) 5. A + B = B + A (Commutativity of matrix ad- dition) 6. (A + B) + C = A + (B + C) (Associativity of matrix addition) 7. k(AB) = A(kB) (Commutativity of Scalar Mul- tiplication) 3 3.2, 3.3 Inverting Matrices P. Danziger The matrix 0 is the identity of matrix addition.
    [Show full text]
  • Clifford Algebra with Mathematica
    Clifford Algebra with Mathematica J.L. ARAGON´ G. ARAGON-CAMARASA Universidad Nacional Aut´onoma de M´exico University of Glasgow Centro de F´ısica Aplicada School of Computing Science y Tecnolog´ıa Avanzada Sir Alwyn William Building, Apartado Postal 1-1010, 76000 Quer´etaro Glasgow, G12 8QQ Scotland MEXICO UNITED KINGDOM [email protected] [email protected] G. ARAGON-GONZ´ ALEZ´ M.A. RODRIGUEZ-ANDRADE´ Universidad Aut´onoma Metropolitana Instituto Polit´ecnico Nacional Unidad Azcapotzalco Departamento de Matem´aticas, ESFM San Pablo 180, Colonia Reynosa-Tamaulipas, UP Adolfo L´opez Mateos, 02200 D.F. M´exico Edificio 9. 07300 D.F. M´exico MEXICO MEXICO [email protected] [email protected] Abstract: The Clifford algebra of a n-dimensional Euclidean vector space provides a general language comprising vectors, complex numbers, quaternions, Grassman algebra, Pauli and Dirac matrices. In this work, we present an introduction to the main ideas of Clifford algebra, with the main goal to develop a package for Clifford algebra calculations for the computer algebra program Mathematica.∗ The Clifford algebra package is thus a powerful tool since it allows the manipulation of all Clifford mathematical objects. The package also provides a visualization tool for elements of Clifford Algebra in the 3-dimensional space. clifford.m is available from https://github.com/jlaragonvera/Geometric-Algebra Key–Words: Clifford Algebras, Geometric Algebra, Mathematica Software. 1 Introduction Mathematica, resulting in a package for doing Clif- ford algebra computations. There exists some other The importance of Clifford algebra was recognized packages and specialized programs for doing Clif- for the first time in quantum field theory.
    [Show full text]
  • Eigenvalues and Eigenvectors
    Section 5.1 Eigenvalues and Eigenvectors At this point in the class, we understand that the operation of matrix multiplication is very different from scalar multiplication. Matrix multiplication is an operation that finds the product of a pair of matrices, whereas scalar multiplication finds the product of a scalar with a matrix. More to the point, scalar multiplication of a scalar k with a matrix X is an easy operation to perform; indeed there's a simple algorithm: multiply each entry of X by the scalar k. Matrix multiplication on the other hand cannot be expressed using a simple algorithm. Indeed, the operation is extremely complex: to multiply a pair of n×n matrices, it takes about n2 operations, a hugely inefficient process. Thus it is surprising that, for any square matrix A, certain matrix multiplications \collapse" (in a sense) into scalar multiplication. As an example, consider the matrices ( ) ( ) ( ) 3 2 −4 10 A = ; w = ; x = : 3 4 −1 15 The product Aw is not particularly interesting: ( ) −14 Aw = ; −16 indeed, there appears to be very little connection between A, w, and Aw. On the other hand, an we see an interesting phenomenon when we calculate Ax: ( ) ( ) 60 10 Ax = = 6 = 6x: 90 15 Forgetting the calculation in the middle, we have Ax = 6x; in a sense, the matrix multiplication Ax collapsed into scalar multiplication 6x. We will see below that the phenomenon described here is more common than one might think. Identifying products like Ax = 6x will allow us to extract a great deal of useful information about A; accordingly, the rest of the section will be devoted to this phenomenon, which we give a name in the next definition.
    [Show full text]
  • Functions of Multivector Variables Plos One, 2015; 10(3): E0116943-1-E0116943-21
    PUBLISHED VERSION James M. Chappell, Azhar Iqbal, Lachlan J. Gunn, Derek Abbott Functions of multivector variables PLoS One, 2015; 10(3): e0116943-1-e0116943-21 © 2015 Chappell et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited PERMISSIONS http://creativecommons.org/licenses/by/4.0/ http://hdl.handle.net/2440/95069 RESEARCH ARTICLE Functions of Multivector Variables James M. Chappell*, Azhar Iqbal, Lachlan J. Gunn, Derek Abbott School of Electrical and Electronic Engineering, University of Adelaide, Adelaide, South Australia, Australia * [email protected] Abstract As is well known, the common elementary functions defined over the real numbers can be generalized to act not only over the complex number field but also over the skew (non-com- muting) field of the quaternions. In this paper, we detail a number of elementary functions extended to act over the skew field of Clifford multivectors, in both two and three dimen- sions. Complex numbers, quaternions and Cartesian vectors can be described by the vari- ous components within a Clifford multivector and from our results we are able to demonstrate new inter-relationships between these algebraic systems. One key relation- ship that we discover is that a complex number raised to a vector power produces a quater- nion thus combining these systems within a single equation. We also find a single formula that produces the square root, amplitude and inverse of a multivector over one, two and three dimensions.
    [Show full text]
  • SCALAR PRODUCTS, NORMS and METRIC SPACES 1. Definitions Below, “Real Vector Space” Means a Vector Space V Whose Field Of
    SCALAR PRODUCTS, NORMS AND METRIC SPACES 1. Definitions Below, \real vector space" means a vector space V whose field of scalars is R, the real numbers. The main example for MATH 411 is V = Rn. Also, keep in mind that \0" is a many splendored symbol, with meaning depending on context. It could for example mean the number zero, or the zero vector in a vector space. Definition 1.1. A scalar product is a function which associates to each pair of vectors x; y from a real vector space V a real number, < x; y >, such that the following hold for all x; y; z in V and α in R: (1) < x; x > ≥ 0, and < x; x > = 0 if and only if x = 0. (2) < x; y > = < y; x >. (3) < x + y; z > = < x; z > + < y; z >. (4) < αx; y > = α < x; y >. n The dot product is defined for vectors in R as x · y = x1y1 + ··· + xnyn. The dot product is an example of a scalar product (and this is the only scalar product we will need in MATH 411). Definition 1.2. A norm on a real vector space V is a function which associates to every vector x in V a real number, jjxjj, such that the following hold for every x in V and every α in R: (1) jjxjj ≥ 0, and jjxjj = 0 if and only if x = 0. (2) jjαxjj = jαjjjxjj. (3) (Triangle Inequality for norm) jjx + yjj ≤ jjxjj + jjyjj. p The standard Euclidean norm on Rn is defined by jjxjj = x · x.
    [Show full text]
  • MATH 304 Linear Algebra Lecture 6: Transpose of a Matrix. Determinants. Transpose of a Matrix
    MATH 304 Linear Algebra Lecture 6: Transpose of a matrix. Determinants. Transpose of a matrix Definition. Given a matrix A, the transpose of A, denoted AT , is the matrix whose rows are columns of A (and whose columns are rows of A). That is, T if A = (aij ) then A = (bij ), where bij = aji . T 1 4 1 2 3 Examples. = 2 5, 4 5 6 3 6 T 7 T 4 7 4 7 8 = (7, 8, 9), = . 7 0 7 0 9 Properties of transposes: • (AT )T = A • (A + B)T = AT + BT • (rA)T = rAT • (AB)T = BT AT T T T T • (A1A2 ... Ak ) = Ak ... A2 A1 • (A−1)T = (AT )−1 Definition. A square matrix A is said to be symmetric if AT = A. For example, any diagonal matrix is symmetric. Proposition For any square matrix A the matrices B = AAT and C = A + AT are symmetric. Proof: BT = (AAT )T = (AT )T AT = AAT = B, C T = (A + AT )T = AT + (AT )T = AT + A = C. Determinants Determinant is a scalar assigned to each square matrix. Notation. The determinant of a matrix A = (aij )1≤i,j≤n is denoted det A or a11 a12 ... a1n a a ... a 21 22 2n . .. an1 an2 ... ann Principal property: det A =0 if and only if the matrix A is singular. Definition in low dimensions a b Definition. det (a) = a, = ad − bc, c d a a a 11 12 13 a21 a22 a23 = a11a22a33 + a12a23a31 + a13a21a32− a31 a32 a33 −a13a22a31 − a12a21a33 − a11a23a32. * ∗ ∗ ∗ * ∗ ∗ ∗ * +: ∗ * ∗ , ∗ ∗ * , * ∗ ∗ . ∗ ∗ * * ∗ ∗ ∗ * ∗ ∗ ∗ * ∗ * ∗ * ∗ ∗ − : ∗ * ∗ , * ∗ ∗ , ∗ ∗ * . * ∗ ∗ ∗ ∗ * ∗ * ∗ Examples: 2×2 matrices 1 0 3 0 = 1, = − 12, 0 1 0 −4 −2 5 7 0 = − 6, = 14, 0 3 5 2 0 −1 0 0 = 1, = 0, 1 0 4 1 −1 3 2 1 = 0, = 0.
    [Show full text]
  • MATHEMATICS INTRO for PHYSICS 1. Scalars and Vectors
    MATHEMATICS INTRO FOR PHYSICS YORK UNIVERSITY 1. Scalars and Vectors 1.1. Preliminaries. Many physics variables use the real-number system to pro- vide a quantitative description. Temperature is one such example: on the Celsius scale we encounter positive and negative real numbers, which allow ordering: we can always compare two temperature readings and state whether one is higher than the other (or lower, or equal). A complete description requires specification of a unit: this distinguishes physics from mathematics. When we adopt the inter- national unit system SI (meters, kilograms, seconds, Ampere) we can avoid writing down the unit. However, when doing so we should make clear that we are using SI to avoid confusion. In a physics measurement one is limited by the accuracy of equipment. Thus, an answer like x = 4 m has to be treated carefully. In mathematics the implication would be that x equals to the integer number 4 (the integers are embedded in the real numbers). No physics measurement can be as precise as that. Therefore, we quote results to a certain number of significant digits. A meaningful physics statement can be: x = 4:00 m - which would imply that we are certain about the result being above x = 3:995 m and below x = 4:005 m. When calculating answers to physics problems in this course pay attention to how many significant digits are provided by the input, use one more digit for intermediate calculations, and quote your final answer to no more than this one additional digit. Ideally, you will round the result to the correct number of significant digits.
    [Show full text]
  • 17. Inner Product Spaces Definition 17.1. Let V Be a Real Vector Space
    17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function h ; i: V × V −! R; which is • symmetric, that is hu; vi = hv; ui: • bilinear, that is linear (in both factors): hλu, vi = λhu; vi; for all scalars λ and hu1 + u2; vi = hu1; vi + hu2; vi; for all vectors u1, u2 and v. • positive that is hv; vi ≥ 0: • non-degenerate that is if hu; vi = 0 for every v 2 V then u = 0. We say that V is a real inner product space. The associated quadratic form is the function Q: V −! R; defined by Q(v) = hv; vi: Example 17.2. Let A 2 Mn;n(R) be a real matrix. We can define a function n n h ; i: R × R −! R; by the rule hu; vi = utAv: The basic rules of matrix multiplication imply that this function is bi- linear. Note that the entries of A are given by t aij = eiAej = hei; eji: In particular, it is symmetric if and only if A is symmetric that is At = A. It is non-degenerate if and only if A is invertible, that is A has rank n. Positivity is a little harder to characterise. Perhaps the canonical example is to take A = In. In this case if t P u = (r1; r2; : : : ; rn) and v = (s1; s2; : : : ; sn) then u Inv = risi. Note 1 P 2 that if we take u = v then we get ri . The square root of this is the Euclidean distance.
    [Show full text]