2 Rank and Matrix Algebra

Total Page:16

File Type:pdf, Size:1020Kb

2 Rank and Matrix Algebra 2 Rank and Matrix Algebra 2.1 Rank In our introduction to systems of linear equations we mentioned that a system can have no solutions, a unique solution, or infinitely many solutions. In this section we're going to introduce an invariant1 of matrices, and when this invariant is computed for the matrix of coefficients of a system of linear equations, it can give us information about the number of solutions to this system. Rank.2The rank of a matrix A is the number of pivots in rref(A), the reduced row- echelon form of A. We denote the rank of A by rank(A). Example 8. Consider the matrix 2 4 3 2 −1 3 6 5 4 3 −1 7 A = 6 7 : 4 −2 −2 −1 2 5 11 6 4 1 This matrix has reduced row-echelon form 2 1 0 0 1 3 6 0 1 0 −3 7 rref(A) = 6 7 ; 4 0 0 1 2 5 0 0 0 0 so rank(A) = 3, since rref(A) has 3 pivots. Note. Here are some of the connections between the rank of a matrix and the number of solutions to a system of linear equations. Suppose we have a system of n linear equations in m variables, and that the n × m matrix A is the coefficient matrix of this system. Then 1. We have rank(A) ≤ n and rank(A) ≤ m, because there cannot be more pivots than there are rows, nor than there are columns. 2. If the system of equations is inconsistent, then rank(A) < n. This is because in row- reducing an inconsistent system we eventually have a row of zeros, augmented by a nonzero solution. This row of zeros can't have a pivot, so the number of pivots is at most n − 1. 1An invariant of a mathematical object is a property that doesn't change when we apply certain oper- ations to the object. In this case, the rank of a matrix is invariant under elementary row operations. 2We'll soon give a different, much better definition of the rank of a matrix. This is a probationary definition from which we'll soon move on. 10 3. If the system has exactly one solution, then rank(A) = m. If rank(A) < m, then the system would have a free variable, meaning that if there is a solution, then there are infinitely many solutions. 4. If the system has infinitely many solutions, then rank(A) < m, because a system with infinitely many solutions must have a free variable. We can combine these observations with some remarks made in the previous section to conclude that if m = n (i.e., we have the same number of equations as variables), then the system has a unique solution if and only if rref(A) = In, the n × n identity matrix. 2.2 Matrix Algebra So far we've only seen matrices used to represent systems of linear equations. They have another (perhaps more important) use as representatives of linear transformations. For example, matrices can be used to model the evolution of the electorate between election cycles. Suppose that each election cycle, 90% of voters who were members of the Republican party remain in the Republican party, while the other 10% join the Democratic party. At the same time, suppose 5% of Democrats become Republicans, while the other 95% remain in the Democratic party. If each party has 500,000 voters in one election cycle, the following matrix equation computes the number of voters each party will have in the next cycle: A 0:95 0:10 500; 000 525; 000 = : (6) 0:05 0:90 500; 000 475; 000 Once we learn how to interpret (6) we'll see that in the next election there should be 525,000 Democratic voters and 475,000 Republican voters. In this equation the matrix A represents the transformation from one election cycle to the next, and this is the idea we'd like to focus on soon: matrices as transformations. Before we can really consider matrices as representing transformations, we'll need to first make sense of expressions such as (6). To this end, we'll make clear some vocabulary sur- rounding matrices and then discuss addition of matrices. Size of a matrix. The size of a matrix is given by the number of rows and columns it has. A matrix with m rows and n columns is said to be \m-by-n", written m × n. We occasionally call a matrix with only one row a row matrix and call a matrix with just one column a column matrix; we will call either of these types of matrices vectors. Matrices which have the same number of rows and columns are called square matrices. Example 9. Below we have a 2 × 3 matrix, a row matrix, a column matrix, and a square matrix, respectively: 263 2 0 −1 1 3 2 7 −3 ; 0 9 1 ; 2 ; −2 2 −4 : 1 0 9 4 5 4 5 3 4 −8 8 11 We use lowercase letters with double subscripts to identify the entries of a matrix. For example, if the square matrix above is A, then a23 = −4. Now we define addition of matrices. This operation is easy, but it is important to remember that addition of matrices is only defined when the matrices involved have the same size. That is, we can't add a 2 × 3 matrix to a 5 × 4 matrix, but we can add two 2 × 3 matrices together, and we do so entry-wise. For instance, 2 7 −3 1 2 3 3 9 0 + = : 1 0 9 4 5 6 5 5 15 We can similarly define subtraction entry-wise, subject to the same restriction on size: 2 7 −3 1 2 3 1 5 −6 − = : 1 0 9 4 5 6 −3 −5 3 Because scalar addition is commutative and we defined matrix addition entry-wise, ma- trix addition is commutative. That is, if A and B are matrices of the same size, then A + B = B + A. We point this out because it will not be true for multiplication. After addition and subtraction, it makes sense to define matrix multiplication. We're going to defer most of this discussion a little longer, but we will at least define the product Ax, where A is a matrix and x is a vector. As with addition and subtraction, multiplication has a size condition: Size condition for the product Ax. If A is an n×m matrix and x is a vector with k component, then the product Ax only exists if m = k. That is, the number of columns of A must equal the number of components of x. If this condition is met, then Ax will be a vector with n components. To prepare ourselves for the definition of Ax, we first define a product between vectors with the same number of components, called the dot product. Dot product. Suppose we have vectors v = hv1; v2; : : : ; vni and w = hw1; w2; : : : ; wni: Then the dot product of v and w is a scalar given by v · w = v1w1 + v2w2 + ··· + vnwn: Notice that w · v = v · w. Example 10. We have 2 2 3 6 3 7 1 1 −1 1 · 6 7 = 1 · 2 + 1 · 3 + (−1) · 6 + 1 · 1 = 0: 4 6 5 1 12 Though it's not necessary, we often write the first vector in a dot product as a row vector and the second as a column vector. We now define the product Ax in terms of dot products. The product Ax. Let A be an n×m matrix and let x be a vector with m components. If the rows of A are the vectors w1;:::; wn, each with m components, then 2 3 w1 · x 6 . 7 Ax = 4 . 5 : wn · x Example 11. Compute each of the following products, or write that the product does not exist: 2 0 3 1 0 8 1 0 8 24 0:95 0:10 525; 000 (a) 6 (b) (c) : 4 −3 2 4 5 4 −3 2 −12 0:05 0:90 475; 000 3 (Solution) 2 0 3 1 0 8 1 · 0 + 0 · 6 + 8 · 3 24 (a) 6 = = : 4 −3 2 4 5 4 · 0 + (−3) · 6 + 2 · 3 −12 3 (b) Since the matrix in the product is 2 × 3 and the vector has two components (rather than three), this product is not defined. (c) This represents another election cycle in the two-party example at the beginning of this section: 0:95 0:10 525; 000 0:95 · 525; 000 + 0:10 · 475; 000 546; 250 = = : 0:05 0:90 475; 000 0:05 · 525; 000 + 0:90 · 475; 000 453; 750 So after another cycle, there will be 546,250 Democratic voters and 453,750 Republican voters. ♦ One helpful property of matrix multiplication is that it plays nicely with vector and matrix addition, as well as scalar multiplication: A(x + y) = Ax + Ay; (A + B)x = Ax + Bx; and A(kx) = k(Ax); for all n × m matrices A and B, vectors x and y with m components, and scalars k. 13 We'll conclude with an example of arguably the most important matrix product in this course. Consider the product 2 1 2 3 3 2 x 3 2 x + 2y + 3z 3 4 2 −1 6 5 4 y 5 = 4 2x − y + 6z 5 : 3 0 −4 z 3x − 4z The entries in the resulting vector look at lot like the left side of a system of linear equations.
Recommended publications
  • ADDITIVE MAPS on RANK K BIVECTORS 1. Introduction. Let N 2 Be an Integer and Let F Be a Field. We Denote by M N(F) the Algeb
    Electronic Journal of Linear Algebra, ISSN 1081-3810 A publication of the International Linear Algebra Society Volume 36, pp. 847-856, December 2020. ADDITIVE MAPS ON RANK K BIVECTORS∗ WAI LEONG CHOOIy AND KIAM HEONG KWAy V2 Abstract. Let U and V be linear spaces over fields F and K, respectively, such that dim U = n > 2 and jFj > 3. Let U n−1 V2 V2 be the second exterior power of U. Fixing an even integer k satisfying 2 6 k 6 n, it is shown that a map : U! V satisfies (u + v) = (u) + (v) for all rank k bivectors u; v 2 V2 U if and only if is an additive map. Examples showing the indispensability of the assumption on k are given. Key words. Additive maps, Second exterior powers, Bivectors, Ranks, Alternate matrices. AMS subject classifications. 15A03, 15A04, 15A75, 15A86. 1. Introduction. Let n > 2 be an integer and let F be a field. We denote by Mn(F) the algebra of n×n matrices over F. Given a nonempty subset S of Mn(F), a map : Mn(F) ! Mn(F) is called commuting on S (respectively, additive on S) if (A)A = A (A) for all A 2 S (respectively, (A + B) = (A) + (B) for all A; B 2 S). In 2012, using Breˇsar'sresult [1, Theorem A], Franca [3] characterized commuting additive maps on invertible (respectively, singular) matrices of Mn(F). He continued to characterize commuting additive maps on rank k matrices of Mn(F) in [4], where 2 6 k 6 n is a fixed integer, and commuting additive maps on rank one matrices of Mn(F) in [5].
    [Show full text]
  • Do Killingâ•Fiyano Tensors Form a Lie Algebra?
    University of Massachusetts Amherst ScholarWorks@UMass Amherst Physics Department Faculty Publication Series Physics 2007 Do Killing–Yano tensors form a Lie algebra? David Kastor University of Massachusetts - Amherst, [email protected] Sourya Ray University of Massachusetts - Amherst Jennie Traschen University of Massachusetts - Amherst, [email protected] Follow this and additional works at: https://scholarworks.umass.edu/physics_faculty_pubs Part of the Physics Commons Recommended Citation Kastor, David; Ray, Sourya; and Traschen, Jennie, "Do Killing–Yano tensors form a Lie algebra?" (2007). Classical and Quantum Gravity. 1235. Retrieved from https://scholarworks.umass.edu/physics_faculty_pubs/1235 This Article is brought to you for free and open access by the Physics at ScholarWorks@UMass Amherst. It has been accepted for inclusion in Physics Department Faculty Publication Series by an authorized administrator of ScholarWorks@UMass Amherst. For more information, please contact [email protected]. Do Killing-Yano tensors form a Lie algebra? David Kastor, Sourya Ray and Jennie Traschen Department of Physics University of Massachusetts Amherst, MA 01003 ABSTRACT Killing-Yano tensors are natural generalizations of Killing vec- tors. We investigate whether Killing-Yano tensors form a graded Lie algebra with respect to the Schouten-Nijenhuis bracket. We find that this proposition does not hold in general, but that it arXiv:0705.0535v1 [hep-th] 3 May 2007 does hold for constant curvature spacetimes. We also show that Minkowski and (anti)-deSitter spacetimes have the maximal num- ber of Killing-Yano tensors of each rank and that the algebras of these tensors under the SN bracket are relatively simple exten- sions of the Poincare and (A)dS symmetry algebras.
    [Show full text]
  • What's in a Name? the Matrix As an Introduction to Mathematics
    St. John Fisher College Fisher Digital Publications Mathematical and Computing Sciences Faculty/Staff Publications Mathematical and Computing Sciences 9-2008 What's in a Name? The Matrix as an Introduction to Mathematics Kris H. Green St. John Fisher College, [email protected] Follow this and additional works at: https://fisherpub.sjfc.edu/math_facpub Part of the Mathematics Commons How has open access to Fisher Digital Publications benefited ou?y Publication Information Green, Kris H. (2008). "What's in a Name? The Matrix as an Introduction to Mathematics." Math Horizons 16.1, 18-21. Please note that the Publication Information provides general citation information and may not be appropriate for your discipline. To receive help in creating a citation based on your discipline, please visit http://libguides.sjfc.edu/citations. This document is posted at https://fisherpub.sjfc.edu/math_facpub/12 and is brought to you for free and open access by Fisher Digital Publications at St. John Fisher College. For more information, please contact [email protected]. What's in a Name? The Matrix as an Introduction to Mathematics Abstract In lieu of an abstract, here is the article's first paragraph: In my classes on the nature of scientific thought, I have often used the movie The Matrix to illustrate the nature of evidence and how it shapes the reality we perceive (or think we perceive). As a mathematician, I usually field questions elatedr to the movie whenever the subject of linear algebra arises, since this field is the study of matrices and their properties. So it is natural to ask, why does the movie title reference a mathematical object? Disciplines Mathematics Comments Article copyright 2008 by Math Horizons.
    [Show full text]
  • The Matrix As an Introduction to Mathematics
    St. John Fisher College Fisher Digital Publications Mathematical and Computing Sciences Faculty/Staff Publications Mathematical and Computing Sciences 2012 What's in a Name? The Matrix as an Introduction to Mathematics Kris H. Green St. John Fisher College, [email protected] Follow this and additional works at: https://fisherpub.sjfc.edu/math_facpub Part of the Mathematics Commons How has open access to Fisher Digital Publications benefited ou?y Publication Information Green, Kris H. (2012). "What's in a Name? The Matrix as an Introduction to Mathematics." Mathematics in Popular Culture: Essays on Appearances in Film, Fiction, Games, Television and Other Media , 44-54. Please note that the Publication Information provides general citation information and may not be appropriate for your discipline. To receive help in creating a citation based on your discipline, please visit http://libguides.sjfc.edu/citations. This document is posted at https://fisherpub.sjfc.edu/math_facpub/18 and is brought to you for free and open access by Fisher Digital Publications at St. John Fisher College. For more information, please contact [email protected]. What's in a Name? The Matrix as an Introduction to Mathematics Abstract In my classes on the nature of scientific thought, I have often used the movie The Matrix (1999) to illustrate how evidence shapes the reality we perceive (or think we perceive). As a mathematician and self-confessed science fiction fan, I usually field questionselated r to the movie whenever the subject of linear algebra arises, since this field is the study of matrices and their properties. So it is natural to ask, why does the movie title reference a mathematical object? Of course, there are many possible explanations for this, each of which probably contributed a little to the naming decision.
    [Show full text]
  • Preserivng Pieces of Information in a Given Order in HRR and GA$ C$
    Proceedings of the Federated Conference on ISBN 978-83-60810-22-4 Computer Science and Information Systems pp. 213–220 Preserivng pieces of information in a given order in HRR and GAc Agnieszka Patyk-Ło´nska Abstract—Geometric Analogues of Holographic Reduced Rep- Secondly, this method of encoding will not detect similarity resentations (GA HRR or GAc—the continuous version of between eye and yeye. discrete GA described in [16]) employ role-filler binding based A quantum-like attempt to tackle the problem of information on geometric products. Atomic objects are real-valued vectors in n-dimensional Euclidean space and complex statements belong to ordering was made in [1]—a version of semantic analysis, a hierarchy of multivectors. A property of GAc and HRR studied reformulated in terms of a Hilbert-space problem, is compared here is the ability to store pieces of information in a given order with structures known from quantum mechanics. In particular, by means of trajectory association. We describe results of an an LSA matrix representation [1], [10] is rewritten by the experiment: finding the alignment of items in a sequence without means of quantum notation. Geometric algebra has also been the precise knowledge of trajectory vectors. Index Terms—distributed representations, geometric algebra, used extensively in quantum mechanics ([2], [4], [3]) and so HRR, BSC, word order, trajectory associations, bag of words. there seems to be a natural connection between LSA and GAc, which is the ground for fututre work on the problem I. INTRODUCTION of preserving pieces of information in a given order. VER the years several attempts have been made to As far as convolutions are concerned, the most interesting O preserve the order in which the objects are to be remem- approach to remembering information in a given order has bered with the help of binding and superposition.
    [Show full text]
  • Math Object Identifiers – Towards Research Data in Mathematics
    Math Object Identifiers – Towards Research Data in Mathematics Michael Kohlhase Computer Science, FAU Erlangen-N¨urnberg Abstract. We propose to develop a system of “Math Object Identi- fiers” (MOIs: digital object identifiers for mathematical concepts, ob- jects, and models) and a process of registering them. These envisioned MOIs constitute a very lightweight form of semantic annotation that can support many knowledge-based workflows in mathematics, e.g. clas- sification of articles via the objects mentioned or object-based search. In essence MOIs are an enabling technology for Linked Open Data for mathematics and thus makes (parts of) the mathematical literature into mathematical research data. 1 Introduction The last years have seen a surge in interest in scaling computer support in scientific research by preserving, making accessible, and managing research data. For most subjects, research data consist in measurement or simulation data about the objects of study, ranging from subatomic particles via weather systems to galaxy clusters. Mathematics has largely been left untouched by this trend, since the objects of study – mathematical concepts, objects, and models – are by and large ab- stract and their properties and relations apply whole classes of objects. There are some exceptions to this, concrete integer sequences, finite groups, or ℓ-functions and modular form are collected and catalogued in mathematical data bases like the OEIS (Online Encyclopedia of Integer Sequences) [Inc; Slo12], the GAP Group libraries [GAP, Chap. 50], or the LMFDB (ℓ-Functions and Modular Forms Data Base) [LMFDB; Cre16]. Abstract mathematical structures like groups, manifolds, or probability dis- tributions can formalized – usually by definitions – in logical systems, and their relations expressed in form of theorems which can be proved in the logical sys- tems as well.
    [Show full text]
  • 8 Rank of a Matrix
    8 Rank of a matrix We already know how to figure out that the collection (v1;:::; vk) is linearly dependent or not if each n vj 2 R . Recall that we need to form the matrix [v1 j ::: j vk] with the given vectors as columns and see whether the row echelon form of this matrix has any free variables. If these vectors linearly independent (this is an abuse of language, the correct phrase would be \if the collection composed of these vectors is linearly independent"), then, due to the theorems we proved, their span is a subspace of Rn of dimension k, and this collection is a basis of this subspace. Now, what if they are linearly dependent? Still, their span will be a subspace of Rn, but what is the dimension and what is a basis? Naively, I can answer this question by looking at these vectors one by one. In particular, if v1 =6 0 then I form B = (v1) (I know that one nonzero vector is linearly independent). Next, I add v2 to this collection. If the collection (v1; v2) is linearly dependent (which I know how to check), I drop v2 and take v3. If independent then I form B = (v1; v2) and add now third vector v3. This procedure will lead to the vectors that form a basis of the span of my original collection, and their number will be the dimension. Can I do it in a different, more efficient way? The answer if \yes." As a side result we'll get one of the most important facts of the basic linear algebra.
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • History of Algebra and Its Implications for Teaching
    Maggio: History of Algebra and its Implications for Teaching History of Algebra and its Implications for Teaching Jaime Maggio Fall 2020 MA 398 Senior Seminar Mentor: Dr.Loth Published by DigitalCommons@SHU, 2021 1 Academic Festival, Event 31 [2021] Abstract Algebra can be described as a branch of mathematics concerned with finding the values of unknown quantities (letters and other general sym- bols) defined by the equations that they satisfy. Algebraic problems have survived in mathematical writings of the Egyptians and Babylonians. The ancient Greeks also contributed to the development of algebraic concepts. In this paper, we will discuss historically famous mathematicians from all over the world along with their key mathematical contributions. Mathe- matical proofs of ancient and modern discoveries will be presented. We will then consider the impacts of incorporating history into the teaching of mathematics courses as an educational technique. 1 https://digitalcommons.sacredheart.edu/acadfest/2021/all/31 2 Maggio: History of Algebra and its Implications for Teaching 1 Introduction In order to understand the way algebra is the way it is today, it is important to understand how it came about starting with its ancient origins. In a mod- ern sense, algebra can be described as a branch of mathematics concerned with finding the values of unknown quantities defined by the equations that they sat- isfy. Algebraic problems have survived in mathematical writings of the Egyp- tians and Babylonians. The ancient Greeks also contributed to the development of algebraic concepts, but these concepts had a heavier focus on geometry [1]. The combination of all of the discoveries of these great mathematicians shaped the way algebra is taught today.
    [Show full text]
  • 2.2 Kernel and Range of a Linear Transformation
    2.2 Kernel and Range of a Linear Transformation Performance Criteria: 2. (c) Determine whether a given vector is in the kernel or range of a linear trans- formation. Describe the kernel and range of a linear transformation. (d) Determine whether a transformation is one-to-one; determine whether a transformation is onto. When working with transformations T : Rm → Rn in Math 341, you found that any linear transformation can be represented by multiplication by a matrix. At some point after that you were introduced to the concepts of the null space and column space of a matrix. In this section we present the analogous ideas for general vector spaces. Definition 2.4: Let V and W be vector spaces, and let T : V → W be a transformation. We will call V the domain of T , and W is the codomain of T . Definition 2.5: Let V and W be vector spaces, and let T : V → W be a linear transformation. • The set of all vectors v ∈ V for which T v = 0 is a subspace of V . It is called the kernel of T , And we will denote it by ker(T ). • The set of all vectors w ∈ W such that w = T v for some v ∈ V is called the range of T . It is a subspace of W , and is denoted ran(T ). It is worth making a few comments about the above: • The kernel and range “belong to” the transformation, not the vector spaces V and W . If we had another linear transformation S : V → W , it would most likely have a different kernel and range.
    [Show full text]
  • 23. Kernel, Rank, Range
    23. Kernel, Rank, Range We now study linear transformations in more detail. First, we establish some important vocabulary. The range of a linear transformation f : V ! W is the set of vectors the linear transformation maps to. This set is also often called the image of f, written ran(f) = Im(f) = L(V ) = fL(v)jv 2 V g ⊂ W: The domain of a linear transformation is often called the pre-image of f. We can also talk about the pre-image of any subset of vectors U 2 W : L−1(U) = fv 2 V jL(v) 2 Ug ⊂ V: A linear transformation f is one-to-one if for any x 6= y 2 V , f(x) 6= f(y). In other words, different vector in V always map to different vectors in W . One-to-one transformations are also known as injective transformations. Notice that injectivity is a condition on the pre-image of f. A linear transformation f is onto if for every w 2 W , there exists an x 2 V such that f(x) = w. In other words, every vector in W is the image of some vector in V . An onto transformation is also known as an surjective transformation. Notice that surjectivity is a condition on the image of f. 1 Suppose L : V ! W is not injective. Then we can find v1 6= v2 such that Lv1 = Lv2. Then v1 − v2 6= 0, but L(v1 − v2) = 0: Definition Let L : V ! W be a linear transformation. The set of all vectors v such that Lv = 0W is called the kernel of L: ker L = fv 2 V jLv = 0g: 1 The notions of one-to-one and onto can be generalized to arbitrary functions on sets.
    [Show full text]
  • 1.3 Matrices and Matrix Operations 1.3.1 De…Nitions and Notation Matrices Are Yet Another Mathematical Object
    20 CHAPTER 1. SYSTEMS OF LINEAR EQUATIONS AND MATRICES 1.3 Matrices and Matrix Operations 1.3.1 De…nitions and Notation Matrices are yet another mathematical object. Learning about matrices means learning what they are, how they are represented, the types of operations which can be performed on them, their properties and …nally their applications. De…nition 50 (matrix) 1. A matrix is a rectangular array of numbers. in which not only the value of the number is important but also its position in the array. 2. The numbers in the array are called the entries of the matrix. 3. The size of the matrix is described by the number of its rows and columns (always in this order). An m n matrix is a matrix which has m rows and n columns. 4. The elements (or the entries) of a matrix are generally enclosed in brack- ets, double-subscripting is used to index the elements. The …rst subscript always denote the row position, the second denotes the column position. For example a11 a12 ::: a1n a21 a22 ::: a2n A = 2 ::: ::: ::: ::: 3 (1.5) 6 ::: ::: ::: ::: 7 6 7 6 am1 am2 ::: amn 7 6 7 =4 [aij] , i = 1; 2; :::; m, j =5 1; 2; :::; n (1.6) Enclosing the general element aij in square brackets is another way of representing a matrix A . 5. When m = n , the matrix is said to be a square matrix. 6. The main diagonal in a square matrix contains the elements a11; a22; a33; ::: 7. A matrix is said to be upper triangular if all its entries below the main diagonal are 0.
    [Show full text]