Math 75 Linear Algebra Class Notes
Prof. Erich Holtmann
For use with Elementary Linear Algebra, 7th ed., Larson
Revised 21-Nov-2015
p. i
Contents Chapter 1: Systems of Linear Equations 1 1.1 Introduction to Systems of Equations. 1 1.2 Gaussian Elimination and Gauss-Jordan Elimination. 9 1.3 Applications of Systems of Linear Equations. 15 Chapter 2: Matrices. 19 2.1 Operations with Matrices. 19 2.2 Properties of Matrix Operations. 23 2.3 The Inverse of a Matrix. 27 2.4 Elementary Matrices. 31 2.5 Applications of Matrix Operations. 37 Chapter 3: Determinants. 41 3.1 The Determinant of a Matrix. 41 3.2 Determinants and Elementary Operations. 45 3.3 Properties of Determinants. 51 3.4 Applications of Determinants. 55 Chapter 4: Vector Spaces. 61 8.1 Complex Numbers (Optional). 61 8.2 Conjugates and Division of Complex Numbers (Optional). 67 4.1 Vectors in Rn. 71 4.2 Vector Spaces. 77 4.3 Subspaces of Vector Spaces. 82 4.4 Spanning Sets and Linear Independence. 86 4.5 Basis and Dimension. 97 4.6 Rank of a Matrix and Systems of Linear Equations. 105 4.7 Coordinates and Change of Basis. 117 Chapter 5: Inner Product Spaces. 121 5.1 Length and Dot Product in Rn. 121 5.2 Inner Product Spaces. 129 5.3 Orthogonal Bases: Gram-Schmidt Process. 137 5.4 Mathematical Models and Least Squares Analysis (Optional). 145 5.5 Applications of Inner Product Spaces (Optional). 151 8.3 Polar Form and De Moivre's Theorem. (Optional) 157 8.4 Complex Vector Spaces and Inner Products. 163
p. i Chapter 6: Linear Transformations. 169 6.1 Introduction to Linear Transformations. 169 6.2 The Kernel and Range of a Linear Transformation. 175 6.3 Matrices for Linear Transformations. 183 6.4 Transition Matrices and Similarity. 191 6.5 Applications of Linear Transformations. 193 Chapter 7: Eigenvalues and Eigenvectors. 201 7.1 Eigenvalues and Eigenvectors. 201 7.2 Diagonalization. 209 7.3 Symmetric Matrices and Orthogonal Diagonalization. 215 7.4 Applications of Eigenvalues and Eigenvectors. 223 8.5 Unitary and Hermitian Spaces. 223
p. ii Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations.
Chapter 1: Systems of Linear Equations
1.1 Introduction to Systems of Equations. Objective: Recognize and solve mn systems of linear equations by hand using Gaussian elimination and back-substitution.
a1x1 + a2x2 + … + anxn = b is a linear equation in standard form in n variables xi.
The first nonzero coefficient ai is the leading coefficient. The constant term is b.
Compare to the familiar forms of linear equation s in two variables y = mx + b and x = a.
Example: Linear and Nonlinear Equations
2 (sin ) x1 – 4x2 = e sin x1 + 2x2 – 3x3 = 0 Linear Nonlinear
An mn system of linear equations is a set of m linear equations in n unknowns.
Example: Systems of Two Equations in Two Variables
Solve and graph each 22 system.
a. 2x – y = 1 b. 2x – y = 1 c. 2x – y = 1 x + y = 5 –4x + 2y = –2 –4x + 2y = 5
For a system of linear equations, exactly one of the following is true. 1) The system has exactly one solution (consistent, nonsingular system). 2) The system has infinitely many solutions (consistent, singular system). Use a free parameter or free variable (or several free parameters) to represent the solution set. 3) The system has no solution (inconsistent, singular system).
p. 1 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations. To solve mn systems of linear equations (when m and n are large) we use a procedure called Gaussian elimination to find an equivalent system of equations in row-echelon form. Then we use back-substitution to solve for each variable.
Row-echelon form means that the leading coefficients of 1 (called “pivots”) and the zero terms below them form a stair-step pattern. You could walk downstairs from the top left. You might have to move more than one column to the right to reach the next step, but you never have to step down more than one row at a time.
1x1 5x2 2x3 1x4 3x5 7 1x 1 Row-echelon form 3 1x4 4x5 8
1x5 9
1x1 5x2 2x3 1x4 3x5 7 1x 1 Row-echelon form 3 0 0 0 0
1x1 5x2 2x3 1x4 3x5 7 1x 3x 2x 1 Not row-echelon form 3 4 5 1x3 2x4 4x5 8
1x4 3x5 9
The goal of Gaussian elimination is to find an equivalent system that is in row-echelon form. The three operations you can use during Gaussian elimination are 1) Swap the order of two equations. 2) Multiply an equation on both sides by a non-zero constant. 3) Add a multiple of one equation to another equation. In Gaussian elimination, you start with Equation 1 (the first equation of your mn system). 1) Find the leading coefficient in the current equation. (Sometimes you need to swap equations in this step.) 2) Eliminate the coefficients of the corresponding variable in all of the equations below the current equation. 3) Move down to the next equation and go back to Step 1. Repeat until you run out of equations or you run out of variables. Solve using back-substitution: solve the last equation for the leading variable, then substitute into the preceding (i.e. second-to-last) equation and solve for its leading variable, then substitute into the preceding equation and solve for its leading variable, etc. Variables that are not leading variables are free parameters, and we often set them equal to t, s, ….
p. 2 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations. Examples: Gaussian Elimination and Back-Substitution on 33 Systems of Linear Equations.
p. 3 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations.
p. 4 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations.
p. 5 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations.
p. 6 Chapter 1: Systems of Linear Equations 1.1 Introduction to Systems of Equations. Example: Chemistry Application
Write and solve a system of linear equations for the chemical reaction
(x1)CH4 + (x2)O2 (x3)CO2 + (x4)H2O
Solution: write a separate equation for each element, showing the balance of that element.
C: 1x1 + 0x2 = 1x3 + 0x4 so 1x1 + 0x2 – 1x3 + 0x4 = 0 H: 4x1 + 0x2 = 0x3 + 2x4 so 4x1 + 0x2 + 0x3 – 2x4 = 0 O: 0x1 + 2x2 = 2x3 + 1x4 so 0x1 + 2x2 – 2x3 – 1x4 = 0
p. 7
Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination.
1.2 Gaussian Elimination and Gauss-Jordan Elimination. Objective: Use matrices and Gauss-Jordan elimination to solve mn systems of linear equations by hand and with software. Objective: Use matrices and Gaussian elimination with back-substitution to solve mn systems of linear equations by hand and with software. Use row-echelon form or reduced row-echelon form to determine the number of solutions of a homogeneous system of linear equations, and (if applicable) the number of free parameters.
A matrix is a rectangular array of numbers, called matrix aaaa entries, arranged in horizontal rows and vertical columns. 11 12 13 1n aaaa Matrices are denoted by capital letters; matrix entries are 21 22 23 2n denoted by lowercase letters with two indices. In a given A 31 32 33 aaaa3n matrix entry a , the first index i is the row, and the second ij index j is the column. The entries a , a , a , … compose the 11 22 33 aaaa main diagonal. If m = n then A is called a square matrix. mmm321 mn
11 1 12 2 13 3 1 nn bxaxaxaxa 1 bxaxaxaxa A linear system 21 1 22 2 23 3 2 nn 2
m m m 332211 mn bxaxaxaxa mn can represented either by a coefficient matrix A and a column vector b
1211 13 aaaa 1n b1 aaaa b 2221 23 2n 2
A 3231 33 aaaa 3n and b b3 mmm 331 aaaa mn bm or by an augmented matrix M, which I will sometimes write as [A | b]
11 12 13 n baaaa 11 baaaa 21 22 23 n 22
M 31 32 33 n baaaa 33 mmm 331 mn baaaa m (The book, Mathematica, and the calculator do not display the dotted vertical line.) To create M in Mathematica from A and b, type m=Join[a,b,2]
To create M on the TI-89 from A and b, type
Matrixaugment(AB.M
p. 9 Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination. In a similar manner to that used for an mn system of linear equations, we can use a Gaussian elimination on the coefficient side A of the augmented matrix [A | b] to find an equivalent augmented matrix [U | c] in row-echelon form. Then we use back-substitution to solve for each variable. U is called an upper triangular matrix because all non-zero entries are on or above the main diagonal.
row-echelon form row-echelon form not row-echelon form
1 5 2 1 3 1 5 2 1 3 1 5 2 1 3 0 0 1 0 5 0 0 1 0 5 0 0 1 3 2 0 0 0 1 4 0 0 0 0 0 0 0 1 2 4 0 0 0 0 2 0 0 0 0 0 0 0 0 1 3
Example: Use Gaussian elimination and back-substitution to solve. The three elementary row operations you can use during Gaussian elimination are 1) Swap the two rows. 2) Multiply a row by a non-zero constant. 3) Add a multiple of one row to another row.
2 1 4 6x 4y 3z 8 6 4 3 8 1 3 2 3 3x 2y 1z 3 3 2 1 3 0 1 3 2 2 1x 1y 1z 2 1 1 1 2 0 0 1 2 z 2 3 3 y 2 z 2 so y 2 2 (2) 1 2 1 4 4 2 1 x 3 y 2 z 3 so x 3 3 (1) 2 (2) 1 Steps:
p. 10 Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination. Instead of using back-substitution, we can take the row-echelon form [U | c] and eliminate the coefficients above the pivots by adding multiples of the pivot rows. The result [R | d] is called reduced row-echelon form.
reduced row-echelon form reduced row-echelon form
1 5 0 0 11 1 5 0 1 3 0 0 1 0 5 0 0 1 3 2 0 0 0 1 4 0 0 0 0 4 0 0 0 0 2 0 0 0 0 3
Example: Use Gauss-Jordan elimination to solve.
2 1 4 6x 4y 3z 8 6 4 3 8 1 3 2 3 1 0 0 1 x 1 3x 2y 1z 3 3 2 1 3 0 1 3 2 0 1 0 1 y 1 2 1x 1y 1z 2 1 1 1 2 0 0 1 2 0 0 1 2 z 2
p. 11 Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination. Example Using software to find the Reduced Row-Echelon Form (Exercise 1.2 #39)
Mathematica Go to the Palettes Menu and open the Basic Math Assistant. Under Basic Commands, open the Matrix Tab. Type a= and click Use the Add Row and Add Column buttons to expand the matrix, so you can enter the augmented matrix
(If you make the matrix to large, use to remove rows and columns.) Press (or on the number pad) when you are finished entering the augmented matrix. Another way to enter the augmented matrix is to type (or download) a={{1,-1,2,2,6,6},{3,-2,4,4,12,14},{0,1,-1,-1,-3,-3},{2,- 2,4,5,15,10},{2,-2,4,4,13,13}} MatrixForm[a] You can download this matrix from http://holtmann75.pbworks.com. Click on Electronic Data Sets and open 1133110878_323834.zip/DataSets/Mathematica/0102039.nb (Ch. 01, Section 02, Problem 039). Notice that A and a are different variables! User-defined variables should always begin with a lower-case letter, because Mathematica’s built-in fuctions and commands begin with capital letters. (For example, N and C are already defined by Mathematica.) On the Basic Math Assistant palette, click on MatrixForm RowReduce and type a so you have MatrixForm[RowReduce[a]]
Converting back to a system of linear equations, we have x1 = 2 x4 = 5 x2 = 2 x5 = 1 x3 = 3
p. 12 Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination. TI-89: Type Data/Matrix Editor New... Type: Matrix Folder: main Variable: A (A is above . If you type = instead of a, use . If a already exists, then use Open... on the previous screen instead of New....) Use and the arrow keys to type in the coefficient matrix
(If you need to insert or delete a row or column, use ) When you are finished, press . Another way to enter the matrix is to type (from the home screen) 1,1,2,2,6,6 3,2,4,4,12,14 0,1,1,1,3,3 2,2,4,5,15,10 2,2,4,4,13,13 .A
After the matrix is entered, type A
Matrix rref( A
Converting back to a system of linear equations, we have x1 = 2 x2 = 2 x3 = 3 x4 = 5 x5 = 1
p. 13 Chapter 1: Systems of Linear Equations 1.2 Gaussian Elimination and Gauss-Jordan Elimination. Mathematica can help you perform row operations. If your augmented matrix is in a, then a[[2]] is row 2 of the matrix. To view a in the usual format, type MatrixForm[a] To swap rows 2 and 3 of a, type a[[{2,3}]]=a[[{3,2}]] To multiply row 2 of a by 7, type a[[2]]=7*a[[2]] To add 7 times row 2 to row 1, type a[[1]]=a[[1]]+7*a[[2]]
Mathematica performs operations in the order that you type them in, not as they appear on the screen. If you go back and edit your work, you can use Evaluate Notebook under the Evaluation menu to recalculate the notebook in the order on the screen. The TI-89 can also help you perform row operations. If your augmented matrix is in a, then … To swap rows 2 and 3 of a and store the result in a1, type Matrix Row opsrowSwap(A,2,3,).A1 To multiply row 2 of a1 by 7 and store the result in a2, type Matrix Row opsmRow(7,A1,2).A2 To add 7 times row 2 of a2 to row 1 and store the result in a3, type Matrix Row opsmRowAdd(7,A2,1,3).A3 1 2 1 2 1 2 About notation: and are matrices, but is a determinant (Ch. 3). 3 4 3 4 3 4 Theorem 1.1. Every homogeneous (constants on right-hand side are all zeroes) system of linear equations is consistent. The number of free parameters in the solution set is the number of variables minus the number of pivots (leading coefficients). If there are zero free parameters, then there is exactly one solution.
1 1 5 0 1 1 5 2 4 0 2 5 3 2 1 0 0 5 1 1 1 0 Examples: 1 0 2 0 6 15 4 5 4 0 1 21 9 5 5 0 8 20 7 7 3 0 2 1 1 0 5 20 28 14 12 0 0 0 15 3 21 0 1 0 2 0 Solutions: 0 1 3 0 0 0 0 0
p. 14 Chapter 1: Systems of Linear Equations 1.3 Applications of Systems of Linear Equations.
1.3 Applications of Systems of Linear Equations. Objective: Set up and solve a system of equations to fit a polynomial function to a set of data points. Objective: Set up and solve a system of equations to represent a network.
Polynomial Curve Fitting.
Given m data points (t1, y1), (t2, y2), …, (tm, ym). We want to find a polynomial of degree m–1 that passes through these points. 2 m 2 m–1 p(t) = c0 + c1t + c2t + … + cmt Notice that yj = c0 + c1(tj)+ c2(tj) + … + cm–1(tj)
m1 1 1 tt1 )( c0 y1 m1 1 tt )( c y This produces an mm linear system: 2 2 1 = 2 m1 1 m ttm )( cm1 ym that you can solve using Gauss-Jordan elimination.
Example 1.3#7
t 21012 y 28 0860
p. 15 Chapter 1: Systems of Linear Equations 1.3 Applications of Systems of Linear Equations. Example 1.3#9
t 2006 2007 2008
y 5 7 12
Network Analysis: write a system of linear equations using Kirchoff’s Laws. 1) Flow into each node (also called vertex) equals flow out. 2) In an electrical network, the sum of the products IR (I = current and R = resistance) around any closed path of edges (lines) is equal to the total voltage in the loop from the batteries.
A resistor is represented by the symbol Resistance is measured in ohms (). 1 k = 1000
A battery is represented by the symbol If the current flows through the battery from the short line (–) to the long line I (+), then the voltage is positive.
Current is measure in amps (A). 1 mA = 0.001 A.
Flow into a node is positive. Flow out of a node is negative. To write a system of equations,
1) Pick a direction (at random) for each current I. 2) For each node, write an equation for the current input and output. 3) For each loop, write the V = IR equation.
p. 16 Chapter 1: Systems of Linear Equations 1.3 Applications of Systems of Linear Equations. Example: 1.3.#32. Solve for the currents.
Example: 1.3.29. The figure shows traffic flow in vehicles/hour through a network of streets.
a) Solve for x1, x2, x3, and x4. b) Find the traffic flow when x4 = 0. c) Find the traffic flow when x4 = 100.
p. 17
Chapter 2: Matrices. 2.1 Operations with Matrices.
Chapter 2: Matrices.
2.1 Operations with Matrices. Objective: Determine whether or not two matrices are equal. Objective: Add and subtract matrices and multiply a matrix by a scalar. Objective: Multiply two matrices. Objective: Write solutions to a system of linear equations in column vector notation. Objective: Partition a matrix and perform block multiplication.
11 12 aaa1n 21 22 aaa2n Three ways to represent the same matrix are A = [aij] = . mm31 aaamn
Two matrices A = [aij] and B = [bij] are equal if and only if they have the same dimensions or order (mn) and aij = bij for all 1 i m and 1 j n.
Comment on logic: An example of “P if Q” (also written P Q) is “It is cloudy if it is raining.” An example of “Q only if P” (i.e. “if Q then P,” also written Q P) is “It is raining only if it is cloudy.” “R iff S” is shorthand for “R if and only if S” (also written R S). R iff S means that R is equivalent to S.
To add (or subtract) two matrices A = [aij] and B = [bij] that have the same dimensions, add (or subtract) corresponding matrix entries.
A + B = [aij + bij] A – B = [aij – bij] The sum (and difference) of two matrices with different dimensions is undefined. To multiply a matrix by a scalar (number), multiply each entry by that scalar.
cA = [caij]
Examples:
421 341 50 Let A = , B = , and D = . Find A + B, A – B, A + D, and 4A. 103 231 31
p. 19 Chapter 2: Matrices. 2.1 Operations with Matrices. In Mathematica, use a+b, a–b, and 4*a or 4a to add matrices, subtract matrices, and multiply 50 scalars c by matrices. Remember, you can enter D = from the Palettes Menu::Basic 31 Math Assistant::Basic Commands::Matrix Tab.
On the TI-89, use ab, a b, and ca or 4a to add matrices A and B, subtract matrices,
and multiply scalars c by matrices. Remember, you can enter D = from the home screen
as [0,5;-1,3].D or from Data/Matrix Editor. (Be sure to use Type: Matrix)
Matrix multiplication is defined in a much more complex manner. For a 11 system, we can write ax = b. We want a definition of matrix multiplication that allows us to write an mn system
11 1 12 2 13 3 1 nn bxaxaxaxa 1 bxaxaxaxa 21 1 22 2 23 3 2 nn 2 as Ax = b,
m m m 332211 mn bxaxaxaxa mn
a11 a12 a1n a a a where A is the coefficient matrix A 21 22 2n , am1 am2 amn
x1 b1 x b and x and b are column matrices (or column vectors): x 2 and b 2 . xn bm
Observe that each row of A was multiplied by the column x to give the corresponding row of b.
If A = [aij] is an mn matrix and B = [bij] is an np matrix, then the product AB is an mp matrix A = [cij] where
n [cij] = aik bkj = ai1b1j + ai2b2j + ai3b3j + … ainbnj k 1 If the column dimension of A does not match the row dimension of B, then the product is undefined
p. 20 Chapter 2: Matrices. 2.1 Operations with Matrices. Examples 2 3 3 2 4 1 0 3 3 2 1 5 2 5
In Mathematica, use a.b to multiply matrices.(Do not use a*b)
On the TI-89, use ab multiply matrices.
Example: Write solutions to a system of linear equations in column vector notation.
x1 2 1 2 1 3 0 x 1 Solve 2 = = t 1 1 0 1 0 x3 1 0 1 1 2 0 x4 1
Block multiplication on partitioned matrices works whenever the dimensions are OK.
Examples 0 1 0 0 0 1 0 0 A A 11 12 with A11 A12 and A 1 0 0 0 1 0 0 0 A21 A22 0 0 2 1 A21 0 0 A22 2 1
1 2 0 1 2 0 B11 B12 1 1 0 B B 1 1 0 B 11 12 with 4 3 1 4 3 1 B21 B22 B21 B22 5 6 2 5 6 2
1 1 0 A B A B A B A B A B 0B A 0 0B Then AB = 11 11 12 21 11 12 12 22 = 11 11 21 11 22 = 1 2 0 A21B11 A22B21 A21B12 A22B22 0B11 A22B21 0 0 A22B22 13 12 0
p. 21 Chapter 2: Matrices. 2.1 Operations with Matrices.
a11 a12 a1n x1 a11x1 a12x2 a1n xn a11 a12 a1n a a a x a x a x a x a a a Ax= 21 22 2n 2 = 21 1 22 2 2n n = 21 x 22 x 2n x 1 2 n am1 am2 amn xn am1 x1 am2 x2 amn xn am1 am2 amn
In block matrix notation, we write [ a1 | a2 | … | an ] = where the ai are
m1 column matrices. The Ax = [ a1 | a2 | … | an ] = a1x1 + a2x2 + … + anxn
Similarly, if B is an lm matrix (so BA is defined), then BA = B[ a1 | a2 | … | an ] = [ Ba1 | Ba2 | … | Ban ], i.e. the columns of BA are Bai. Notice that B is lm, ai is m1, Bai is l1, and BA is ln.
On the other hand, we could partition A into 1m row matrices ri
a11 a12 a1n r1 a11 a12 a1n x1 r1 r1x a a a r a a a x r r x A = 21 22 2n = 2 . Then Ax = 21 22 2n 2 = 2 x = 2 . am1 am2 amn rm am1 am2 amn xn rm rm x
Notice that ri is 1n, x is n1, rix is 11 (i.e. a number), and Ax is m1.
e11 e12 e1m e1 e1 e1 A e e e e e e A If E = 21 22 2m = 2 then EA = 2 A = 2 . el1 el 2 elm e l e l el A
Notice that ei is 1m, A is mn, eiA is 1n, and EA is ln
If e = [ e1 | e2 | … | em], then eA = [ e1 | e2 | … | em] = e1r1 + e2r2 + … + emrm
Notice that e is 1m, A is mn, ei is a number, ri is 1n, and eA is 1n.
p. 22 Chapter 2: Matrices. 2.2 Properties of Matrix Operations.
2.2 Properties of Matrix Operations. Objective: Know and use the properties of matrix operations (matrix addition and subtraction, scalar multiplication, and matrix multiplication), and of the zero and identity matrices. Objective: Know which properties of fields do not hold for matrices (commutativity of matrix multiplication and existence of multiplicative inverses). Objective: Find the transpose of a matrix and know properties of the transpose.
The real numbers R, together with the operations of addition and multiplication, is an example of a mathematical field. (The complex numbers C with addition and multiplication is another example.) Fields and their operations have all of the usual properties.
1) Closure under addition: if a and b R, then a + b R. 2) Addition is associative: (a + b) + c = a + (b + c) 3) Addition is commutative: a + b = b + a 4) Additive identity (zero): R contains 0, which has the property that a + 0 = a for all a. 5) Additive inverses (opposites): every a R. has an opposite –a, such that a + (–a) = 0. We define subtraction a – b as a + (–b). 6) Closure under multiplication: if a and b R, then ab R. 7) Multiplication is associative: (ab)c = a(bc) 8) Multiplication is commutative: ab = ba 9) Multiplication distributes over addition: a(b + c) = ab +ac 10) Multiplicative identity (one): R contains 1, which has the property that 1a = a for all a. 11) Multiplicative inverses (reciprocals): every a R. has an inverse a–1, such that aa–1 = 1. We define division a b as ab–1.
Matrix addition (and subtraction) has all of the usual properties: closure, associativity, commutativity, zero matrices, and opposites. A zero matrix has zero in all entries, but because the matrix can have any dimensions mn, we have many zero matrices 0mn.
The opposite of a matrix [aij] is –[aij] = [–aij]. Multiplication of a scalar by a matrix has all of the usual properties: closure, associativity, commutativity, multiplicative identity (the scalar 1), and distribution. For distribution, we have both c(A + B) = cA + cB and (c + d)A = cA + dA.
Examples: 98 98 98 295 295 000 (–1) = – = , + = = 023 97 97 97 356 356 000
p. 23 Chapter 2: Matrices. 2.2 Properties of Matrix Operations. Multiplication of matrices is closed, is associative, and distributes over matrix addition. The multiplicative identity matrices are square matrices with ones on the main diagonal and zeros 1 0 0 0 1 0 everyplace else: Inn = . If A is mn, then Imm Amn = Amn and Amn Inn = Amn 0 0 1 k We define exponents for square matrices A and positive integers k: A = AAA . Also, we k times define A0 = I. Multiplication of matrices is not commutative in general. Many matrices do not have multiplicative inverses. Division of matrices is undefined.
Examples:
1 0 0 1 0 295 = 0 1 0 = 0 1 356 0 0 1
98 = = 97
1 0 1 0 = = 2 1 2 1 3 1 3 1
Theorem 2.5: For a system of linear equations, exactly one of the following is true. 1) The system has no solution (inconsistent, singular system). 2) The system has exactly one solution (consistent, nonsingular system). 3) The system has infinitely many solutions (consistent, singular system). Use a free parameter or free variable (or several free parameters) to represent the solution set.
Proof using matrix operations:
Given a system of linear equations Ax = b. Exactly one of the following is true: the system has no solution, the system has exactly one solution, or the system has at least two solutions (call them x1 and x2).
p. 24 Chapter 2: Matrices. 2.2 Properties of Matrix Operations.
If the system has two solutions, then Ax1 = b and Ax2 = b so A(x1 – x2) = Ax1 – Ax2 = b – b = 0. Let xh = x1 – x2, so xh is a nonzero solution to the homogenous equation Ax = 0. Then for any scalar t, x1 + txh is a solution of Ax = b because
A(x1 + txh) = Ax1 + tAxh = b+ t 0 = b. Thus, in the last case, the system has infinitely many solutions with a parameter t.
The transpose AT of a matrix A is formed by writing its columns as rows. For example,
a11 a12 a13 a1n a11 a21 a31 am1 a a a a a a a a 21 22 23 2n 12 22 32 m2 T if A a31 a32 a33 a3n then A a13 a23 a33 am3 . am1 am2 am3 amn a1n a2n a3n amn T Equivalently, the transpose of A is formed by writing rows as columns. The ij entry of A is aji.
9 3 7 0 Example: Find the transpose of A = 3 0 8 4 1 5 2 8
Mathematica: to take the transpose of a matrix, either type Transpose[a] (also available on the Basic Math Assistant palette) or type a followed by tr (four separate keystrokes). After the first three keystrokes, you will see atr. After the fourth keystroke, atr will change to aT. Of course, at the end, type
TI-89: AMatrixT
Properties of the transpose: 1) (AT)T = A 2) (A + B)T = AT +BT * 3) (cA) T =cAT 4) (AB) T = BTAT
p. 25 Chapter 2: Matrices. 2.2 Properties of Matrix Operations. Example of Property #4:
Consider A = and B =
The 32 entry of (AB)T is the 23 entry of AB =
which is a21b13 + a22b23 + a23b33
If we look at AT and BT, we much reverse the order of multiplication to obtain row column.
BTAT =
T T T T T The 32 entry B A is b13a21 + b23a22 + b33a23. You can see that (AB) = B A
A matrix M is symmetric iff MT = M. 10 6 5 Example: M = is symmetric. Prove that AAT is symmetric for any matrix A. 6 4 2 5 2 5 Proof:
p. 26 Chapter 2: Matrices. 2.3 The Inverse of a Matrix.
2.3 The Inverse of a Matrix. Objective: Find the inverse of a matrix (if it exists) by Gauss-Jordan elimination. Objective: Use properties of inverse matrices. Objective: Use an inverse matrix to solve a system of linear equations.
A square nn matrix is invertible (or nonsingular) when there exists an nn matrix A–1 such that
–1 –1 AA = Inn and A A = Inn
–1 where Inn is the identity matrix. A is called the (multiplicative) inverse of A. A matrix that does not have an inverse is called singular (or noninvertible). Nonsquare matrices do not have inverses.
Theorem 2.7: If A is an invertible matrix, then the inverse is unique.
Proof: let B and C be inverses of A. Then BA = I an AC = I. So B = BI = B(AC) = (BA)C = IC =C. Therefore, B = C and the inverse of A is unique.
53 Example: Find the inverse of A = . 21
xx 1211 01 Solution: We need to solve the system AX = I, or = , xx 2221 10 xxxx 053153 which gives four equations 11 21 12 22 11 21 12 xxxx 22 121021
To solve these four equations, we take the reduced row echelon forms
153 201 053 501 –1 52 and , so A = X = 021 110 121 310 31
Since the row operations performed to find the reduced row echelon form depend only on the 53 coefficient part of the augmented matrix , we could solve all four equations 21 simultaneously by using a doubly augmented matrix
0153 5201 –1 [ A | I ] = = [ I | A ] 1021 3110
To create the doubly augmented matrix in Mathematica from A, type m=Join[a, IdentityMatrix[2],2] To create the doubly augmented matrix on the TI-89 from A, type
p. 27 Chapter 2: Matrices. 2.3 The Inverse of a Matrix. Matrixaugment(A,Matrixidentity(2)).M We used IdentityMatrix[2] and identity(2) because we wanted a 22 matrix.
To find the inverse of an nn matrix A by Gauss-Jordan elimination, find the reduced row echelon form of th n2n augmented matrix [ A | I ]. If the nn block on the left can be reduced to –1 I, then the nn block on the right is A * [ A | I ] [ I | A–1]
If the nn block on the left cannnot be reduced to I, then A is not invertible.
1 1 1 Example: Invert A = using Gauss-Jordan elimination. 1 2 2 1 2 3
1 1 1 Example: Invert A = using Gauss-Jordan elimination. 1 2 2 2 3 3
a b Example: Invert M = using Gauss-Jordan elimination. c d
You can also use software. First clear the variables using Clear[a,b,c,d] or Clear a-z, then type Inverse[m] or M^1
(also on Basic Math Assistant palette)
p. 28 Chapter 2: Matrices. 2.3 The Inverse of a Matrix. Theorem 2.8 Properties of Inverse Matrices If A is an invertible matrix, k is a positive integer, and c is a nonzero scalar, then A–1, Ak, cA, and AT are invertible, and 1) (A–1) –1 = A * 2) (Ak)–1 = (A–1)k –1 1 –1 3) (cA) = c A 4) (AT)–1 = (A–1)T
Proof:
1)
2) Proof by Induction (see the Appendix) When k = 1,
If (Ak)–1 = (A–1)k,
By mathematical induction, we conclude that (Ak)–1 = (A–1) for all positive integers k.
3)
4)
p. 29 Chapter 2: Matrices. 2.3 The Inverse of a Matrix.
Theorem 2.9 Inverse of a Product * If A and B are invertible nn matrices, then (AB)–1 = B–1A–1.
Proof:
Theorem 2.10 Cancellation Properties Let C be an invertible matrix. * 1) If AC = BC, then A = B. (Right cancellation property)
2) If CA = CB, then A = B. (Left cancellation property)
Theorem 2.11 Systems of Equations with Unique Solutions.* If A is an invertible matrix, then the system Ax = b has a unique solution x = A–1b.
Review: What is wrong with the following “proof” that if AB = I and CA = I then B = C?
AB CA
A A
B = C
What is wrong with the following “proof” that if AB = I and CA = I and A is invertible then B = C?
AB = CA
A–1AB = CAA–1
IB = CI
B = C
p. 30 Chapter 2: Matrices. 2.4 Elementary Matrices.
2.4 Elementary Matrices. Objective: Factor a matrix into a product of elementary matrices. Objective: Find the PA = LDU factorization of a matrix.
An elementary matrix is a square matrix of dimensions nn (or of order n) is a matrix that can be obtained from the nn identity matrix by a single elementary row operation.
The three elementary row operations, with examples of corresponding elementary matrices, are
0001 1000 1) Swapping two rows, e.g. R2 R4 E1 = 0100 0010
001 2) Multiplying a single row by a nonzero constant, e.g. 3R R E = 3 3 2 010 300
0001 0010 3) Adding a multiple of one row to another row, e.g. –2R1 + R3 R3 E3 = 0102 1000
Theorem 2.12 Representing Elementary Row Operations
If we premultiply (multiply on the left) a matrix A by an elementary matrix, we obtain the same result as if we had applied the corresponding elementary row operation to A.
Examples:
0001 0189 r1 1000 1913 r2 = = R2 R4 0100 9497 r3 0010 2514 r4
001 189 = 3R R 010 913 3 3 300 497
p. 31 Chapter 2: Matrices. 2.4 Elementary Matrices.
1 0 0 0 9 8 1 0 0 1 0 0 3 1 9 1 = –2R1 + R3 R3 2 0 1 0 7 9 4 9 0 0 0 1 4 1 5 2
Gaussian elimination can be represented by a product of elementary matrices. For example,
0 1 3 2 A = 1 2 1 5 3 6 9 3 0 1 0 R R E = 1 2 1 1 0 0 0 0 1 1 2 1 5 0 1 3 2 3 6 9 3 1 0 0 3R + R R E = 1 3 3 2 0 1 0 3 0 1 1 2 1 5 0 1 3 2 0 0 6 12 1 0 0 1 R R E = 0 1 0 6 3 3 3 0 0 1/ 6 1 2 1 5 0 1 3 2 0 0 1 2
1 0 0 1 0 0 so = E (E (E A)) = 3 2 1 0 1 0 0 1 0 0 0 1/ 6 3 0 1
Two nn matrices A and B are row-equivalent when there exist a finite number of elementary matrices such that B = EnEn–1…E2E1A.
p. 32 Chapter 2: Matrices. 2.4 Elementary Matrices.
A square matrix L is lower triangular if all entries above the main diagonal are zero, i.e. lij = 0 whenever i < j. A square matrix U is upper triangular if all entries below the main diagonal are zero, i.e. uij = 0 whenever i > j. A square matrix D is diagonal if all entries not on the main diagonal are zero, i.e. dij = 0 whenever i j.
0 0 0 0 L = U = D = 0 0 0 0 0 0 0 0
Theorem 2.13+ Elementary Matrices are Invertible
If E is an elementary matrix, then E–1 exists and is an elementary matrix. Moreover, if E is lower triangular, then E–1 is also lower triangular. And if E is diagonal, then E–1 is also diagonal.
Examples:
0 1 0 E = R R E –1 = R R 1 1 0 0 1 2 1 1 2 0 0 1 1 0 0 1 0 0 E = 3R + R R E –1 = –3R + R R 2 0 1 0 1 3 3 2 0 1 0 1 3 3 3 0 1 3 0 1 1 0 0 1 0 0 E = 0 1 0 1 R R E –1 = 0 1 0 6R R 3 6 3 3 3 3 3 0 0 1/ 6 0 0 6
You can check these using matrix multiplication. 1 0 0 1 0 0 E.g. = and = . 0 1 0 0 1 0 0 0 1 0 0 1
Theorem 2.14 Invertible Matrices are Row-Equivalent to the Identity Matrix
A square matrix A is invertible if and only if it is row equivalent to the identity matrix:
A = En…E2E1I = En…E2E1 if and only if A is a product of elementary matrices.
Proof of “If A is invertible, then A is a product of elementary matrices”:
Since A is invertible, Ax = b has a unique solution (namely, x = A–1b). But this means that we can use row operations to reduce [ A | b ] to [ I | c ] (where c = A–1b, of course). If the
p. 33 Chapter 2: Matrices. 2.4 Elementary Matrices.
corresponding elementary matrices are E1, E2, …, En, then I = EnEn–1…E2E1A so –1 –1 –1 –1 A = E1 E2 … En–1 En , which is a product of elementary matrices.
Proof of “If A is a product of elementary matrices, then A is invertible”:
–1 –1 –1 –1 –1 –1 If A = E1E2… En–1En, then A exists and A = En En–1 …E2 E1 because every elementary matrix is invertible.-
Theorem 2.15 Equivalent Conditions for Invertibility
If A is an n n matrix, then the following statements are equivalent.
1) A is invertible. 2) Ax = b has a unique solution for every n1 column matrix b (namely, x = A–1b). 3) Ax = 0 has only the trivial solution. 4) A is row-equivalent to Inn. 5) A can be written as a product of elementary matrices.
Returning to our example of Gaussian elimination using elementary matrices, we found the 0 1 3 2 reduced echelon form of the augmented matrix 1 2 1 5 3 6 9 3
1 2 1 5 1 0 0 1 0 0 0 1 0 = 0 1 3 2 0 1 0 0 1 0 1 0 0 0 0 1 2 0 0 1/ 6 3 0 1 0 0 1
Look just at the 33 coefficient matrix instead of the 34 augmented matrix. We have
row-echelon form mRows mRowAdds row swaps 1 2 1 1 0 0 1 0 0 0 1 0 0 1 3 0 1 3 = 0 1 0 0 1 0 1 0 0 1 2 1 0 0 1 0 0 1/ 6 3 0 1 0 0 1 3 6 9 U A
The row-echelon form is upper triangular. In general, we may have more than one mRow (multiply a row by a nonzero constant) elementary matrix, more than one mRowAdd (add a multiple of one row to another row) elementary matrix, and more than one row swap matrix. The product of and arbitrary number of row swap matrices is called a permutation matrix P.
mRows mRowAdds
U= Fn F2 F1 Em E2 E1 PA
p. 34 Chapter 2: Matrices. 2.4 Elementary Matrices. mRowAdds mRows E 1E 1 E 1 F 1F 1 F 1 U= PA 12m 1 2n lower triangular diagonal
Lemma The product of diagonal matrices is diagonal. The product of lower triangular matrices is lower triangular.
LDU = PA E 1E 1E 1 F 1F 1F 1 12m 12 n row-echelon form row swaps 1 0 0 1 0 0 1 2 1 0 1 0 0 1 3 0 1 0 0 1 0 0 1 3 = 1 0 0 1 2 1 3 0 1 0 0 1/ 6 0 0 1 0 0 1 3 6 9 L D U P A
Theorem LU-Factorization
Every square matrix can be factored as PA = LDU, where P is a permutation matrix, L is lower triangular with all ones on the main diagonal, D is diagonal, and U is upper triangular with all ones on the main diagonal.
A variation on this is PA = LU, where this L equals the LD from above, and does not necessarily have ones on the diagonal.
The PA = LU factorization is the usual method used by computers for solving systems of linear equations, finding inverse matrices, and calculating determinants (Chapter 3). It is also useful in proofs.
p. 35 Chapter 2: Matrices. 2.4 Elementary Matrices.
0 1 1 Example: Find the PA = LDU factorization of A = 1 0 1 2 3 6
0 1 1 1 0 1 1 0 1 Solution: A = R1R2 2R1R3R3 3R2R3R3 1 0 1 0 1 1 0 1 1 2 3 6 2 3 6 0 3 4 1 0 1 1 0 1 R2 R2 = U 0 1 1 0 1 1 0 0 1 0 0 1
1 0 1 1 0 0 1 0 0 1 0 0 0 1 0 0 1 1 * = 0 1 0 0 1 0 0 1 1 0 1 0 1 0 0 1 0 1 0 0 1 0 0 1 0 3 1 2 0 1 0 0 1 2 3 6 U F E2 E1 P A
1 0 0 1 0 0 1 0 0 1 0 1 0 1 0 0 1 0 0 1 0 = 0 1 1 2 0 1 0 3 1 0 0 1 0 0 1 1 1 1 U E1 E2 F
1 0 0 1 0 0 = 0 1 0 0 1 0 2 3 1 0 0 1 L D
p. 36 Chapter 2: Matrices. 2.5 Applications of Matrix Operations.
2.5 Applications of Matrix Operations. Objective: Write and use a stochastic (Markov) matrix. Objective: Use matrix multiplication to encode and decode messages. Objective: Use matrix algebra to analyze and economic system (Leontief input-output model).
Consider a situation in which members of a population occupy a finite number of states {S1, S2, …, Sn}. For example, a multinational company has $4 trillion in assets (the population). Some of the money is in the Americas, some in Asia, and the rest is in Europe (the three states). In a Markov process, at each discrete step in time, members of the population may move from one state to another, subject to the following rules:
1) The total number of individuals stays the same. 2) The numbers in each state never become negative. 3) The new state depends only on the current state (history is disregarded). from The behavior of a Markov process is described by a matrix SSS of transition probabilities (or stochastic matrix or Markov 21 n 1211 ppp 1n S1 matrix). 2221 ppp 2n S2 P to pij is the probability that a member of the population will th th change from the j state to the i state. The rules above ppp S become nn 21 nn n
1) Each column of the transition matrix adds up to one. 2) Every probability entry is 0 ≤ pij ≤ 1.
1 Example: Stochastic Matrix. A chemistry course is taught in two sections. Every week, 4 of the 1 1 students in Section A and 3 of the students in Section B drop, and 6 of each section transfer to the other section. Write the transition matrix. At the beginning of the semester, each section has 144 students. Find the number of students in each section and the number of students who have dropped after one week and after two weeks. 1 7 6 1 Solution: 12 A B 2
* 1 1 1 (A) 7 1 1 4 6 6 0 12 6 0 1 1 1 (B) 1 1 P = 6 6 3 01 = 6 2 0 . 1 1 1 1 (d) 1 1 1 4 0 1 4 3 4 3 0 3 144 108 (A) 79 P = (B) . P = 144 96 66 drop 0 84 (d) 143
1
p. 37 Chapter 2: Matrices. 2.5 Applications of Matrix Operations. Matrix multiplication can be used to encode and decode messages. The encoded messages are called cryptograms.
To begin, assign a number to each letter of the alphabet (and assign 0 to a space).
0 1 2 3 4 5 6 7 8 9 10 11 12 13 _ A B C D E F G H I J K L M
14 15 16 17 18 19 20 21 22 23 24 25 26 N O P Q R S T U V W X Y Z
Use these numbers to convert a message in to a row matrix, including spaces but ignoring punctuation. Then partition the row matrix into 13 uncoded row matrices.
Example:
M A K E _ I T _ S O _ _ [13 1 11] [5 0 9] [20 0 19] [15 0 0]
1 4 1 Example: Use the invertible matrix A = to encode the message MAKE IT SO. 2 7 6 0 1 3 Solution:
*--[ 13 1 11] = [11 –56 26] [20 0 19] = [20 –99 37]
[5 0 9] = [5 –29 22] [15 0 0] = [15 –60 –15]
The sequence of encoded matrices is [11 –56 26] [5 –29 22] [20 –99 37] [15 –60 –15]
Removing the brackets yields the cryptogram 11 –56 26 5 –29 22 20 –99 37 15 –60 –15
In order to decrypt a message, we need to know the encryption matrix A.
p. 38 Chapter 2: Matrices. 2.5 Applications of Matrix Operations.
1 4 1 Example: Use the invertible matrix A = to decode 2 7 6 0 1 3 –8 26 31 13 –73 50 19 –97 44 16 –64 –16. Solution:
27 13 17 A–1 = 6 3 4 2 1 1 *-–-
[–8 26 31] = [2 5 1] [19 –97 44] = [19 0 21]
[13 –73 50] = [13 0 21] [16 –64 –16] = [16 0 0]
[2 5 1] [13 0 21] [19 0 21] [16 0 0] B E A M _ U S _ U P _ _
User (Output) In economics, an input-output model (developed by
Leontief) consists of n different industries In, each of I1 I 2 I n
which needs inputs (e.g. steel, food, labor, …) and has d11 d12 d1n I1 and output. To produce a unit (e.g. $1 million) of output, d 21 d 22 d 2n I 2 Supplier an industry may use the outputs of other industries and of D itself. For example, production of steel may use steel, (Input) food, and labor. d n1 d n2 d nn I n
Let dij be the amount of output the industry j needs from industry i to produce one unit of output per year. (We assume the dij are constant, i.e. fixed prices.) The matrix of these coefficients is called the input-output matrix or consumption matrix D. A column represents all of the inputs to a given industry. For this model to work, 0 ≤ dij ≤ 1 and the sum of the entries in each column must be less than or equal to 1. (Otherwise, it costs more than one unit to produce a unit in that industry.)
Let xi be the total output matrix of industry i, and X = [xi]. If the economic system is closed (self- sustaining: total output = “intermediate demand,” i.e. what is needed to produce it), then X = DX. If the system is open with external demand matrix E (e.g. exports) then X = DX + E.
To find what output matrix is needed to produce a given external demand matrix, we solve
p. 39 Chapter 2: Matrices. 2.5 Applications of Matrix Operations. X = DX + E X – DX = E (I – D)X = E X = (I – D)–1E
Example: Input-Output Economic Model
Production of one unit of steel requires 0.4 units of steel, no food, and 0.5 units of labor. Production of one unit of food requires no steel, 0.1 units of food, and 0.7 units of labor. Production of one unit of labor requires 0.1 units of steel, 0.8 units of food, and 0.1 units of *-–labor.- Find the output matrix when the external demands are 300 units of steel, 200 units of food, and no labor.
Solution:
1 0.4 0 0.1 300 1 0 0 0.4 0 0.1 848 D = . E = . X = 0 1 0 0 0.1 0.8 0 0.1 0.8 200 2076 0.5 0.7 0.1 0 0 0 1 0.5 0.7 0.1 2086
We need 848 units of steel, 2076 units of food, and 2086 units of labor.
p. 40 Chapter 3: Determinants. 3.1 The Determinant of a Matrix.
Chapter 3: Determinants.
3.1 The Determinant of a Matrix. Objective: Find the determinant of a 22 matrix. Objective: Find the minors and cofactors of a matrix. Objective: Use expansion by cofactors to find the determinant of a matrix. Objective: Find the determinant of a triangular matrix.
Every square matrix can be associated with a scalar called its determinant. Historically, determinants were recognized as a pattern of nn systems of linear equations. The system
11 1 12 bxaxa12 1 1222baab2 11 12abba21 has the solution x1 and x2 . 21 1 22 bxaxa22 aaaa21122211 aaaa21122211
The determinant of a 11 matrix A = a11 is det(A) = |A| = a11. The |…| symbols mean determinant, not absolute value.
aa 1211 aa 1211 The determinant of a 22 matrix A = is det(A) = |A| = = a11a22 – a12a21. aa 2221 aa 2221
Geometrically, the signed area of a parallelogram with vertices at (0, 0), (x1, y1), (x2, y2), and (x1 + x2, y1 + y2) is yx A = 11 yx 22 (x2, y2)
(The area is positive if the angle from (x1, y1) (x1, y1) to (x2, y2) is counterclockwise; otherwise, the area is negative.)
Proof:
The area A of the parallelogram is A = area of large rectangle areas of four triangles areas of two small rectangles 1 1 = (x1 + x2)(y1 + y2) 2 x1y1 2 x1y1 x2y2 x2y2 2x2y1
= x1y1 + x1y2 + x2y1 + x2y2 x1y1 x2y2 2x2y1
= x1y2 x2y1 =
p. 41 Chapter 3: Determinants. 3.1 The Determinant of a Matrix. To define the determinant of a square matrix A of order (dimensions) higher than 2, we define minors and cofactors. The minor Mij of the entry aij is the determinant of the matrix obtained by i+j deleting row i and column j of A. The cofactor Cij of the entry aij is Cij = (–1) Mij. Notice that (–1)i+j is a “checkerboard” pattern: (–1)i+j =
Examples: Finding Cofactors. Let A =
a12 a13 Find C21. Solution: C21 = – = –a12a33 + a13a32 a32 a33
a11 a13 Find C22. Solution: C22 = + = a11a33 – a13a31 a31 a33
The determinant of an nn matrix A (n ≥ 2) is the sum of the entries in the first row of A multiplied by their respective cofactors.
n det(A) = a1 j C1 j = a11C11 + a12C12 + … + a1nC1n j1
3 3 4 8 7 6 7 6 8 Example: 6 8 7 = 3 + 3 + 4 5 9 3 9 3 5 3 5 9 = 3(–72 – 35) +3(54 – 21) + 4(30 + 24) = 3(–107) +3(33) +4(54) = –6
Theorem 3.1 Expansion by Cofactors
Let A be a square matrix of order n. Then the determinant of A is given by an expansion in any row i
n det(A) = Ca ijij = ai1Ci1 + ai2Ci2 + … + ainCin j1
and also by an expansion in any column j
n det(A) = aijCij = a1jC1j + a2jC2j + … + anjCnj i1
p. 42 Chapter 3: Determinants. 3.1 The Determinant of a Matrix.
When expanding, you don’t need to find the cofactors of zero entries, because aijCij = (0)Cij = 0.
The definition of the determinant is inductive, because it uses the determinant of a matrix of order n – 1 to define the determinant of a matrix of order n.
Example: Expanding by Cofactors to Find a Determinant
= = –1 + 3(– )
= –1(2 +7 ) +3(–2 ) = –1(2(3) + 7(–6)) + 3(–2)(0) = –(6 – 42) = 36
To find a determinant using Mathematica, type Det[a] (also on Basic Math Assistant, More drop-down menu)
To find a determinant on the TI-89, type Matrixdet(A
To find a determinant of a 33 matrix, you can also use the following shortcut. Copy Columns 1 and 2 into Columns 4 and 5. To calculate the determinant, add and subtract the indicated products. subtract
a11 a12 a13 a11 a12 a13 a11 a12
a21 a22 a23 a21 a22 a23 a21 a22
a31 a32 a33 a31 a32 a33 a31 a32 add
= a11a22a33 + a12a23a31 + a13a21a32 – a31a22a13 – a32a23a11 – a33a21a12
12 30 0 1 1 3 1 1 3 1 1 Example: 0 4 6 0 4 6 0 4 so = 32 + 6 + 0 – 12 – 30 – 0 = –4 1 5 8 1 5 8 1 5 32 6 0
p. 43 Chapter 3: Determinants. 3.1 The Determinant of a Matrix. Theorem 3.2 Determinant of a Triangular Matrix
The determinant of a triangular matrix A of order n is the product of the entries on the main diagonal. det(A) = a11a22… ann
Proof by Induction for upper triangular matrices:
When k = 1, A = [a11] so |A| = a11
Assume that the theorem holds for all upper triangular matrices of order k. Let A be an upper triangular matrix of order k + 1. Then expanding in the last row,
a11 a12 a1k a1,k 1 a11 a12 a1k 0 a22 a2k a2,k 1 0 a22 a2k |A| = = ak+1,k+1 0 0 akk ak,k 1 0 0 akk 0 0 0 ak1,k 1
= ak+1,k+1(a11a22…akk) = a11a22…akk ak+1,k+1
The proof for lower triangular matrices is similar.
Optional application to multivariable calculus:
u2 x2 du Remember integration by substitution: f (u)du = f (u(x)) dx dx u1 x1 cos(x) optional For example, cot(x)dx = dx sin(x) du cos(x) 1 du 1 Let u = sin(x), = cos(x) so dx = dx = du = ln|u| + C = ln|sin(x)| + C dx sin(x) u dx u
In multivariable calculus, u u u x y z v v v f (u,v, w)dudvdw = f (u(x, y, z),v(x, y, z), w(x, y, z)) dxdydz V V x y z w w w x y z
The determinant is called the Jacobian.
p. 44 Chapter 3: Determinants. 3.2 Determinants and Elementary Operations.
3.2 Determinants and Elementary Operations. Objective: Use elementary row operations to evaluate a determinant. Objective: Use elementary column operations to evaluate a determinant. Recognize conditions that yield zero determinants.
In practice, we rarely evaluate determinants using expansion by cofactors. The properties of determinants under elementary operations provide a much quicker way to evaluate determinants.
Theorem 3.9 det(AT) = det(A). [Proof is in Section 3.4]
Theorem 3.3 Elementary Row (Column) Operations and Determinants.
Let A and B be nn square matrices.
a) When B is obtained from A by swapping two rows (two columns) of A, det(B) = –det(A). b) When B is obtained from A by adding a multiple of one row of A to another row of A (or one column of A to another column of A), det(B) = det(A). c) When B is obtained from A by multiplying of a row (column) of A by a nonzero constant c, det(B) = c det(A).
Theorem 3.4 Conditions that Yield a Zero Determinant.
If A is an nn square matrix and any one of the following conditions is true, then det(A) = 0
a) An entire row (or an entire column) consists of zeros. b) Two rows (or two columns) are equal. c) One row is a multiple of another row (or one column is a multiple of another column).
Proof by Induction of 3.3a (for rows):
aa 1211 aa 2221 When k = 2, A = and B = so aa 2221 aa 1211 det(B) = a21a12 – a22a11 = –(a11a22 – a12a21) = –det(A)
Assume that the theorem holds for all matrices of order k. Let A be a matrix of order k + 1 and B be a matrix obtained by swapping two rows of A. To find det(A) and det(B), expand in any row other than the swapped rows. The respective cofactors are opposites, because they come from kk matrices that have two rows swapped. Thus, det(B) = –det(A).
Proof of 3.4a (for rows): Suppose that row i of A is all zeroes. Expand by cofactors in row i. n n det(A) = Ca ijij = 0 Cij = 0 j1 j1
p. 45 Chapter 3: Determinants. 3.2 Determinants and Elementary Operations. Proof of 3.4b: Let B be the matrix obtained from A by swapping the two identical rows (columns) of A, so det(B) = –det(A). But B = A, so det(A) = –det(A) so det(A) = 0.
Proof of 3.3b (for rows): Suppose B is obtained from A by adding c times row k to row i. Expand by cofactors in row i. Note that the cofactors of Cij are the same for matrices A and B, because the matrices are the same everywhere except row i. n n n n det(B) = Cbijij = ca kj )(Caijij = kj Cacij + Caijij = c·0 + det(A) = det(A). j1 j1 j1 j1 n because kj Caij is the determinant of a matrix with two identical rows (row k and row i). j1 See Theorem 3.4b.
Another way of writing this is
a11 a1n 11 aa 1n 11 aa 1n
ak1 akn k1 aa kn k1 aa kn = c + = c·0 + det(A) = det(A).
ca aik 11 ca kn ain )()( k1 aa kn i1 aa in
an1 ann n1 aa nn n1 aa nn
Proof of 3.3c (for rows): Suppose B is obtained from A by multiplying row i by a nonzero scalar c. Expand by cofactors in row i. n n n det(B) = Cb ijij = ca Cijij = c Ca ijij = c det(A) j1 j1 j1
Proof of 3.4c: Suppose B is a matrix with two equal rows (or two equal columns), and A is obtained from B by multiplying one of those rows (or columns) by a nonzero scalar c. Using 3.3c on that row (or column), det(A) = c det(B). Using 3.4b, det(B) = 0. Thus, det(A) = 0.
Geometrically, the signed area of a parallelogram with edges from (0, 0) to (x1, y1) and from (0, 0) to (x2, y2) has the same properties yx as 11 when you perform an elementary row operation. Also, yx 22 the signed area of a parallelepiped with edges from (0, 0) to (x1, y1, z1), from (0, 0) to (x2, y2, z2), and from (0, 0) to (x3, y3, z3) has the
zyx 111 same properties as zyx 222 when you perform an elementary
zyx 333 row operation.
p. 46 Chapter 3: Determinants. 3.2 Determinants and Elementary Operations. Row Swapping (Theorem 3.3a)
yx x y If we swap two rows, e.g. 11 2 2 yx 22 x1 y1
zyx111 x2 y2 z2
or zyx222 x1 y1 z1
zyx333 x3 y3 z3
then the sign of the area/volume changes because we change from a right-hand orientation to a left-hand orientation.
Adding a multiple of one row to another row (Theorem 3.3b)
x y If we add a multiple of one row to another row, e.g. 1 1 , then x2 0.5x1 y2 0.5y1 the A = bh is unchanged.
p. 47 Chapter 3: Determinants. 3.2 Determinants and Elementary Operations. Adding a multiple of one row to another row (Theorem 3.3b)
yx cx cy If we multiply a row by a nonzero constant c, e.g. 11 1 1 , then the A = bh yx 22 x2 y2 (cb)h = cA is also multiplies by the constant c.
Example: Finding a Determinant Using Elementary Row Operations
4 3 2 14 11 0 5 4 1 = 5 4 1 R + 2R R = –1[14(–13) – 11(–22)] 1 2 1
2 3 4 2 3 4 = –1(60) 14 11 0 = –60 = 5 4 1 R3 – 4R2 R3
22 13 0 14 11 = –1 22 13
p. 48 Chapter 3: Determinants. 3.2 Determinants and Elementary Operations. Example: Finding a Determinant Using Elementary Column Operations
C2 3C1 C2 1 3 3 4 1 0 0 0 2 4 6 5 2 ______ C3 ____ C3 7 5 1 3 = 7 ______ 9 5 1 2 9 ______ C4 ____ C4 1 0 0 0 2 1 __ __
= (__) 7 ______C2 C2
9 ______1 0 0 0 ______ C3 2 1 0 0 = (__) 7 ____________ C 9 ______ 4 1 0 0 ______= (__)(__) = (__)(__)(__) ______
=
p. 49
Chapter 3: Determinants. 3.3 Properties of Determinants.
3.3 Properties of Determinants. Objective: Find the determinant of a matrix product and of a scalar multiple of a matrix. Find the determinant of an inverse matrix and recognize equivalent conditions for a nonsingular matrix. Find the determinant of the transpose of a matrix.
Theorem 3.5 Determinant of a Matrix Product
If A and B are square matrices of the same order, then det(AB) = det(A) det(B).
Proof: To begin, let E be an elementary matrix. By Thm 2.12, EB is the matrix obtained from applying the corresponding row operation to B. By Thm. 3.3, B)det( exchanging two rows det(EB) = B)det( if the row operation is ofmultiple one a adding another torow Bc)det( constantgmultiplyin a c row nonzero aby Also by Thm 3.3, 1 det(E) = det(EI) = 1 if the row operation is c Thus, det(EB) = det(E) det(B). This can be generalized by induction to conclude that |Ek…E2E1B| = |Ek| |…| |E2| |E1| |B| where the Ei are elementary matrices. If A is nonsingular, then by Thm. 2.14, it can be written as the product A = Ek…E2E1 so |AB| = |A| |B|.
If A is singular, then A is row-equivalent to a matrix with an entire row of zeroes (for example, the reduced row echelon form). From Thm 3.4, we know |A| = 0. Moreover, because A is singular, it follows that AB must be singular. (Proof by contradiction: if AB were nonsingular, then A[B(AB)-1] = I would show that A is not singular, because A–1 = B(AB)-1.) Therefore, |AB| = 0 = |A| |B|.
Comment on Proof by Contradiction: “P implies Q” is equivalent to “not Q implies not P.”
Theorem 3.6 Determinant of a Scalar Multiple of a Matrix
If A is a square matrix of order n and c is a scalar, then det(cA) = cndet(A).
Proof: Apply Property (c) of Thm. 3.3 to each of the n rows of A to obtain n factors of c.
Theorem 3.7 Determinant of an Invertible Matrix
A square matrix A is invertible (nonsingular) if and only if det(A) 0.
Proof: On the one hand, if A is invertible, then AA–1 = I, so . |A| |A–1| = | I | = 1. Therefore, |A| 0. On the other hand, assume det(A) 0. Then use Gauss-Jordan elimination to find the reduced row-echelon form R. Since R is in reduced row-echelon form, it is either the identity matrix or
p. 51 Chapter 3: Determinants. 3.3 Properties of Determinants. it must have at least one row of all zeroes. The second case is not possible: if R had a row of all zeroes, then det(R) = 0, but then det(A) = 0 (which contradicts the assumption). Therefore, A is row-equivalent to R = I, so A is invertible.
Theorem 3.8 Determinant of an Inverse Matrix
1 If A is an invertible matrix, then det (A–1) = det(A)
1 Proof: AA–1 = I, so . |A| |A–1| = | I | = 1 and |A| 0, so |A–1| = . | A |
Equivalent Conditions for a Nonsingular nn Matrix (Summary)
1) A is invertible. 2) Ax = b has a unique solution for every n1 column matrix b. 3) Ax = 0 has only the trivial solution for the n1 column matrix 0. 4) A is row-equivalent to I. 5) A can be written as a product of elementary matrices. 6) det(A) 0.
Theorem 3.9 Determinant of a the Transpose of a Matrix
If A is a square matrix, then det(AT) = det(A).
Proof: Let A be a square matrix of order n. From Section 2.4, we know that A can be factored as PA = LDU, where P is a permutation matrix, L is lower triangular with all ones on the main diagonal, D is diagonal, and U is upper triangular with all ones on the main diagonal. L is obtained from I by adding a multiple of the rows containing the diagonal ones to the rows below the diagonal, so |L| = |I | = 1 by Thm. 3.3b. Likewise, U is obtained from I by adding a multiple of the rows containing the diagonal ones to the rows above the diagonal, so |U| = |I | = 1 by Thm. 3.3b. By Thm. 3.2, |D| = d11d22… dnn
i.e. the product of its diagonal elements. P is a product of elementary row-swap matrices, each of which has determinant –1. So |P| is the product of some number of –1’s.
|P| = 1 if the number of row swaps is even; |P| = –1 if the number of row swaps is odd.
p. 52 Chapter 3: Determinants. 3.3 Properties of Determinants.
T 1 0 0 e i 1 0 1 0 eT i2 Let e1 = , e2 = , …, en = be n1 matrices. Then P = where i1, i2, …, in is 0 0 1 eT in some permutation of 1, 2, …, n. Now PT = e e e , so i1 i2 in eT e eT e eT e 1 0 0 i1 i1 i1 i2 i1 in T T T T ei ei ei ei ei ei 0 1 0 PP = 2 1 2 2 2 n = = I, eT e eT e eT e 0 0 1 in i1 in i2 in in
and by Thm. 3.5, det(P) det(PT ) = det(PPT ) = det(I) = 1.
Then either det(P) = 1 so det(PT) = 1, or det(P) = –1 so det(PT) = –1. In both cases,
det(P) = det(PT).
So we have PA = LDU which gives us |P| |A| = |L| |D| |U| = (1) |D| (1) =|D|, so | D | |A| = . | P | Taking the transpose, we have ATPT = UTDTLT which gives us |AT| |PT| = |UT| |DT| |LT|. Now |PT| = |P|; |DT| = |D| because DT = D since D is diagonal; |LT| = 1
because LT is upper triangular with all ones on the main diagonal; and
|UT| = 1
because UT is lower triangular with all ones on the main diagonal. Thus, |AT| |PT| = |UT| |DT| |LT| becomes |AT| |PT| = (1) |D| (1), so
| D | | D | |AT| = = = |A|. | PT | | P |
p. 53
Chapter 3: Determinants. 3.4 Applications of Determinants.
3.4 Applications of Determinants. Objective: Find the adjoint of a matrix and use it to find the inverse of a matrix. Objective: Use Cramer’s Rule to solve a system of n linear equations in n unknowns. Objective: Use determinants to find area, volume, and the equations of lines and planes.
Using the adjoint of a matrix to calculate the inverse is time-consuming and inefficient. In practice, Gauss-Jordan elimination is used for 33 matrices and larger. However, “adjoint” is vocabulary you may be expected to know in future classes.
i+j Recall from Section 3.1 that the cofactor Cij of a matrix A is (–1) times the determinant of the matrix obtained by deleting row i and column j of A.
11 12 CCC1n CCC The matrix of cofactors of A is 21 22 2n . nn21 CCCnn
11 21 CCC n1 CCC The adjoint of A is the transpose of matrix of cofactors: adj(A) = 12 22 n2 21 nn CCC nn
Theorem 3.10 The Inverse of a Matrix Given by Its Adjoint
1 If A is an invertible matrix, then A–1 = adj(A). A)det(
1211 aaa 1n 2221 aaa 2n 11 21 j1 CCCC n1 12 22 j2 CCCC n2 Proof: Consider A[adj(A)] = ii 21 aaa ki 21 nn jn CCCC nn optional nn 21 aaa nn
The ij entry of this product is ai1Cj1 + ai2Cj2 + … + ainCjn. If i = j, this is det(A) (expanded by cofactors in row i). If i j, this is the determinant of the matrix B, which is the same as A except that row j has been replaced with row i.
p. 55 Chapter 3: Determinants. 3.4 Applications of Determinants.
a11 a12 a1n a11 a12 a1n a a a a a a 21 22 2n 21 22 2n 11 21 j1 CCCCn1 a a a a a a CCCC A = i1 i2 ki , B[adj(A)] = i1 i2 ki 12 22 j2 n2 CCCC a j1 a j2 a jn ai1 ai2 ain 21nn jn nn an1 an2 ann an1 an2 ann The j column of the cofactor matrix is unchanged, because it does not depend on the j row of A or B. Since two rows of B are the same, the cofactor expansion for i j is zero.
det(A) 0 0 0 det(A) 0 Thus, A[adj(A)] = = det(A)I. 0 0 det(A)
1 det(A) 1 because A is invertible, so we can write A[ adj(A)] = I, so A–1 = adj(A). A)det(
For 33 matrices and larger, Gauss-Jordan elimination is much more efficient than the adjoint method for finding the inverse of a matrix.
a b d b –1 1 a b However, for a 22 matrix A = , adj(A) = so A = . c d c a ad bc c d
Cramer’s Rule to solve n linear equation in n variables is time-consuming and inefficient. In practice, Gaussian elimination is used to solve linear systems. However, Cramer’s Rule is vocabulary you may be expected to know in future classes.
Theorem 3.11 Cramer’s Rule
If an nn system Ax = b has a coefficient matrix with nonzero determinant |A| 0, then det(A ) det(A ) det(A ) x 1 , x 2 , … x n 1 det(A) 2 det(A) n det(A)
where Ai is the matrix A but with column i replace by b.
p. 56 Chapter 3: Determinants. 3.4 Applications of Determinants.
x1 11 21 CCCn1 b1 x 1 CCC b Proof: 2 x = A–1b = adj(A)b = 12 22 n2 2 A)det( xn 21nn CCCnn bn
so xi = (b1C1i + b2C2i + … + bnCni). The sum in parentheses is the cofactor expansion of
det(Ai ) det(Ai), so xi = det(A)
1x 2y 10 Example: Solve using Cramer’s Rule. 3x 4y 15
10 2 1 10 15 4 10 3 15 15 15 Solution: x = = = –5; y = = = 1 2 2 1 2 2 2 3 4 3 4
Area, Volume and Equations of Lines and Planes:
We already know that the signed area of a parallelogram yx is given by a 22 determinant (Section 3.1). A = 11 yx 22
x1 y1 1
The signed area of a triangle with vertices (x1, y1), (x2, y2), and (x3, y3) is A = x2 y2 1
x3 y3 1
Proof: (x3, y3) The area of a triangle is half of the area of a parallelogram. So the area of the triangle we want is
x1 y1 1 x y x y x y 1 1 2 2 + 3 3 + 1 1 = x y 1 2 x y x y x y 2 2 3 3 1 1 2 2 2 (x2, y2) x3 y3 1 pos neg neg (x1, y1)
x1 y1 1 1 For a triangle the signed area is A = x2 y2 1 2 x3 y3 1
If the vertices (x1, y1), (x2, y2), and (x3, y3) are ordered clockwise then the area is positive; otherwise, it is negative. (The homework asks for the absolute value of the area.)
p. 57 Chapter 3: Determinants. 3.4 Applications of Determinants. The area of the triangle is zero if and only if the three points are collinear.
x1 y1 1
(x1, y1), (x2, y2), and (x3, y3) are collinear if and only if x2 y2 1 = 0
x3 y3 1
x y 1 The equation of a line through distinct points (x1, y1) and (x2, y2) is x1 y1 1 = 0
x2 y2 1
Similarly to the two-dimensional case of a parallelogram, the signed volume of a parallelepiped is given
zyx111
by a 33 determinant (Section 3.1). V = zyx222
zyx333
Let’s find the volume of the tetrahedron (pyramid with four triangular faces) with vertices at (x1, y1, z1), (x2, y2, z2), (x3, y3, z3), and (x4, y4, z4).
1 The volume of the tetrahedron with vertices at (0, 0, 0), (x1, y1, z1), (x2, y2, z2), and (x3, y3, z3) is 6
x1 y1 z1 1 times the volume of the parallelepiped, i.e. x y z . For the triangle above 6 2 2 2 x3 y3 z3
x1 y1 1 1 ( 2 x2 y2 1 ), we had 3 sides and the areas of 3 triangles to add/subtract. Now we have 4faces
x3 y3 1 and the volumes of 4 tetrahedrons to add/subtract. The signed volume of the tetrahedron with vertices at (0, 0, 0), (x1, y1, z1), (x2, y2, z2), and (x3, y3, z3) is (imagine for the moment that (x4, y4, z4) in front of the triangle (x1, y1, z1), (x2, y2, z2), (x3, y3, z3)
x1 y1 z1 x4 y4 z4 x2 y2 z2 x3 y3 z3 1 1 1 1 V = x y z + x y z + x y z + x y z 6 4 4 4 6 1 1 2 6 3 3 3 6 2 2 2 x3 y3 z3 x2 y2 z2 x4 y4 z4 x1 y1 z1
For a tetrahedron, the signed volume is x y z 1 1 1 1 1 x2 y2 z2 1 V = – 6 x3 y3 z3 1
x4 y4 z4 1
p. 58 Chapter 3: Determinants. 3.4 Applications of Determinants. If when you wrap the fingers of your right from (x1, y1, z1) to (x2, y2, z2) to (x3, y3, z3), your thumb points toward (x4, y4, z4), then the signed volume is positive. If when you wrap the fingers of your left from (x1, y1, z1) to (x2, y2, z2) to (x3, y3, z3), your thumb points toward (x4, y4, z4), then the signed volume is negative. (The homework asks for the absolute value of the volume.)
Four points (x1, y1, z1), (x2, y2, z2), (x3, y3, z3), and (x4, y4, z4) are coplanar if and only if x y z 1 1 1 1 x y z 1 2 2 2 = 0 because that is when the tetrahedron has zero volume.
x3 y3 z3 1
x4 y4 z4 1
The equation of a plane through distinct points (x1, y1, z1), (x2, y2, z2), and (x3, y3, z3) is x y z 1
x1 y1 z1 1 = 0 x2 y2 z2 1
x3 y3 z3 1
p. 59
Chapter 4: Vector Spaces. 8.1 Complex Numbers (Optional).
Chapter 4: Vector Spaces.
8.1 Complex Numbers (Optional). Objective: Use the Quadratic Formula to find all zeroes of a quadratic polynomial with real coefficients. Objective: Add and subtract two complex numbers and multiply a complex number by a real scalar. Objective: Graphically represent complex numbers in the complex plane as directed line segments. Objective: Multiply two complex numbers. Objective: Multiply and find determinants of matrices with complex entries. Objective: Perform Gaussian and Gauss-Jordan elimination on matrices with complex entries.
So far, the scalars we have been using have been real numbers R. However, any mathematical field can be used as the scalars.
Properties of a Field F
Let a, b, and c be any elements of F. Then F is a field if it has two operations, addition and multiplication, and it contains two distinct elements 0 and 1, such that the following properties are true.
1) a + b is an element of F. Closure under addition 2) a + b = b + a Commutative property of addition 3) (a + b) + c = a + (b + c) Associative property of addition 4) a + 0 = a Additive identity property 5) There exists a –a such that a + (–a) = 0 Additive inverse property 6) ab is an element of F. Closure under multiplication 7) ab = ba Commutative property of multiplication 8) (ab)c = a(bc) Associative property of multiplication 9) 1a = a Multiplicative identity property 10) Except for a = 0, there exists an a–1 such that a(a–1) = 1 Multiplicative inverse property 11) a(b + c) = ab + ac Distributive property
Three familiar examples of fields are the rational numbers Q = , the real numbers R, and the complex numbers C.
A familiar set that is not a field is the integers Z = {…, –2, –1, 0, 1, 2, …}. Why not?
We are going to need to use the field of complex numbers as our scalars, because we will need to solve polynomial equations. All polynomials can be solved using complex numbers (complex numbers are an “algebraically closed field”), but the same is not true of real numbers.
p. 61 Chapter 4: Vector Spaces. 8.1 Complex Numbers (Optional).
def The imaginary unit is i 1 , so i2 = –1.
(Electrical engineers often write j 1 because they use i for electric current.)
Example: Solve 5x2 – 6x + 5 = 0
Aside: the complex roots of a polynomial with real coefficients are complex conjugates (Section 8.2) of each other (a + bi and a – bi).
Notice that we have assumed a definition of multiplication by a real number: 1 1 10 (6 + 8i) = ( 10 6) + ( 8)i
Example: Solving the polynomial equation 2x3 +3x2 +50x +75 = 0 using software.
Solution:
Mathematica: inputting Clear[x];Solve[2x^3+3x^2+50x+75==0,x]
yields output
You can also find Solve in the Palettes Menu::Basic Math Assistant::y = x menu.
TI-89: ComplexcSolve(2x^3+3x^2+50x+75=0,x)
A complex number is a number of the form a + bi, where a and b are real numbers. a is the real part and bi is the imaginary part of a + bi. The form a + bi is the standard form of a complex number, for example 1 + 2i, 0 + 3i, and –4 + 0i.
Geometrically, a complex number a + bi is represented in the complex plane by a directed line segment from the origin to (a, b), where a and b are Cartesian coordinates. In other words, the horizontal axis is the real axis and the vertical axis is the imaginary axis.
Operations in the Set of Complex Numbers C def Addition: (a + bi) + (c + di) (a + c) + (b + d)i def Multiplication by a real number: c(a + bi) ca + cbi Negative: –(a + bi) –a + –bi. Notice that –(a + bi) = (–1)(a + bi). Subtraction: (a + bi) + (c + di) (a + bi) + –(c + di) = (a – c) + (b – d)i
p. 62 Chapter 4: Vector Spaces. 8.1 Complex Numbers (Optional). Examples
Let z = 2 + 4i and w = –4 + 1 i. Illustrate the following graphically. z, w, z + w, 1.5z, – w, w – z p. 63 Chapter 4: Vector Spaces. 8.1 Complex Numbers (Optional). Multiplication of complex numbers is defined using the distributive property and using i 2 = –1. def (a + bi)(c + di) (ac – bd) + (ad + bc)i
because (a + bi)(c + di) = ac + adi + bci + bdi 2 = ac + adi + bci + bdi 2 = ac + adi + bci – bd
Warning: When you multiply square roots of negative numbers, convert into standard form a + bi (not using the square root of a negative number) before you multiply. For example, 1 1 = i·i = –1, not 1 = 1()()1 = 1 = 1.
Application (Electrical Engineering): V(t)= I(t)Z, I(t)= current = I0 cos(t) + i I0 sin(t) 1 V(t) = volatge, Z = impedance = R + + Li, Ci R = resistance, C = capacitance, L = inductance
3 4 Example: Use software to check that i is a zero of the polynomial 5x2 – 6x + 5. 5 5
Solution:
Mathematica: to input the imaginary unit i, denoted by i in Mathematica, type ii (four separate keystrokes). After the first three keystrokes, you will see ii. After the fourth keystroke, ii will change to i.
Type x=3/5+4ii/5 5x^2-6x+5
TI-89: (3/5+4/5). 5x^2-6x+5
Complex matrices 413 i Example: Find the inverse of A = by hand. Check your answer by multiplying. i 541
–1 1 415 i Solution: A = . A)det( i 341 det(A) = 5(3) – (–1 + 4i)(–1 – 4i) = 15 – (1 + 4i – 4i + 16) = –2 1 A–1 = 2
p. 64 Chapter 4: Vector Spaces. 8.1 Complex Numbers (Optional).
–1 413i 1 415i Check: AA = i 541 2 i 341 1 15 441 ii 16 3 12 3 12ii 02 01 = = = 2 5 20 5 20 441 iiii 1516 20 10 Example: By hand, find the determinant of
Example: Perform Gauss-Jordan elimination using row operations to solve
)2( xiw iy z 22 iw 31)53(3)42(2 iziyxi ()62()36(3 16 3)7 iziyixiw
p. 65
9781133110873_0801.qxp 3/10/12 6:52 AM Page 392
392 Chapter 8 Complex Vector Spaces
8.1 Complex Numbers
Use the imaginary unit i to write complex numbers. Graphically represent complex numbers in the complex plane as points and as vectors. Add and subtract two complex numbers, and multiply a complex number by a real scalar. Multiply two complex numbers, and use the Quadratic Formula to find all zeros of a quadratic polynomial. Perform operations with complex matrices, and find the determinant of a complex matrix.
REMARK COMPLEX NUMBERS When working with products So far in the text, the scalar quantities used have been real numbers. In this chapter, you involving square roots of will expand the set of scalars to include complex numbers. negative numbers, be sure to In algebra it is often necessary to solve quadratic equations such as convert to a multiple of i before x2 3x 2 0. The general quadratic equation is ax 2 bx c 0, and its multiplying. For instance, consider the following solutions are given by the Quadratic Formula operations. b Ίb2 4ac x Ί 1Ί 1 i i 2a i 2 where the quantity under the radical,b2 4ac, is called the discriminant. If 2 1 Correct b 4ac 0, then the solutions are ordinary real numbers. But what can you conclude about the solutions of a quadratic equation whose discriminant is negative? For Ί Ί Ί͑ ͒͑ ͒ 1 1 1 1 instance, the equation x2 4 0 has a discriminant of b2 4ac 16, but there is Ί1 no real number whose square is 16. To overcome this deficiency, mathematicians 1 Incorrect invented the imaginary unit i, defined as i Ί 1 where i 2 1. In terms of this imaginary unit, Ί 16 4Ί 1 4i. With this single addition of the imaginary unit i to the real number system, the system of complex numbers can be developed.
Definition of a Complex Number If and ba are real numbers, then the number a bi is a complex number, where a is the real part and bi is the imaginary part of the number. The form a bi is the standard form of a complex number.
Some examples of complex numbers written in standard form are 2 2 0i, 4 3i, and 6i 0 6i. The set of real numbers is a subset of the set of complex numbers. To see this, note that every real number a can be written as a complex number using b 0. That is, for every real number, a a 0i. A complex number is uniquely determined by its real and imaginary parts. So, two complex numbers are equal if and only if their real and imaginary parts are equal. That is, if a bi and c di are two complex numbers written in standard form, then a bi c di if and only if a c and b d. 9781133110873_0801.qxp 3/10/12 6:52 AM Page 393
8.1 Complex Numbers 393
Imaginary THE COMPLEX PLANE axis Because a complex number is uniquely determined by its real and imaginary parts, (a, b) or a + bi it is natural to associate the number a bi with the ordered pair ͑a, b͒. With this association, complex numbers can be represented graphically as points in a coordinate b plane called the complex plane. This plane is an adaptation of the rectangular Real axis coordinate plane. Specifically, the horizontal axis is the real axis and the vertical axis a is the imaginary axis. The point that corresponds to the complex number a bi is ͑a, b͒, as shown in Figure 8.1. The Complex Plane Figure 8.1 Plotting Numbers in the Complex Plane
Plot each number in the complex plane. a.4 3i b. 2 i c. 3i d. 5 SOLUTION Figure 8.2 shows the numbers plotted in the complex plane.
a.Imaginary b. Imaginary axis axis 4 2 3 4 + 3i 1 or (4, 3) Real 2 axis −31−2 −1 2 3 1 Real −2 − i axis or (−2, −1) −2−1 1234 −3 −2 −4
c.Imaginary d. Imaginary axis axis 2 4 1 3 Real axis 2 −3123−2 −1 1 5 or (5, 0) Real −2 axis −1 12345 −3 −3i or (0, −3) −4 −2
Figure 8.2
Another way to represent the complex number a bi is as a vector whose horizontal component is a and whose vertical component is b. (See Figure 8.3.) (Note that the use of the letter i to represent the imaginary unit is unrelated to the use of i to represent a unit vector.)
Imaginary axis
1 Horizontal component Real axis
−1
−2 Vertical component 4 − 2i −3
Vector Representation of a Complex Number Figure 8.3 9781133110873_0801.qxp 3/10/12 6:52 AM Page 394
394 Chapter 8 Complex Vector Spaces
ADDITION, SUBTRACTION, AND SCALAR MULTIPLICATION OF COMPLEX NUMBERS Because a complex number consists of a real part added to a multiple of i, the operations of addition and multiplication are defined in a manner consistent with the rules for operating with real numbers. For instance, to add (or subtract) two complex numbers, add (or subtract) the real and imaginary parts separately.
Definition of Addition and Subtraction of Complex Numbers The sum and difference of a bi and c di are defined as follows. ͑a bi͒ ͑c di͒ ͑a c͒ ͑b d͒i Sum ͑ ͒ ͑ ͒ ͑ ͒ ͑ ͒ REMARK a bi c di a c b d i Difference Note in part (a) of Example 2 that the sum of two complex numbers can be a real number. Adding and Subtracting Complex Numbers
a. ͑2 4i͒ ͑3 4i͒ ͑2 3͒ ͑ 4 4͒i 5 b. ͑1 3i͒ ͑3 i͒ ͑1 3͒ ͑ 3 1͒i 2 4i
Using the vector representation of complex numbers, you can add or subtract two complex numbers geometrically using the parallelogram rule for vector addition, as shown in Figure 8.4.
Imaginary Imaginary axis axis z = 3 + 4i 4 2 3 w = 3 + i 1 2 Real 1 z w axis + = 5 Real −3 1 23 axis −1 162345
−2 −3 − −3 z = 1 3i −4 w = 2 − 4i z − w = −2 − 4i Addition of Complex Numbers Subtraction of Complex Numbers Figure 8.4 Many of the properties of addition of real numbers are valid for complex numbers as well. For instance, addition of complex numbers is both associative and commutative. Moreover, to find the sum of three or more complex numbers, extend the definition of addition in the natural way. For example, ͑ 2 i͒ ͑3 2i͒ ͑ 2 4i͒ ͑2 3 2͒ ͑1 2 4͒i 3 3i. 9781133110873_0801.qxp 3/10/12 6:52 AM Page 395
8.1 Complex Numbers 395
Another property of real numbers that is valid for complex numbers is the distributive property of scalar multiplication over addition. To multiply a complex number by a real scalar, use the definition below.
Definition of Scalar Multiplication If c is a real number and a bi is a complex number, then the scalar multiple of c and a bi is defined as c͑a bi͒ ca cbi.
Scalar Multiplication with Complex Numbers
a. 3 ͑2 7i͒ 4͑8 i͒ 6 21i 32 4i 38 17i b. 4͑1 i͒ 2͑3 i͒ 3͑1 4i͒ 4 4i 6 2i 3 12i 1 6i
Geometrically, multiplication of a complex number by a real scalar corresponds to the multiplication of a vector by a scalar, as shown in Figure 8.5.
Imaginary Imaginary axis axis
4 3 3 2 2z = 6 + 2i z = 3 + i 2 1 z = 3 + i Real 1 axis Real 1 23 axis − − − 1 2345 6 z = 3 i −1 −2 −2 −3
Multiplication of a Complex Number by a Real Number Figure 8.5 With addition and scalar multiplication, the set of complex numbers forms a vector space of dimension 2 (where the scalars are the real numbers). You are asked to verify this in Exercise 55.
LINEAR Complex numbers have some useful applications in ALGEBRA electronics. The state of a circuit element is described by APPLIED two quantities: the voltage V across it and the current I flowing through it. To simplify computations, the circuit element’s state can be described by a single complex number z V li, of which the voltage and current are simply the real and imaginary parts. A similar notation can be used to express the circuit element’s capacitance and inductance. When certain elements of a circuit are changing with time, electrical engineers often have to solve differential equations. These can often be simpler to solve using complex numbers because the equations are less complicated.
Adrio Communications Ltd/shutterstock.com 9781133110873_0801.qxp 3/10/12 6:52 AM Page 396
396 Chapter 8 Complex Vector Spaces
MULTIPLICATION OF COMPLEX NUMBERS The operations of addition, subtraction, and scalar multiplication of complex numbers have exact counterparts with the corresponding vector operations. By contrast, there is no direct vector counterpart for the multiplication of two complex numbers.
Definition of Multiplication of Complex Numbers The product of the complex numbers a bi and c di is defined as ͑a bi͒͑c di͒ ͑ac bd͒ ͑ad bc͒i.
Rather than try to memorize this definition of the product of two complex numbers, simply apply the distributive property, as follows. TECHNOLOGY ͑a bi͒͑c di͒ a͑c di͒ bi͑c di͒ Distributive property Many graphing utilities and 2 software programs can ac ͑ad͒i ͑bc͒i ͑bd͒i Distributive property calculate with complex ac ͑ad͒i ͑bc͒i ͑bd͒͑ 1͒ Use i2 1. numbers. For example, on ͑ ͒ ͑ ͒ some graphing utilities, you ac bd ad i bc i Commutative property can express a complex number ͑ac bd͒ ͑ad bc͒i Distributive property a bi as an ordered pair ͑a, b͒. Try verifying the result of Example 4(b) by multiplying Multiplying Complex Numbers ͑2, 1͒ and ͑4, 3͒.You should obtain the ordered pair ͑11, 2͒. a. ͑ 2͒͑1 3i͒ 2 6i b. ͑2 i͒͑4 3i͒ 8 6i 4i 3i 2 8 6i 4i 3͑ 1͒ 8 3 6i 4i 11 2i
Complex Zeros of a Polynomial
Use the Quadratic Formula to find the zeros of the polynomial p͑x͒ x2 6x 13 and verify that p͑x͒ 0 for each zero. SOLUTION Using the Quadratic Formula, b Ίb2 4ac 6 Ί 16 6 4i x 3 2i. 2a 2 2 Substitute each value of x into the polynomial p͑x͒ to verify that p͑x͒ 0. p͑3 2i͒ ͑3 2i͒2 6͑3 2i͒ 13 ͑3 2i͒͑3 2i͒ 6͑3 2i͒ 13 REMARK 9 6i 6i 4 18 12i 13 0 A well-known result from p͑3 2i͒ ͑3 2i͒2 6͑3 2i͒ 13 algebra states that the complex ͑ ͒͑ ͒ ͑ ͒ zeros of a polynomial with real 3 2i 3 2i 6 3 2i 13 coefficients must occur in 9 6i 6i 4 18 12i 13 0 conjugate pairs. (See Review Exercise 81.) In Example 5, the two complex numbers 3 2i and 3 2i are complex conjugates of each other (together they form a conjugate pair). More will be said about complex conjugates in Section 8.2. 9781133110873_0801.qxp 3/10/12 6:52 AM Page 397
8.1 Complex Numbers 397
COMPLEX MATRICES Now that you are able to add, subtract, and multiply complex numbers, you can apply these operations to matrices whose entries are complex numbers. Such a matrix is called complex.
Definition of a Complex Matrix A matrix whose entries are complex numbers is called a complex matrix.
All of the ordinary operations with matrices also work with complex matrices, as demonstrated in the next two examples.
Operations with Complex Matrices
Let A and B be the complex matrices i 1 i 2i 0 A ΄ ΅ and B ΄ ΅ 2 3i 4 i 1 2i and determine each of the following. a.3A b.͑2 i͒B c.A B d. BA SOLUTION i 1 i 3i 3 3i a. 3A 3΄ ΅ ΄ ΅ 2 3i 4 6 9i 12 2i 0 2 4i 0 b. ͑2 i͒B ͑2 i͒΄ ΅ ΄ ΅ i 1 2i 1 2i 4 3i i 1 i 2i 0 3i 1 i c. A B ΄ ΅ ΄ ΅ ΄ ΅ 2 3i 4 i 1 2i 2 2i 5 2i 2i 0 i 1 i d. BA ΄ ΅΄ ΅ i 1 2i 2 3i 4 2 0 2i 2 0 ΄ ΅ 1 2 3i 4i 6 i 1 4 8i 2 2 2i ΄ ΅ 7 i 3 9i
Finding the Determinant of a Complex Matrix
Find the determinant of the matrix TECHNOLOGY 2 4i 2 A ΄ ΅. Many graphing utilities and 3 5 3i software programs can perform matrix operations on complex SOLUTION matrices. Try verifying the 2 4i 2 calculation of the determinant det͑A͒ Խ 3 5 3iԽ of the matrix in Example 7. ͑ ͒͑ ͒ ͑ ͒͑ ͒ You should obtain the same 2 4i 5 3i 2 3 answer, ͑ 8, 26͒. 10 20i 6i 12 6 8 26i 9781133110873_0801.qxp 3/10/12 6:52 AM Page 398
398 Chapter 8 Complex Vector Spaces
8.1 Exercises
Simplifying an Expression In Exercises 1Ð6, determine 42. p͑x͒ x3 2x2 11x 52 Zero: x 4 the value of the expression. 43. p͑x͒ 2x3 3x2 50x 75 Zero: x 5i Ί Ί Ί Ί Ί Ί 1.2 3 2.8 8 3. 4 4 44. p͑x͒ x3 x2 9x 9 Zero: x 3i 4.i 3 5.i 4 6. ͑ i͒7 Operations with Complex Matrices In Exercises Equality of Complex Numbers In Exercises 7Ð10, 45Ð54, perform the indicated matrix operation using the determine x such that the complex numbers in each pair complex matrices A and B. are equal. 1 ؉ i 1 1 ؊ i 3i ΅ ΄ ؍ and B ΅ ΄ ؍ A x 3i, 6 3i 2 ؊ 2i ؊3i ؊3 ؊i .7 8. ͑2x 8͒ ͑x 1͒i, 2 4i 45.A B 46. B A 9. ͑x 2 6͒ ͑2x͒i, 15 6i 1 47.2A 48. 2B 10. ͑ x 4͒ ͑x 1͒i, x 3i 1 49.2iA 50. 4iB ͑ ͒ ͑ ͒ Plotting Complex Numbers In Exercises 11Ð16, plot 51. det A B 52. det B the number in the complex plane. 53.5AB 54. BA 11.z 6 2i 12.z 3i 13. z 5 5i 55. Proof Prove that the set of complex numbers, with the 14.z 7 15.z 1 5i 16. z 1 5i operations of addition and scalar multiplication (with real scalars), is a vector space of dimension 2. Adding and Subtracting Complex Numbers In Exercises 17Ð24, find the sum or difference of the complex numbers. Use vectors to illustrate your answer. 56. Consider the functions ͑ ͒ 2 ͑ ͒ 2 17.͑2 6i͒ ͑3 3i͒ 18. ͑1 Ί2i͒ ͑2 Ί2i͒ p x x 6x 10 and q x x 6x 10. 19.͑5 i͒ ͑5 i͒ 20. i ͑3 i͒ (a) Without graphing either function, determine whether the graphs of p and q have x -intercepts. ͑ ͒ ͑ ͒ ͑ ͒ 21.6 2i 22. 12 7i 3 4i Explain your reasoning. ͑ ͒ ͑ ͒ ͑ ͒ ͑ ͒ 23.2 i 2 i 24. 2 i 2 i (b) For which of the given functions is x 3 i a Scalar Multiplication In Exercises 25 and 26, use zero? Without using the Quadratic Formula, find the vectors to illustrate the operations geometrically. Be sure other zero of this function and verify your answer. to graph the original vector. 25. u and 2u, where u 3 i 57. (a) Evaluate in for n 1, 2, 3, 4, and 5. 3 2010 26. 3u and 2u, where u 2 i (b) Calculate i . (c) Find a general formula for in for any positive integer n. Multiplying Complex Numbers In Exercises 27Ð34, 0 i find the product. 58. Let A ΄ ΅. ͑ ͒͑ ͒ ͑ ͒͑ 2 ͒ i 0 27.5 5i 1 3i 28. 3 i 3 i (a) Calculate An for n 1, 2, 3, 4, and 5. 29.͑Ί7 i͒͑Ί7 i͒ 30. ͑4 Ί2i͒͑4 Ί2 i ͒ (b) Calculate A2010. 31.͑a bi͒2 32. ͑a bi͒͑a bi͒ (c) Find a general formula for An for any positive 33.͑1 i͒3 34. ͑2 i͒͑2 2i͒͑4 i͒ integer n. Finding Zeros In Exercises 35Ð40, determine all the True or False? In Exercises 59 and 60, determine zeros of the polynomial function. whether each statement is true or false. If a statement is ͑ ͒ 2 ͑ ͒ 2 35.p x 2x 2x 5 36. p x x x 1 true, give a reason or cite an appropriate statement from 37.p͑x͒ x 2 5x 6 38. p͑x͒ x 2 4x 5 the text. If a statement is false, provide an example that 39.p͑x͒ x4 16 40. p͑x͒ x4 10x 2 9 shows the statement is not true in all cases or cite an appropriate statement from the text. Finding Zeros In Exercises 41Ð44, use the given zero 59.Ί 2Ί 2 Ί4 2 60. ͑Ί 10͒2 Ί100 10 to find all zeros of the polynomial function. 41. p͑x͒ x3 3x2 4x 2 Zero: x 1 61. Proof Prove that if the product of two complex numbers is zero, then at least one of the numbers must be zero. 9781133110873_08_ANS.qxd 3/10/12 6:58 AM Page A1
Answer Key
Ϫ5 Ϫ15 ϩ 10i Section 8.1 51. Ϫ5 Ϫ 3i 53. ΄ ΅ 55. Proof 25i 15 ϩ 30i 57. (a) i1 ϭ i (b) i2010 ϭϪ1 1.ϪΊ6 3.Ϫ4 5. 1 7.x ϭ 6 9. x ϭ 3 i2 ϭϪ1 11. Imaginary 13. Imaginary i3 ϭϪi axis axis 4 ϭ z = −5 + 5i i 1 5 2 5 i ϭ i Real 4 1 n ϭ 4k axis 3 i n ϭ 4k ϩ 1 246 (c) in ϭ , where k is an integer. −2 2 ΆϪ1 n ϭ 4k ϩ 2 − z = 6 2i Ϫ ϭ ϩ −4 1 i n 4k 3 Real 59. False. See the Remark, page 392. 61. Proof axis −5 −4 −3 −2 −1 15. Imaginary axis
5 z = 1 + 5i 4 3 2 1 Real axis 12345 17. 5 ϩ 3i 19. 2i Imaginary Imaginary axis axis
10 6 8 − u = 2 + 6i 4 u v = 2i 6 2 u = 5 + i 4 u + v = 5 + 3i Real 2 axis Real −2 8 axis v = 5 − i −2 4 6810 −4 −4 v = 3 − 3i 21. 6 ϩ 2i 23. 4 ϩ 2i Imaginary Imaginary axis axis
4 4 u − v = 6 + 2i 3 2 u + v = 4 + 2i Real 2 axis v = 2 + i 24u = 6 1 u = 2 + i −2 v = −2i Real axis 1234 25. Imaginary axis
4 −u = −3 + i Real axis −4 −2 6 −2 2u = 6 − 2i − 4 u = 3 − i 27. 20 ϩ 10i 29. 8 31. ͑a2 Ϫ b2͒ ϩ 2abi Ϫ ϩ Ϫ1 ± 3 ± ± 33. 2 2i 35. 2 2i 37. 2, 3 39. 2, 2 2 1 ϩ 3i 41. 1, 1 ± i 43. Ϫ3, ±5i 45. ΄ ΅ 2 Ϫ1 Ϫ 2i Ϫ4i 2 ϩ 2i 2 Ϫ2 ϩ 2i 2i 47. ΄ ΅ 49. ΄ ΅ 4 Ϫ 4i Ϫ6i 4 ϩ 4i 6 Chapter 4: Vector Spaces. 8.2 Conjugates and Division of Complex Numbers (Optional).
8.2 Conjugates and Division of Complex Numbers (Optional). Objective: Find the conjugate of a complex number. Objective: Find the modulus of a complex number. Objective: Divide complex numbers. Objective: Perform Gaussian on and find the inverses of matrices with complex entries.
The conjugate of the complex number z = a + bi is denoted by 푧̅ or z* and is given by z* = a – bi
Theorem 8.1 Properties of Complex Conjugates
For a complex numbers z = a + bi, 1) zz* = 푧푧̅ = a2 + b2 2) zz* = 푧푧̅ 0 3) zz* = 푧푧̅ = 0 if and only if z = 0 4) (z*)* = (푧̅) = z
The modulus of the complex number z = a + bi is denoted by |z| and is given by |z| = ba 22
Theorem 8.2 The modulus of a complex number. |z| = zz* = √푧푧̅
Example: Find z* and |z| if z = 4 – i. Solution: z* = 4 + i |z| = 2 )1(4 2 = 17
The quotient of two complex numbers z = a + bi and w = c + di, w 0, is z z w* zw* (a bi)(c di) adbcbdac )( i = = = = w w w* w || 2 dc 22 dc 22
72 i Example: Find . 34 i 72 i 72( )( ii )34 68 ii 2128 13 34 Solution: = = = i 34 i 34( )( ii )34 22 )34( 25 25
p. 67 Chapter 4: Vector Spaces. 8.2 Conjugates and Division of Complex Numbers (Optional). Theorem 8.3 Properties of Complex Conjugates
For complex numbers z and w (w 0), 1) (z + w)* = z* + w* i.e. 푧̅̅̅+̅̅̅푤̅̅ = 푧̅ + 푤̅ 2) (z – w)* = z* – w* i.e. 푧̅̅̅−̅̅̅푤̅̅ = 푧̅ − 푤̅ 3) (zw)* = z*w* i.e. 푧푤̅̅̅̅ = 푧̅ 푤̅ 4) (z/w)* = z*/w* i.e. 푧̅̅/̅̅푤̅ = 푧̅/푤̅
5 5i 4 i Example: Find the inverse of A = by hand. 3 4i 1 3i Solution
p. 68 Chapter 4: Vector Spaces. 8.2 Conjugates and Division of Complex Numbers (Optional). Example: Perform Gaussian elimination using row operations to solve
(3 i)x (8 4i)y (9 3i)z 52 4i 2ix 7iy (12 10i)z 67 57i (5 5i)x 20iy (15 16i)z 57 100i
Solution:
p. 69
9781133110873_0802.qxp 3/14/12 8:23 AM Page 399
8.2 Conjugates and Division of Complex Numbers 399
8.2 Conjugates and Division of Complex Numbers
Find the conjugate of a complex number. Find the modulus of a complex number. Divide complex numbers, and find the inverse of a complex matrix.
COMPLEX CONJUGATES In Section 8.1, it was mentioned that the complex zeros of a polynomial with real coefficients occur in conjugate pairs. For instance, in Example 5 you saw that the zeros of p͑x͒ ϭ x2 Ϫ 6x ϩ 13 are 3 ϩ 2i and 3 Ϫ 2i. In this section, you will examine some additional properties of complex conjugates. You will begin with the definition of the conjugate of a complex number.
Definition of the Conjugate of a Complex Number The conjugate of the complex number z ϭ a ϩ bi is denoted by z and is given by z ϭ a Ϫ bi.
REMARK In part (d) of Example 1, note Finding the Conjugate of a Complex Number that 5 is its own complex conjugate. In general, it can Complex Number Conjugate be shown that a number is its a. z ϭϪ2 ϩ 3i z ϭϪ2 Ϫ 3i own complex conjugate if and only if the number is real. (See b. z ϭ 4 Ϫ 5i z ϭ 4 ϩ 5i Exercise 39.) c. z ϭϪ2i z ϭ 2i d. z ϭ 5 z ϭ 5 Imaginary axis Geometrically, two points in the complex plane are conjugates if and only if they z = −2 + 3i 3 are reflections in the real (horizontal) axis, as shown in Figure 8.6. Complex conjugates have many useful properties. Some of these are shown in Theorem 8.1. 2
Real THEOREM 8.1 Properties of Complex Conjugates axis −4 −3 −112 For a complex number z ϭ a ϩ bi, the following properties are true. −2 1. zz ϭ a2 ϩ b2 2. zz Ն 0 −3 3. zz ϭ 0 if and only if z ϭ 0. 4. ͑z͒ ϭ z z = −2 − 3i
Imaginary axis PROOF z = 4 + 5i 5 To prove the first property, let z ϭ a ϩ bi. Then z ϭ a Ϫ bi and 4 3 zz ϭ ͑a ϩ bi͒͑a Ϫ bi͒ ϭ a2 ϩ abi Ϫ abi Ϫ b2i 2 ϭ a 2 ϩ b2. 2 1 Real The second and third properties follow directly from the first. Finally, the fourth axis −3567−2 23 property follows from the definition of the complex conjugate. That is, −2 ͑z͒ ϭ ͑ ϩ ͒ ϭ Ϫ ϭ ϩ ϭ −3 a bi a bi a bi z. −4 −5 − z = 4 5i Finding the Product of Complex Conjugates Conjugate of a Complex Number Figure 8.6 When z ϭ 1 ϩ 2i, you have zz ϭ ͑1 Ϫ 2i͒͑1 ϩ 2i͒ ϭ 12 ϩ 22 ϭ 1 ϩ 4 ϭ 5. 9781133110873_0802.qxp 3/10/12 6:57 AM Page 400
400 Chapter 8 Complex Vector Spaces
THE MODULUS OF A COMPLEX NUMBER REMARK Because a complex number can be represented by a vector in the complex plane, it The modulus of a complex makes sense to talk about the length of a complex number. This length is called the number is also called the modulus of the complex number. absolute value of the number. In fact, when z is a real number, ԽzԽ Ίa2 02 ԽaԽ. Definition of the Modulus of a Complex Number The modulus of the complex number z a bi is denoted byԽzԽ and is given by ԽzԽ Ίa2 b2.
Finding the Modulus of a Complex Number
For z 2 3i and w 6 i, determine the value of each modulus. a.zԽ b.ԽwԽԽ c. ԽzwԽ SOLUTION a. ԽzԽ Ί22 32 Ί13 b. ԽwԽ Ί62 ͑ 1͒2 Ί37 c. Because zw ͑2 3i͒͑6 i͒ 15 16i, you have ԽzwԽ Ί152 162 Ί481.
Note that in Example 3,ԽzwԽ ԽzԽ ԽwԽ. In Exercise 41, you are asked to prove that this multiplicative property of the modulus always holds. Theorem 8.2 states that the modulus of a complex number is related to its conjugate.
THEOREM 8.2 The Modulus of a Complex Number For a complex number z, ԽzԽ2 zz.
PROOF Let z a bi, then z a bi and zz ͑a bi͒͑a bi͒ a2 b2 ԽzԽ2.
LINEAR Fractals appear in almost every part of the universe. They ALGEBRA have been used to study a wide variety of applications such APPLIED as bacteria cultures, the human lungs, the economy, and galaxies. The most famous fractal is called the Mandelbrot Set, named after the Polish-born mathematician Benoit Mandelbrot (1924–2010). The Mandelbrot Set is based on the following sequence of complex numbers. ͑ ͒2 zn zn 1 c, z1 c The behavior of this sequence depends on the value of the complex number c. For some values of c, the modulus of
each term zn in the sequence is less than some fixed number N, and the sequence is bounded. This means that c is in the Mandelbrot Set, and its point is colored black. For other values of c, the moduli of the terms of the sequence become infinitely large, and the sequence is unbounded. This means that c is not in the Mandelbrot Set, and its point is assigned a color based on “how quickly” the sequence diverges. Andrew Park/Shutterstock.com 9781133110873_0802.qxp 3/10/12 6:57 AM Page 401
8.2 Conjugates and Division of Complex Numbers 401
DIVISION OF COMPLEX NUMBERS One of the most important uses of the conjugate of a complex number is in performing division in the complex number system. To define division of complex numbers, consider z a bi and w c di and assume that c and d are not both 0. For the quotient z x yi w to make sense, it has to be true that z w͑x yi͒ ͑c di͒͑x yi͒ ͑cx dy͒ ͑dx cy͒i. But, because z a bi, you can form the linear system below. cx dy a dx cy b Solving this system of linear equations for x and y yields ac bd bc ad x and y . ww ww Now, because zw ͑a bi͒͑c di͒ ͑ac bd͒ ͑bc ad͒i, the following definition is obtained.
Definition of Division of Complex Numbers The quotient of the complex numbers z a bi and w c di is defined as z a bi REMARK w c di If c2 d 2 0, then c d 0, ac bd bc ad i and w 0. In other words, as c2 d 2 c2 d 2 is the case with real numbers, 1 division of complex numbers ͑ zw͒ by zero is not defined. ԽwԽ2 provided c2 d 2 0.
In practice, the quotient of two complex numbers can be found by multiplying the numerator and the denominator by the conjugate of the denominator, as follows. a bi a bi c di c di c di c di ͑a bi͒͑c di͒ ͑c di͒͑c di͒ ͑ac bd͒ ͑bc ad͒i c2 d 2 ac bd bc ad i c2 d 2 c2 d 2
Division of Complex Numbers
1 1 1 i 1 i 1 i 1 1 a. i 1 i 1 i 1 i 12 i 2 2 2 2 2 i 2 i 3 4i 2 11i 2 11 b. i 3 4i 3 4i 3 4i 9 16 25 25 9781133110873_0802.qxp 3/10/12 6:57 AM Page 402
402 Chapter 8 Complex Vector Spaces
Now that you can divide complex numbers, you can find the (multiplicative) inverse of a complex matrix, as demonstrated in Example 5.
Finding the Inverse of a Complex Matrix
Find the inverse of the matrix 2 i 5 2i A ΄ ΅ 3 i 6 2i 1 and verify your solution by showing that AA I2. SOLUTION Using the formula for the inverse of a 2 2 matrix from Section 2.3, 1 6 2i 5 2i A 1 ΄ ΅. ԽAԽ 3 i 2 i Furthermore, because ԽAԽ ͑2 i͒͑ 6 2i͒ ͑ 5 2i͒͑3 i͒ ͑ 12 6i 4i 2͒ ͑ 15 6i 5i 2͒ 3 i it follows that 1 6 2i 5 2i A 1 ΄ ΅ TECHNOLOGY 3 i 3 i 2 i If your graphing utility or 1 1 ͑ 6 2i͒͑3 i͒ ͑5 2i͒͑3 i͒ ΄ ΅ software program can perform 3 i 3 i ͑ 3 i͒͑3 i͒ ͑2 i͒͑3 i͒ operations with complex 1 20 17 i matrices, then you can verify ΄ ΅. the result of Example 5. If you 10 10 7 i have matrix A stored on a To verify your solution, multiply A and A 1 as follows. graphing utility, evaluate A 1. 2 i 5 2i 1 20 17 i 1 10 0 1 0 AA 1 ΄ ΅ ΄ ΅ ΄ ΅ ΄ ΅ 3 i 6 2i 10 10 7 i 10 0 10 0 1 The last theorem in this section summarizes some useful properties of complex conjugates.
THEOREM 8.3 Properties of Complex Conjugates For the complex numbers z and w, the following properties are true. 1. z w z w 2. z w z w 3. zw z w 4. z͞w z͞w
PROOF To prove the first property, let z a bi and w c di. Then z w ͑a c͒ ͑b d͒i ͑a c͒ ͑b d͒i ͑a bi͒ ͑c di͒ z w. The proof of the second property is similar. The proofs of the other two properties are left to you. 9781133110873_0802.qxp 3/10/12 6:57 AM Page 403
8.2 Exercises 403
8.2 Exercises
Finding the Conjugate In Exercises 1Ð6, find the Finding the Inverse of a Complex Matrix In Exercises complex conjugate z and geometrically represent both 31Ð36, determine whether the complex matrix A has an z and z. inverse. If A is invertible, find its inverse and verify that ؍ ؊1 1.z 6 3i 2. z 2 5i AA I. 6 3i 2i 2 i 3.z 8i 4. z 2i 31.A ΄ ΅ 32. A ΄ ΅ i i i 5.z 4 6. z 3 2 3 3 1 i 2 1 i 2 33.A ΄ ΅ 34. A ΄ ΅ Finding the Modulus In Exercises 7Ð12, find the 1 1 i 0 1 i ؊3 ؉ 2i, and؍ ؉ i, w 2 ؍ indicated modulus, where z i 0 0 1 0 0 .؊5i؍ v 35.A 0 i 0 36. A 0 1 i 0 2 ΄ ΅ ΄ ΅ 7.ԽzԽ 8. Խz Խ 0 0 i 0 0 1 i 9.ԽzwԽ 10. ԽwzԽ 11.ԽvԽ 12. Խzv2Խ Singular Matrices In Exercises 37 and 38, determine all values of the complex number z for which A is (.and solve for z 0 ؍ Verify that ԽwzԽ ԽwԽԽzԽ ԽzwԽ, where z 1 i and singular. (Hint: Set detͧAͨ .13 w 1 2i. 2 2i 1 i 5 z 14. Verify that Խzv2Խ ԽzԽԽv2Խ ԽzԽԽvԽ2, where z 1 2i 37.A ΄ ΅ 38. A 1 i 1 i z 3i 2 i ΄ ΅ and v 2 3i. 1 0 0
Dividing Complex Numbers In Exercises 15Ð20, 39. Proof Prove that z z if and only if z is real. perform the indicated operations. 2 i 1 1 i 15. 16. 40. Consider the quotient . i 6 3i 6 2i 3 Ί2i 5 i (a) Without performing any calculations, describe 17. 18. 3 Ί2i 4 i how to find the quotient. ͑2 i͒͑3 i͒ 3 i (b) Explain why the process described in part (a) 19. 20. 4 2i ͑2 i͒͑5 2i͒ results in a complex number of the form a bi. (c) Find the quotient. Operations with Complex Rational Expressions In Exercises 21Ð24, perform the operation and write the result in standard form. 41. Proof Prove that for any two complex numbers z and w, each of the statements below is true. 2 3 2i 5 21. 22. (a) ԽzwԽ ԽzԽԽwԽ 1 i 1 i 2 i 2 i (b) If w 0, then Խz͞wԽ ԽzԽ͞ԽwԽ. i 2i 1 i 3 23. 24. 42. Graphical Interpretation Describe the set of 3 i 3 i i 4 i points in the complex plane that satisfies each of the Finding Zeros In Exercises 25Ð28, use the given zero to statements below. find all zeros of the polynomial function. (a) ԽzԽ 3 (b) Խz 1 iԽ 5 25. p͑x͒ 3x3 4x2 8x 8 Zero: 1 Ί3i (c) Խz iԽ 2 (d) 2 ԽzԽ 5 26. p͑x͒ 4x3 23x2 34x 10 Zero: 3 i 43. (a) Evaluate ͑1͞i͒n for n 1,2, 3, 4, and 5. 27. p͑x͒ x4 3x3 5x2 21x 22 Zero: 3 Ί2i (b) Calculate ͑1͞i͒2000 and ͑1͞i͒2010. 28. p͑x͒ x3 4x2 14x 20 Zero: 1 3i (c) Find a general formula for ͑1͞i͒n for any positive integer n. Powers of Complex Numbers In Exercises 29 and 30, 1 i 2 find each power of the complex number z. 44. (a) Verify that i. Ί2 a)z 2 (b)z 3 (c)z؊1 (d) z؊2) (b) Find the two square roots of i. 29.z 2 i 30. z 1 i (c) Find all zeros of the polynomial x 4 1. Answer Key
Section 8.2 1. 6 ϩ 3i 3. 8i Imaginary Imaginary axis axis
4 z = 6 + 3i 8 z = 8i 2 4 Real axis 246 −8 −4 84 −2 −4 −4 − z = 6 3i −8 z = −8i
5. 4 Imaginary axis
3 2 1 z and z = 4 Real axis −1 31245 −2
7.Ί5 9.Ί65 11. 5 13. ԽwzԽ ϭ ԽϪ3 ϩ iԽ ϭ Ί10 ԽwԽԽzԽ ϭ Ί5Ί2 ϭ Ί10 ԽzwԽ ϭ ԽϪ3 ϩ iԽ ϭ Ί10 7 6Ί2 13 9 15. 1 Ϫ 2i 17. Ϫ i 19. ϩ i 11 11 10 10 Ϫ1 Ϫ 5 1 ϩ 9 Ϫ2 ± Ί 21. 2 2i 23. 10 10i 25. 3, 1 3i 27. 1, 2, Ϫ3 ± Ί2i Ϫ Ϫ 2 ϩ 1 3 ϩ 4 29. (a)3 4i (b) 2 11i (c)5 5i (d) 25 25i i Ϫ3i 31. AϪ1 ϭϪ1΄ ΅ 33. Not invertible 3 Ϫ2 ϩ i 6 Ϫi 0 0 Ϫ1 ϭ Ϫ ϪϪ5 10 35. A ΄ 0 i 0΅ 37. 3 3 i 0 0 Ϫ i 39. Proof 41. (a) and (b) Proofs 43. (a) ͑1͞i͒1 ϭ 1͞i ϭϪi, ͑1͞i͒2 ϭϪ1, ͑1͞i͒3 ϭ i, ͑1͞i͒4 ϭ 1, ͑1͞i͒5 ϭϪi (b) ͑1͞i͒2000 ϭ 1, ͑1͞i͒2010 ϭϪ1 1 n ϭ 4k Ϫi n ϭ 4k ϩ 1 (c) ͑1͞i͒n ϭ , where k is an integer ΆϪ1 n ϭ 4k ϩ 2 i n ϭ 4k ϩ 3 Chapter 4: Vector Spaces. 4.1 Vectors in Rn.
4.1 Vectors in Rn. Objective: Represent a vector in the plane as a directed line segment. Objective: Perform basic vector operations in R2 and represent them graphically. Objective: Perform basic vector operations in Rn. Prove basic properties about vectors and their operations in Rn.
In physics and engineering, a vector is an object with magnitude and direction and represented graphically by a directed line segment. In mathematics we have a much more general definition of a vector.
Geometrically, a vector in the plane is represented by a directed line segment with its initial point at the origin and its terminal (final) point at (x1, x2). The same ordered pair used to represent the terminal point is used to represent the vector. That is, x = (x1, x2). The coordinates x1 and x2 are called the components of the vector x. Two vectors u = (u1, u2) and v = (v1, v2) are equal iff u1 = v1 and u2 = v2.
Vector operations in R2
def Vector Addition: u + v = (u1, u2) + (v1, v2) (u1 + v1, u2 + v2). def Scalar Multiplication: cu = c(u1, u2) (cu1, cu2).
Negative: –u (–u1, –u2). Notice that –u = (–1)u.
Subtraction: u – v u + (–v) = (u1, u2) + (–v1, –v2) = (u1 – v1, u2 – v2).
The zero vector in R2 is 0 = (0, 0).
p. 71 Chapter 4: Vector Spaces. 4.1 Vectors in Rn. Examples
Let u = (2, 4) and v = (–4, 1). Illustrate the following graphically. u, v, u + v, 1.5u, – v, v – u p. 72 Chapter 4: Vector Spaces. 4.1 Vectors in Rn. Theorem 4.1 Properties of Vector Addition and Scalar Multiplication in the Plane (R2)
Let u, v, and w be vectors in R2, and let c and d be scalars.
1) u + v is a vector in R2. Closure under addition 2) u + v = v + u Commutative property of addition 3) (u + v) + w = u + (v + w) Associative property of addition 4) u + 0 = u Existence of additive identity 5) u + (–u) = 0 Existence of additive inverses 6) cv is a vector in R2. Closure under scalar multiplication 7) c(u + v) = cu + cv Distributive property over vector addition 8) (c + d)u = cu + du Distributive property over scalar addition 9) c(du) = (cd)u Associative property 10) 1u = u Multiplicative identity property
Proof of (3): Associative property of addition
(u + v) + w = [(u1, u2) + (v1, v2)] + (w1, w2)
=
=
= Assoc. prop. of addition of real numbers
=
=
= u + (v + w)
p. 73 Chapter 4: Vector Spaces. 4.1 Vectors in Rn. Proof of (8): Distributive property of scalar multiplication over real number addition
(c + d)u = (c + d)(u1, u2)
=
= Distributive property of real numbers
=
=
= cu + du
To add (1, 4) + (2, –2) in Mathematica, type {1,4}+{2,-2} You can also assign a variable by typing u={1,4} You can perform scalar multiplication by 3u or 3*u
To add (1, 4) + (2, –2) on the TI-89, you can type 1,4+2,-2 or 14+2-2 You can also assign a variable by typing 1,4.U You can perform scalar multiplication by 3u or 3u
Vector operations in Rn
We can generalize from the 2-dimensional plane R2 to an n-space Rn of ordered n-tuples. For example, R1 = R = set of all real numbers; R2 = 2-space = set of all ordered pairs of real numbers; R3 = 3-space = set of all ordered triples of real numbers.\
n An n-tuple (x1, x2, …, xn) can be viewed as a point in R with the xi as its coordinates, or as a vector with the xi as its components.
The standard vector operations in Rn are
def Vector Addition: u + v = (u1, u2, …, un) + (v1, v2, …, vn) (u1 + v1, u2 + v2, …, un + vn). def Scalar Multiplication: cu = c(u1, u2, …, un) (cu1, cu2, …, cun).
Negative: –u (–u1, –u2, …, –un). Notice that –u = (–1)u.
Subtraction: u – v u + (–v) = (u1, u2, …, un) + (–v1, –v2, …, –vn) = (u1 – v1, u2 – v2, …, un – vn).
The zero vector in Rn is 0 = (0, 0, …, 0).
p. 74 Chapter 4: Vector Spaces. 4.1 Vectors in Rn. Theorem 4.2 Properties of Vector Addition and Scalar Multiplication in the Plane (Rn)
Let u, v, and w be vectors in Rn, and let c and d be scalars.
1) u + v is a vector in Rn. Closure under addition 2) u + v = v + u Commutative property of addition 3) (u + v) + w = u + (v + w) Associative property of addition 4) u + 0 = u Existence of additive identity 5) u + (–u) = 0 Existence of additive inverses 6) cv is a vector in Rn. Closure under scalar multiplication 7) c(u + v) = cu + cv Distributive property over vector addition 8) (c + d)u = cu + du Distributive property over scalar addition 9) c(du) = (cd)u Associative property 10) 1u = u Multiplicative identity property
The vector 0 is called the additive identity in Rn and –v is the additive inverse of v.
Theorem 4.3 Properties of Vector Addition and Scalar Multiplication in Rn
Let v be a vector in Rn and let c be a scalar. Then
1) The additive identity is unique. That is, if v + u = v, then u = 0. 2) The additive inverse of v is unique. That is, if v + u = 0, then u = –v. 3) 0v = 0 4) c0 = 0 5) If cv = 0, then c = 0 or v = 0. 6) –(–v) = v
Proof of (1): Uniqueness of the additive identity
v + u = v Given
(v + u) + (–v) = v + (–v) Add –v to both sides
= v + (–v)
= v + (–v)
= 0
u = 0
p. 75 Chapter 4: Vector Spaces. 4.1 Vectors in Rn. Proof of (2): Uniqueness of the additive inverse
v + u = 0 Given
(–v) + (v + u) = (–v) + 0
= (–v) + 0
= (–v) + 0
= (–v) + 0
u = –v
p. 76 Chapter 4: Vector Spaces. 4.2 Vector Spaces.
4.2 Vector Spaces. Objective: Define a vector space and recognize some important examples of vector spaces. Objective: Show that a given set is not a vector space. (Optional)
Theorem 4.2 listed ten properties of vector addition and scalar multiplication in Rn. However, there are many other sets (Cn, sets of matrices, polynomials, functions) besides Rn that can be given suitable definitions of vector addition and scalar multiplication so that they too satisfy the same ten properties. Hence, one branch of mathematics, linear algebra, can study all of these.
Definition of a Vector Space
Let V be a set on which two operations (vector addition and scalar multiplication) are defined. If the axioms listed below are satisfied for every u, v, and w in V and every scalar c and d in a given field F (usually, F = R or F = C), then V is called a vector space over F.
1) u + v is in V. Closure under addition 2) u + v = v + u Commutative property * 3) (u + v) + w = u + (v + w) Associative property 4) V has a zero vector 0 such that Existence of additive identity for every u in V, u + 0 = u 5) For every u in V, there is a vector Existence of additive inverses (opposites) denoted by –u such that u + (–u) = 0
6) cv is a vector in V. Closure under scalar multiplication 7) c(u + v) = cu + cv Distributive property over vector addition 8) (c + d)u = cu + du Distributive property over scalar addition 9) c(du) = (cd)u Associative property 10) 1u = u Scalar identity
Notice that a vector space actually consists of four entities: a set V of vectors, a field F of scalars, and two defined operations (vector addition and scalar multiplication). Be sure all four entities are clearly understood. (For example, I could keep the set V of vectors, the field F of scalars, the same definition of scalar multiplication, but change the definition of how to add vectors and end up with a different vector space, or end up with something that is no longer a vector space.)
Examples of Vector Spaces. (Unless otherwise stated, assume the field is R.)
R2 with the standard operations * def def u + v = (u1, u2) + (v1, v2) (u1 + v1, u2 + v2). cu = c(u1, u2) (cu1, cu2). 0 = (0, 0) –u = (–u1, –u2) Rn with the standard operations. Note that this includes R2, which is just R with the usual addition and multiplication.
Cn over the field C with the standard operations
p. 77 Chapter 4: Vector Spaces. 4.2 Vector Spaces. More Examples of Vector Spaces. (Unless otherwise stated, assume the field is R.)
The vector space M2,3of all 23 real matrices with the standard operations
a11 a12 a13 b11 b12 b13 a11 b11 a12 b12 a13 b13 A + B = + = a21 a22 a23 b21 b22 b23 a21 b21 a22 b22 a23 b23
ca11 ca12 ca13 and cA = c = ca 21 ca 22 ca 23
The vector space Mm,n of all mn real matrices with the standard operations.
The vector space P2 of all polynomials of degree 2 or less with the usual operations. 2 2 Let p(x) = a0 + a1x + a2x and q(x) = b0 + b1x + b2x . def Define the usual operations (p + q)(x) p(x) + q(x) and (cp)(x) c[p(x)]
We can verify closure under addition: 2 2 (p + q)(x) = p(x) + q(x) = a0 + a1x + a2x + b0 + b1x + b2x 2 = (a0 + b0) + (a1 + b1)x + (a2 + b2)x which is a polynomial of degree 2 or less (less if a2 + b2 = 0). Notice that we have used the commutative and distributive properties of real numbers. The other axioms can be verified in a similar manner. . Note that 0(x) = 0 + 0x + 0x2.
The vector space Pn of all polynomials of degree n or less with the usual operations.
The vector space P of all polynomials with the usual operations.
The vector space C(–,) of continuous real- valued functions on the domain (–,) For 8 example, x2 + 1, , sin(x), and ex are vectors 1 x 2 in this space. Addition and scalar multiplication are defined in the usual way. (f + g)(x) f(x) + g(x) and (cf )(x) c[f (x)]
f, g, and f + g are vectors in C(–,), just as u, v, and u + v and are vectors in Rn. f (x) can be
thought of as a component of f, just as ui is a component of u. u has n components: u1, u2, …, 2 un. f has an infinite number of components: …, f (–2), …, f ( 3 ), …, f (0), …, f ( ), ….
The additive identity (zero function) is f0(x) = 0 (the x-axis), and given f (x), the additive inverse of f is [–f ](x) = –[ f (x)].
p. 78 Chapter 4: Vector Spaces. 4.2 Vector Spaces. Another Example of a Vector Spaces . The vector space C[a, b] of continuous real-valued functions on the domain [a, b] over the field R.
The most important reason for defining an abstract vector space using the ten axioms above is that we can make general statements about all vector spaces. I.e. the same proof can be used for Rn and for C[a, b].
Theorem 4.4 Properties of Vector Addition and Scalar Multiplication
Let v be a vector in Vn and let c be a scalar. Then
1) 0v = 0 2) c0 = 0 3) If cv = 0, then c = 0 or v = 0. 4) –1v = –v
Proof of (2): c0 = 0
c0 = 0 Given
c0 = c(0 + 0) Additive identity
c0 =
c0 + –( c0)
c0 + –( c0) =
0 = c0 Additive inverse
Proof of (3): If cv = 0, then c = 0 or c 0. If c 0, then
cv = c0 Given
Multiply both sides by Multiplicative c–1cv = c–10 inverse in R
p. 79 Chapter 4: Vector Spaces. 4.2 Vector Spaces.
c(c–1)v = c–10 Commutative property of multiplication
1v = c–10 Multiplicative inverse in R
v c–10 Scalar identity
v = 0 Theorem 4.3(2) – just proved
Thus, either c = 0 or v = 0.
Examples that are not Vector Spaces
2 2
Z (ordered pair of integers) over the field R. Z is not closed under scalar multiplication, for 1 1 2 example 2 (1, 2) = ( 2 , 1) Z .
Aside on notation: 1 Z means “1 is an element of (is a member) of the set of integers. optional Z means “1 is not an element of the set of integers. Although Z2 satisfied Axioms 1–5 and 10 of a vector space, it is not a vector space because not all axioms are satisfied. .() V = The set of second-degree polynomials is not a vector space because it is not closed under addition. For example, let p(x) = x2 and q(x) = –x2 + x + 1. Then p(x) + q(x) = x + 1 is a first degree polynomial.
Let V = R2 with the standard vector addition but nonstandard scalar multiplication defined by
c(u1, u2) = (cu1, 0). Show that V is not a vector space.
It turns out that the only axiom that is not satisfied in this case is (10) Scalar identity. For example, 1(2, 3) = (2, 0) (2, 3).
p. 80 Chapter 4: Vector Spaces. 4.2 Vector Spaces.
Another Example that is not a Vector Spaces
Rotations in three dimensions represented as arrows using the right-hand rule. The direction of the arrow represents the direction of the rotation, via
the right-hand rule, while the length of the arrow represents the optional magnitude of the direction in degrees. Scalar multiplication is the standard operation (stretching the arrow, or reversing the direction if
the scalar is negative). “Vector addition” (e.g. ) is the first rotation followed by
the second. This is not a vector space because “vector addition” is not commutative.
.
(In Chapter 6, we will see that rotations can be represented not as vectors, but as matrices.)
p. 81
Chapter 4: Vector Spaces. 4.3 Subspaces of Vector Spaces.
4.3 Subspaces of Vector Spaces. Objective: Determine whether a subset W of a vector space V is a subspace of V. Objective: Determine subspaces of Rn.
Many vector spaces are subspaces of larger spaces.
.() A V nonempty = subset W of a vector space V is a called a subspace of V when is a vector space under the operations of vector addition and scalar multiplication defined in V.
Theorem 4.5 Test for a Subspace
If W is a nonempty subset of V, then W is a subspace of V if and only if the following .() conditionsV = hold.
0) W is not empty. 1) If u and v are in W, then u + v is in W. Closure under addition 2) If u is in W and c is a scalar, then cu is in W. Closure under scalar multiplication
Proof: If W is a subspace of V, then W is a vector space satisfying the closure axioms, so u + v is in W and cu is in W. On the other hand, assume W is closed under vector addition and scalar multiplication. By assumption, two axioms are satisfied. 1) u + v is in W. Closure under addition 6) cv is a vector in W. Closure under scalar multiplication Then if u, v, and w are in W then they are also in V, so the following axioms are automatically satisfied. 2) u + v = v + u Commutative property 3) (u + v) + w = u + (v + w) Associative property 7) c(u + v) = cu + cv Distributive property over vector addition 8) (c + d)u = cu + du Distributive property over scalar addition 9) c(du) = (cd)u Associative property 10) 1u = u Scalar identity Because W is closed under scalar multiplication, we know that for any v in W, 0v and (–1)v are also in W. From Thm. 4.4, we know that 0v = 0 and (–1)v = –v so the remaining axioms are also satisfied. 4) W contains the zero vector 0. Additive identity 5) For every v in W, W contains –v. Additive inverse (opposite)
p. 82 Chapter 4: Vector Spaces. 4.3 Subspaces of Vector Spaces. Examples of Subspaces and Sets that are not Subspaces.
Show that the set W = {(v1, 0, v3): v1 and v3 are real numbers} is a subspace of R3 with the standard operations.
Graphically, W is the x-z plane in R3. W is nonempty, because it contains (0, 0, 0). W is closed under addition because if u, v W, then u + v = (u1, 0, u3) + (v1, 0, v3) = (u3 + v1, 0, u3 + v3) W. W is closed under scalar multiplication because if c is a scalar and v W, then cv = c(v1, 0, v3) = (cv1, 0, cv3) W.
Is Z2 (ordered pair of integers) with the standard operations a subspace of R2?
Z2 is closed under addition. But as we saw in 4.2, Z2 is not closed under 1 1 2 2 scalar multiplication, for example 2 (1, 2) = ( 2 , 1) Z . So Z is not a subspace of R2.
Show that the set W = {(v1, v2): v1 = 0 or v2 = 0} with the standard operations is not a subspace of R2.
W is closed under scalar multiplication, but W is not closed under addition. * For example, (1, 0) + (0, 1) = (1, 1) W.
Is that the set W = {(v1, v2): v1 0 and v2 0} with the standard operations a subspace of R2?
W is closed under addition, but W is not closed under scalar multiplication when c < 0 and v 0. For example, (–1)(2, 3) = (–2, –3) W.
In general, a subspace is “straight” (line) or “flat” (plane or higher-dimensional object), is “infinite” in all directions, has no “holes,” and contains the origin.
Let W be the set of all symmetric 22 matrices. Show that W is a subspace of M2,2 with the standard operations.
T 0 0 “A is symmetric” means that A = A . W is nonempty, because it contains . W is closed 0 0 under addition because if A, B W, then (A + B)T = AT + BT = A + B, so A + B is symmetric. W is closed under scalar multiplication because if c is a scalar and A W, then (cA)T = c(AT) cA, so cA is symmetric.
p. 83 Chapter 4: Vector Spaces. 4.3 Subspaces of Vector Spaces. More Examples of Subspaces and Sets that are not Subspaces.
Let W be the set of all singular matrices of order 2. Show that W is not a subspace of M2,2 with the standard operations.
1 0 0 0 1 0 0 0 1 0 and are singular matrices (det = 0), but + = is not 0 0 0 1 0 0 0 1 0 1 singular, so W is not closed under addition and is therefore not a subspace. (Notice that a b W = { : ad – bc = 0}, and ad – bc = 0 is not a linear equation.) c d
2 2 2 Let W be the unit circle in R , i. e. W = {(v1, v2): v1 + v2 = 1}, with the standard operations. Is W a subspace of R2
No. W is not closed under addition: (1, 0) + (0, 1) = (1, 1) W. Another reason is that W is not closed under scalar multiplication: 0(1, 0) = (0, 0) W.
Is W = {(0, 0)} with the standard operations a subspace of R2
Yes, the only addition to check is 0 + 0 = 0 W. Scalar multiplication is also easy to check: c0 = 0 W. Is W = {0} and W = V are called the trivial subspaces of c.
Which of these two subsets is a subspace of R3 with the standard operations? 3 a) U = {(v1, v2, v3) R : 2v1 + 3v2 + 4v3 = 12} This is a the equation of a plane through the points (6, 0, 0), (0, 4, 0), and (0, 0, 3). U is not closed under addition, because (6, 0, 0) + (0, 4, 0) = (6, 4, 0), but 2(6) + 3(4) + 4(0) = 24 12. Moreover, U is not closed under scalar multiplication, because 0(6, 0, 0) = (0, 0, 0), but 2(0) + 3(0) + 4(0) = 0 12. Either one of these reasons is enough to show that U is not a subspace.
3 b) W= {(v1, v2, v3) R : 2v1 –3 v2 + 4v3 = 0} This is a the equation of a plane parallel to U, but W contains 0. If v and w W, then
v + w = (v1, v2, v3) + (w1, w2, w3) = (v1 + w1, v2 + w2, v3 + w3), and 2(v1 + w1) –3(v2 + w2) + 4(v3 + w3) = 2v1 + 2w1 –3v2 –3w2 + 4v3 + 4w3
= (2v1 –3v2 + 4v3) + (2w1 –3w2 + 4w3) = 0 + 0 = 0, so v + w W.
Also, if c is a scalar and v W, then cv = (cv1, cv2, cv3) and 2(cv1) –3(cv2) + 4(cv3)
= c(2v1 –3v2 + 4v3) = c·0 = 0, so cv W. Therefore, W is a subspace.
p. 84 Chapter 4: Vector Spaces. 4.3 Subspaces of Vector Spaces. More Example of Subspaces.
P[0, 1] is the set of all real-valued polynomial funtions on the domain [0, 1]. C1[0, 1] is the set of all real-valued, continuously differentiable functions on the domain [0, 1]. C0[0, 1], also written as C[0, 1], is the set of all real-valued, continuous functions on the domain [0, 1]. Let W be the set of all real-valued, integrable functions on the interval [0, 1]. Let V be the set of all real-valued functions on the interval [0, 1].
If we take the usual definitions of vector addition and scalar multiplication, and we use R as the field of scalars, then V is a vector space, and the other four sets are subspaces of V. In fact, P[0, 1] is a subspace of C1[0, 1], which is is a subspace of C0[0, 1], which is is a subspace of W, which is is a subspace of V.
Theorem 4.6 The Intersection of Two Subspaces is a Subspace.
If V and W are both subspaces of a vector space U, then the interaction of V and W (denoted V W) is also a subspace of U. (Note: V W is the set of all vectors that are in both V W.)
Proof: Because V and W are both subspaces of U, we know that they both .() V =contain 0, so V W contains 0 and is not empty. To show that V W is closed under addition, let u1 and u2 be two vectors in V V W. Because u1 and u2 are both in V, and—being a subspace—V is closed, u1 + u2 is also in V. Likewise, u1 + u2 is also in W. Since V W U u + u is in both V and W, u + u is in V W, so V W is closed 1 2 1 2 W under vector addition. A similar argument shows that V W is closed under scalar multiplication, so V W is a subspace of U.
p. 85 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
4.4 Spanning Sets and Linear Independence. Objective: Write a vector as a linear combination of other vectors in a vector space V. Objective: Determine whether a spanning set S of vectors in a vector space is a spanning set of V. Objective: Determine whether a set of vectors in a vector space V is linearly independent. Objective: Prove results about spanning sets and linear independence.
A vector v in a vector space V is called a linear combination of the vectors u1, u2, …, uk in V if and only if v can be written in the form
v = c1u1 + c2u2 + … + ckuk where c1, c2, …, ck are scalars.
Example: Write the vector v = (–1, –2, 2) as a linear combination of the vectors in the set S = {(2, –1, 3), (5, 0, 4)} (if possible).
Solution: We need to solve v = c1u1 + c2u2 for the scalars ci. Substituting in the given vectors, we have
(–1, –2, 2) = c1(2, –1, 3) + c2(5, 0, 4)
(–1, –2, 2) = (2c1, –1c1, 3c1) + (5c2, 0c2, 4c2)
(–1, –2, 2) = (2c1 + 5c2, –1c1 + 0c2, 3c1 + 4c2) cc 152 62 1 21 c which gives the system cc 201 or 01 1 2 21 c2 cc 21 243 43 2 Solve this system by finding the reduced row-echelon form (using software) 162 201 01 2 c1 201 110 or 10 1 so c1 = 2 and c2 = –1. c2 243 000 00 0
Answer: (–1, –2, 2) = 2(2, –1, 3) – 1(5, 0, 4).
1 4 1 Example: Write the vectors v = (–3, 15, 18) and w = ( 3 , 3 , 2 ) as linear combinations of the vectors in the set S = {(2, 0, 7), (2, 4, 5), (2, –12, 13)} (if possible).
Solution: We need to solve
p. 86 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
.
24 48 Example: Write the vector v = as linear combinations of the vectors in the set 57 20 1 2 2 7 4 9 6 5 S = { , , , } (if possible). 5 4 6 2 11 12 4 5 Solution:
p. 87 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
Let S = {v1, v2, …, vk} be a subset of a vector space V. Then S is called a spanning set of V if and only if every vector in V can be written as a linear combination of vector in S. In such cases, we say that S spans V.
Examples of spanning sets: 3 3 The set {(1, 0, 0), (0, 1, 0), (0, 0, 1)} spans R because any vector v = (v1, v2, v3) in R can be written as v = v1(1, 0, 0) + v2(0, 1, 0) + v3(0, 0, 1).
2 2 The set {1, x, x } spans P2 because any vector (polynomial) p(x) = ax + bx + c in P2 can be written as p(x) = c(1) + b(x) + a(x2).
3 Example: Determine whether the set S1 = {(5, 7, 6), (4, 2, 4), (1, –3, 2)} spans R .
3 R consists of all the vectors of the form (v1, v2, v3), where v1, v2, and v3 are real numbers. S1 3 spans R if we can always solve for the scalars c1, c2, and c3 in the equation
(v1, v2, v3) = c1(5, 7, 6) + c2(4, 2, 4) + c3(1, –3, 2)
(v1, v2, v3) = (5c1, 7c1, 6c1) + (4c2, 2c2, 4c2) + (1c3, –3c3, 2c3)
(v1, v2, v3) = (5c1 + 4c2 + 1c3, 7c1 + 2c2 – 3c3, 6c1 4c2 + 2c3)
v1 5c1 4c2 1c3 This vector equation is equivalent to the system v2 7c1 2c2 3c3 v3 6c1 4c2 2c3
5 4 1 c1 v1 or the matrix equation 7 2 3c v . 2 2 6 4 2 c3 v3
p. 88 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
5 4 1
This equation can always be solved for c1, c2, and c3, because 7 2 3 0 , so the matrix is 6 4 2 3 invertible. (To be precise, the determinant is –32.) Therefore, S1 spans R .
3 Example: Determine whether the set S2 = {(5, 7, 6), (3, 2, 4), (1, –3, 2)} spans R . only this number has changed.
Solution:
Geometrically, the vectors in S2 all lie in the same plane (a “2-dimensional” object), while the vectors in S1 do not. We need a “3-dimensional” space to contain S1.
S1 S2.
p. 89 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
If S = {v1, v2, …, vk} is a set of vectors in a vector space V, then the span of S is the set of all linear combinations of the vectors in S.
span(S) = {c1v1 + c2v2 + …+ ckvk: c1, c2, … ck, are scalars}
We sometimes write span{v1, v2, …, vk} instead of span(S).
Another notation for span(S)—which we will avoid because it will be confusing later—is v1, v2, …, vk.
Theorem 4.7 Span(S) is a Subspace of V
* If S = {v1, v2, …, vk} is a set of vectors in a vector space V, then span(S) is a subspace of V. Moreover, span(S) is the smallest subspace of V that contains (S), in the sense that every subspace that contains S must also contain span(S).
Proof First, we want to show that span(S) is a subspace of V.
So we need to show that
Let c be a scalar and let u and w be any vectors in span(S). Then
Next, we want to show that every subspace that contains S must also contain span(S). This is Lab Problem 4.4.55.
p. 90 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
Sometimes, one vector can be written “in terms of” other vectors. For example, in S2 = {(5, 7, 6), (5,7,6) (1,3,2) (3, 2, 4), (1, –3, 2)} from above, (3, 2, 4) = . We could say that (3, 2, 4) “is 2 dependent” upon (5, 7, 6) and (1, –3, 2). But we could just as easily solve for (5, 7, 6); there is no good reason to treat (3, 2, 4) as special. A more equitable equation is
1(5, 7, 6) – 2(3, 2, 4) + 1(1, –3, 2) = (0, 0, 0).
In fact, there are an infinite number of solutions to c1(5, 7, 6) + c2(3, 2, 4) + c3(1, –3, 2) = 0.
c1 = t, c2 = –2t, c3 = t. (This includes c1 = c2 = c3 = 0.)
On the other hand, for S1 = {(5, 7, 6), (4, 2, 4), (1, –3, 2)}, the only solution to
5 4 1 c1 0 c (5, 7, 6) + c (4, 2, 4) + c (1, –3, 2) = (0, 0, 0) or 7 2 3c 0 1 2 3 2 6 4 2 c3 0 1 c1 5 4 1 0 0 is c 7 2 3 0 0 . This is called the trivial solution. 2 c3 6 4 2 0 0
If S = {v1, v2, …, vk} is a set of vectors in a vector space V, then S is called linearly independent if and only if the equation
c1v1 + c2v2 + …+ ckvk = 0
has only the trivial solution c1 = c2 = …= ck = 0. S is called linearly dependent if and only if there are also nontrivial solutions.
Examples #1-4: Let w1 = (7, 3, –5), w2 = (1, 4, 6), w3 = (9, 11, 7), w4 = (12, –5, –4), and w5 = (–6, 5, 8).
3 Example #1: Is {w1, w2} linearly independent? Does {w1, w2} span R ? Solution: To decide linear independence, we want to solve c1w1 + c2w2 = 0. 3 To decide spanning R , we want to solve c1w1 + c2w2 = (v1, v2, v3) for arbitrary v1, v2, and v3.
Both equations look like c1(7, 3, –5) + c2(1, 4, 6) = (v1, v2, v3). (For linear independence, the right-hand side is v1 = v2 = v3 = 0.)
p. 91 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
So we have (7c1, 3c1, –5c1) + (1c2, 4c2, 6c2) = (v1, v2, v3)
(7c1 + 1c2, 3c1 + 4c2, –5c1 + 6c2) = (v1, v2, v3)
7 1 v c 1 3 4 1 v 2 c2 5 6 v3
7 1 3 4 is not square, so we cannot take the determinant as we did before. 5 6
7 1 v1 Instead, find the reduced row-echelon form: 3 4 v 2 5 6 v3
(Don’t worry about reproducing this result; I’ll explain soon why Mathematica and the TI-89 give a different answer.)
Writing this as a system of equations, we have
When the right-hand side is v1 = v2 = v3 = 0, we have only the trivial solution c1 = c2 = 0, because each variable ci corresponds to a pivot. Therefore, {w1, w2} is linearly independent.
On the other hand, there are many choices of v1, v2, and v3 47 25 for which the last equation 0 = v1 – 38 v2 + 38 v3 can be solved. This happens whenever the (reduced) row- echelon form has a row of all zeroes. Therefore,{w1, w2} does not span R3.
Notice that to answer these questions we didn’t need to pay attention to the coefficients on the right-hand side of the line in the augmented matrix. All that we needed to know was the coefficient matrix on the left-hand side of the line.
1 0 0 (Mathematica and the TI-89 both give 0 1 0 because they assume 0 0 1
you can divide Row 3 by v1 – v2 + v3, and they do not know that we intend the last
p. 92 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence. column to be on the right-hand side of the equation. However, the right-hand column is not important to deciding linear independence and spanning.)
3 Example #2: Is {w1, w2, w3} linearly independent? Does {w1, w2, w3} span R ? Solution: To decide linear independence, we want to solve c1w1 + c2w2 + c3w3 = 0. 3 To decide spanning R , we want to solve c1w1 + c2w2 + c3w3 = v for arbitrary v.
c1 v1 Using block multiplication notation, we write [ w | w | w ] c = v 1 2 3 2 2 c3 v3
7 1 9 c1 v1 so 3 4 11 c = v . 2 2 5 6 7 c3 v3
is singular (not invertible) because its
determinant (found using software) is zero. This tells us that sometimes we cannot solve c1w1 + c2w2 + c3w3 = v, so {w1, 3 w2, w3} does not span R . This also tells us that c1w1 + c2w2 + c3w3 = 0 has an infinite number of solutions, so {w1, w2, w3} is linearly dependent.
7 1 9 1 0 1 For another perspective, the reduced row-echelon form of 3 4 11 is 0 1 2 . 5 6 7 0 0 0 To investigate linear independence, we set the right-hand side equal to zero. The third column in the matrix, which doesn’t have a pivot, gives us a free parameter. (Fewer pivots than variables.)
1c1 0c2 1c3 0 c3 t 0c1 1c2 2c3 0 has the solutions c2 2t where t is a free parameter. 0c1 0c2 0c3 0 c1 t Since we have non-trivial solutions, {w1, w2, w3} is linearly dependent.
To investigate spanning, we set the right-hand side equal to arbitrary numbers. The third row of all zeroes in the matrix gives us 0c1 + 0c2 + 0c3 = #, which does not have a solution when the right-hand side is not zero. (Fewer pivots than equations.) Therefore, {w1, w2, w3} does not span R3.
3 Example #3: Is {w1, w2, w4} linearly independent? Does {w1, w2, w4} span R ? Solution: To decide linear independence, we want to solve c1w1 + c2w2 + c4w4 = 0. 3 To decide spanning R , we want to solve c1w1 + c2w2 + c4w4 = v for arbitrary v.
p. 93 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence.
c1 v1 Using block multiplication notation, we write [ w | w | w ] c = v 1 2 4 2 2 c4 v3
7 1 12 c1 v1 so 3 4 5 c = v . 2 2 5 6 4 c4 v3
is invertible (non-singular) because its
determinant (found using software) is 591. This tells us that we can solve c1w1 + c2w2 + c4w4 = v, so {w1, w2, 3 w4} spans R . This also tells us that c1w1 + c2w2 + c4w4 = 0 has a unique of solution, so {w1, w2, w4} is linearly independent.
3 Example #4: Is {w1, w2, w4, w5} linearly independent? Does {w1, w2, w4, w5} span R ? Solution: To decide linear independence, we want to solve c1w1 + c2w2 + c4w4 + c5w5 = 0. 3 To decide spanning R , we want to solve c1w1 + c2w2 + c4w4 + c5w5 = v for arbitrary v.
c1 c2 Using block multiplication notation, we write [ w1 | w2 | w4 | w5] = c4 c5 c 7 1 12 6 1 c 2 so 3 4 5 5 = . c4 5 6 4 8 c5
We cannot take the determinant of a non-square matrix. The reduced row-echelon form of 128 7 1 12 6 1 0 0 591 3 4 5 5 is 0 1 0 506 . 591 263 5 6 4 8 0 0 1 591
To investigate linear independence, we see that the fourth column in the matrix (which doesn’t have a pivot) gives us a free parameter, so {w1, w2, w4, w5} is linearly dependent.
To investigate spanning, we see that there is no row of all zeroes (every row has a pivot), so 3 we will never have an inconsistent equation 0 = #. Therefore, {w1, w2, w4, w5} spans R .
p. 94 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence. Summary:
n {w1, …, wk} in R is linearly independent if and* only if the reduced row-echelon form of the matrix [w1 | … | wk ] has as many pivots as columns (variables). n {w1, …, wk} spans R if and only if the reduced row-echelon form of the matrix [w1 | … | wk ] has as many pivots as rows (equations), i.e. no rows of all zeroes.
Theorem 4.8 A Property of Linearly Dependent Sets
A set S = {v1, v2, …, vk}, k 2, is linearly dependent if and only if at least one of the vectors vj can be written as a linear combination of the other vectors in S. Proof
To prove “only if (),” assume set S is linearly dependent and k 2.
Then
Because one of the coefficients must be nonzero, we can assume without loss of generality (WOLOG) that c1 0. Then *
Conversely (which means changing to ), suppose v1 is a linear combination of the other vectors in S. Then
v1 =
0 =
Thus, {v1, v2, …, vk} is linearly dependent because
Example: Given that {w1, w2, w4, w5} is linearly dependent, and that
c5 t 263 c4 509t c1w1 + c2w2 + c4w4 + c5w5 = 0 has the solutions , write w2 as a linear c 506t 2 509 128 c2 509t combination of the other vectors in the set.
p. 95 Chapter 4: Vector Spaces. 4.4 Spanning Sets and Linear Independence. Solution:
128 506 263 509 tw1 – 509 tw2 + 509 tw4 + tw5 = 0
so 128w1 + 263w4 + 509w5 = 506w2
128 263 509 so w2 = 506 w1 + 506 w4 + 506 w5
Theorem 4.9 Corollary
Two vectors v1 and v2 in a vector space V are linearly dependent if and only if one is a scalar multiple of the other. *
Proof: From the proof of Theorem 4.8,
Theorem
Any set S = {0, v1, v2, …, vk}, containing the zero vector is linearly dependent. *
Proof: 10 + 0v1 + 0v2 + … + 0vk = 0.
Since not all coefficients are zero, S is linearly dependent.
p. 96
Chapter 4: Vector Spaces. 4.5 Basis and Dimension.
4.5 Basis and Dimension. n Objective: Recognize bases in the vector spaces R , Pn, and Mm,n. Objective: Find the dimension of a vector space. Objective: Prove results about spanning sets and linear independence.
A set of vectors S = {v1, v2, …, vk} in a vector space V is called a basis for V when
1) S spans V, and 2) S is linearly independent.
“S spans V” says that the set S is not too small to be a basis; “S is linearly independent” says that the set S is not too big to be a basis. (The plural of “basis” is “bases.”)
3 In the diagram on the left, you can see that S1 = {u1, u2} is too small to span R . For example, v is outside of span(S1). In the figure on the right, you can see that S1 = {u1, u2, u3} is large enough to span R3. For example, you can see the linear combination that gives v. Finally, In the figure on the right, you can see that S3 = {u1, u2, u3, v} is large to be linearly independent, because v is a linear combination of u1, u2, and u3.
p. 97 Chapter 4: Vector Spaces. 4.5 Basis and Dimension. Example. The standard basis for R3: Show that {(1, 0, 0), (0, 1, 0), (0 0, 1)} is a basis for R3.
Solution:
To decide linear independence, we want to solve c1(1, 0, 0) + c2(0, 1, 0) + c3(0 0, 1) = 0. 3 To decide spanning R , we want to solve c1(1, 0, 0) + c2(0, 1, 0) + c3(0 0, 1) = v for arbitrary v. * 1 0 0 c1 0 1 0 0 c1 v1 These equations are 0 1 0 c = 0 and 0 1 0 c = v 2 2 2 0 0 1 c3 0 0 0 1 c3 v3 The first equation has only the trivial solution, and the second equation always has a solution, so {(1, 0, 0), (0, 1, 0), (0 0, 1)}is linearly independent and it spans R3. Therefore, {(1, 0, 0), (0, 1, 0), (0 0, 1)} is a basis for R3.
e1 1, 0, , 0 n e 2 0, 1, , 0 The standard basis in R is e n 0, 0, , 1
Example: Show that {(7, –2, 5), (–3, –9, –1), (1, 7, –7)} is a basis for R3.
Solution: To decide linear independence, we want to solve
.
To decide spanning R3, we want to solve
* As matrix equations, we obtain
The matrix is invertible (its determinant is 474). Because of this, the first equation has only the trivial solution, and the second equation always has a solution, so {(7, –2, 5), (–3, –9, –1), (1, 7, –7)} is linearly independent and it spans R3. Therefore, {(7, –2, 5), (–3, –9, –1), (1, 7, –7)} is a basis for R3.
p. 98 Chapter 4: Vector Spaces. 4.5 Basis and Dimension. 2 n The standard basis for Pn is {1, x, x , …, x }.
1 0 0 1 0 0 0 0 0 0 0 0 The standard basis for M2,3 is { 0 0 , 0 0 , 1 0 , 0 1 , 0 0 , 0 0 }. 0 0 0 0 0 0 0 0 1 0 0 1
Theorem 4.9 Uniqueness of a Basis Representation
If S = {v1, v2, …, vk} is a basis for a vector space V, then every vector in V, can be written in one and only one way as a linear combination of vectors in S.
Proof: Let u be an arbitrary vector in V,
u can be written in at least one way as a linear combination of vectors in S because
Now suppose that u can be written as a linear combination of vectors in S in two ways:
u = b1v1 + b2v2 + …+ bkvk and u = c1v1 + c2v2 + …+ ckvk * Subtracting these two equations gives
Since S is linearly independent,
Thus, u can be written as a linear combination of vectors in S in only one way.
Example: Using the basis {(7, –2, 5), (–3, –9, –1), (1, 7, –7)} for R3 from the previous example,
find the unique representation of u = (u1, u2, u3) using this basis. In other words, find the unique solution for the constants c1, c2, c3 in the equation
u = c1(7, –2, 5), + c2(–3, –9, –1), + c3(1, 7, –7)}
p. 99 Chapter 4: Vector Spaces. 4.5 Basis and Dimension.
Solution: As a matrix equation, we obtain
Lemma 4.9½ Dimension of Spanning Sets and Linearly Independent Sets
If S1 = {v1, v2, …, vn} spans a vector space V and S2 = {u1, u2, …, um} is a set of m linearly independent vectors in V, then m n.
Proof by contradiction:
Suppose m > n. To show that S2 is linearly dependent (a contradiction), we need to find scalars k1, k1, …, km (not all zero) such that
u1 k u + k u + … + k u = k k = 0 1 1 2 2 m m 1 m 1m u m m1
Because S1 spans V, each ui is a linear combination of vectors in S1:
u1 v1 u1 C11v1 ... C1n v n C i.e. mn u v u C v ... C v m m1 n n1 m m1 1 mn n
Substituting into the first equation gives
v1 k k C = 0 1 m 1m mn v n n1
k1 0 Now consider [C T ] nm k 0 m m1 n1
This is n equations for m unknowns (ki), with n < m, so we have (infinitely many) nontrivial (non-zero) solutions (k1, k2, …, km). Taking the transpose, we have a non-zero solution to
p. 100 Chapter 4: Vector Spaces. 4.5 Basis and Dimension.
0 01n =k1 kn 1m Cmn Therefore, we have a non-zero solution to
v1 u1 0 = k k C = k k 1 m 1m mn 1 m 1m v u n n1 m m1
This says that S2 = {u1, u2, …, um} is linearly dependent, which contradicts the premise, thus completing the proof by contradiction.
Theorem 4.10 Bases and Linear Dependence * If S = {v1, v2, …, vn} is a basis for a vector space V, then every set containing more than n vectors in linearly dependent.
Proof:
S = {v1, v2, …, vn} spans V so by Lemma 4.9½, any set of m linearly independent vectors in V has m n. Therefore, any set of m vectors in V where m > n must be linearly dependent.
Note: the last step is true because “P Q” is logically equivalent to its contrapositive “(not Q) (not P)”.
Theorem 4.11 Bases and Linear Dependence * If a vector space V has one basis with n vectors, then every basis for V has n vectors.
Proof:
Let S1 = {u1, u2, …, un} be one basis for Vand let S2 = {v1, v2, …, vm} be another basis for V. Because S1 is a basis, and S2 is linearly independent, Thm. 4.10 tells us that m < n. Similarly, because S2 is a basis, and S1 is linearly independent, Thm. 4.10 tells us that n < m. Therefore, m = n.
If a vector space V has basis consisting of n vectors,* then the number n is called the dimension of V, denoted by dim(V) = n. We define the dimension of the trivial vector space {0} to be n = 0.
The dimension of a vector space is well-defined (unambiguous) because of Thm. 4.11.
p. 101 Chapter 4: Vector Spaces. 4.5 Basis and Dimension. Examples:
dim(Rn) = n
dim(Pn) = n + 1
dim(Mm,n) = mn
Example: Find the dimension of the subspace of R3 given by W= {(2a, a –3b, a + b): a and b are real numbers}.
Solution:We can write (2a, a –3b, a + b) = a(2, 1, 1) + b(0, –3, 1), so W is spanned by S = . {(2, 1, 1), (0, –3, 1)}. Moreover, S is linearly independent, because the reduced row echelon 2 0 1 0 0 a form of 1 3 is 0 1 (every column has a pivot). The only solution to = 0 b 1 1 0 0 0 is a = b = 0. So S is a basis, and dim(W) = 2.
Example: Find the dimension of the subspace W of P4 spanned by S = {–3 – 5x + x2 + 2x4, –5x – 4x2 + 4x3, –12 – 10x + 11x2 – 12x3 + 11x4, 3 + 10x + 4x2 – 5x4}
Solution: S spans W, but it might not be linearly independent. To decide linear independence, we solve
p. 102 Chapter 4: Vector Spaces. 4.5 Basis and Dimension. Theorem 4.12 Basis Tests in an n-Dimensional Space
Let V be a vector space of dimension n. *
1) If S = {v1, v2, …, vn}is a linearly independent set of vectors in V, then S is a basis for V. 2) If S = {v1, v2, …, vn}spans V, then S is a basis for V.
Proof by Contradiction of Part(1) Assume that S is not a basis for V. Since S is linearly independent, it must not span V. Choose a vector u in V that is not in span(S) Then the set {v1, v2, …, vn, u} is also linearly independent.
To see this, note that c1v1 + c2v2 + …+ cnvn = u has no solution.
c1v1 + c2v2 + …+ cnvn = –cn+1u
has a solution only when cn+1 = 0, and in that case, c1 = c2 = … = cn = 0 because {v1, v2, …, vn} is linearly independent.
{v1, v2, …, vn, u} being linearly independent contradicts Thm. 4.10—we have n + 1 linearly independent vectors in an n-dimensional vector space. So S must be a basis for V
Proof by Contradiction of Part(2) Assume that S is not a basis for V. Since S spans V, it must not be linearly independent. So c1v1 + c2v2 + …+ cnvn = 0 has a solution where not all of the ci are zero. Without loss of generality, we can assume that cn 0. Then the set {v1, v2, …, vn–1}also spans V.
c1 c2 cn1 To see this, first observe that vn = – v1 – v2 – … – vn–1. cn cn cn Then
p. 103
Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations.
4.6 Rank of a Matrix and Systems of Linear Equations. Objective: Find a basis for the row space, a basis for the column space, and the rank of a matrix. Find the nullspace of a matrix. Find the solution of a consistent system Ax = b in the form xp + xh. Objective: Prove results about subspaces associated with matrices.
Given an mn matrix A, in this section we will speak of its row vectors, which are in Rn
11 12 aaa1n 11 12,, aaa1n aaa ,, aaa A = 21 22 2n 21 22 2n mm31 aaamn m m21,, aaamn
and its column vectors, which are in Rm
11 12 aaa 1n a11 a12 a1n aaa a a a A = 21 22 2n 21 22 2n mm 31 aaa mn a 1 amm 2 amn
Given an mn matrix A.
The row space of A is the subspace of Rn spanned by the row vectors of A. The column space of A is the subspace of Rm spanned by the column vectors of A.
Theorem 4.13 Row-Equivalent Matrices Have the Same Row Spaces .( If an mn matrix A is row-equivalent to an mn matrix U, then the row space of A is equal to the row space of U.
Proof Because the rows of U can be obtained from the rows of A by elementary row operations (scalar multiplication and addition), it follows that the row vectors of U are linear combinations of the row vectors of A. The row vectors of U lie in the row space of A, and the subspace .( spanned by the row vectors of U is contained in the row space of A: span(U) span(A). Since the rows of A can also be obtained from the rows of U by elementary row operations, we also have that subspace spanned by the row vectors of A is contained in the row space of U: span(A) span(U). Therefore, the two row spaces are equal: span(U) = span(A).
p. 105 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Theorem 4.14 Basis for the Row Space of a Matrix .( If a matrix A is row-equivalent to a matrix U in row-echelon form, the then the nonzero row vectors of U form a basis for the row space of A.
34291 26410 Example: Show that the nonzero rows of 31000 (a matrix in row-echelon form) 10000 00000 are linearly independent.
c1 34291 c 26410 Solution: We want to solve 00000 = 2 c3 31000
c4 10000
The first column of the right-hand side sum is c1, so c1 = 0. The second column of the right- hand side sum is –9c1 + c2 = 0 + c2, so c2 = 0. The fourth column of the right-hand side sum is 0 + 0 + c3 (because c1 = c2 = 0), so c3 = 0. Finally, the fifth column of the right-hand side sum is 0 + 0 + 0 + c4 (because c1 = c2 = c3 = 0), so c4 = 0. Therefore the nonzero rows are linearly independent.
Alternate Solution: The transpose of =
0 1 0 0 0 0001 c1 0 9 1 0 0 0019 c2 is 0 = c1 2 1 + c2 4 + c3 0 + c4 0 = 0042 . c3 0 4 6 1 0 0164 c4 0 3 2 3 1 1323
0 0 Solving this by “forward substitution” (first row down to the last row) yields = . 0 0
p. 106 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations.
4981 4 224433 Example: Find a basis for the row space of 4 457637 9 5211376 40 3032
1 8 9 4 0 1 8 6 Solution: The row-echelon form is 0 0 0 1 , so a basis for the row space is 0 0 0 0 0 0 0 0
{[1, –8, –9, 4], [0, 1, 8, –6], [0, 0, 0, 1]}. 1 0 55 0 0 1 8 0 You could also take the reduced row-echelon form 0 0 0 1 to obtain the basis 0 0 0 0 0 0 0 0 {[1, 0, 55, 0], [0, 1, 8, 0], [0, 0, 0, 1]}.
Example: Find a basis for the subspace of R3 spanned by {[1, –6, 3], [1, –5, 7], [–7, 38, –37]}.
Solution:
Lemma: The pivot columns of a matrix U in row-echelon form (or reduced row-echelon form) are linearly independent.
p. 107 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations.
1 xxxxxx 100 xxxx Example: U = 1000 xxx . If we look only at the pivot columns and solve 100000x 0000000 1 xxx 0 c 0 10 xx 1 0 c 0 100x 2 = 0 , we see by back substitution that the unique solution is = c3 0 1000 0 c4 0 0000 0 , so the pivot columns of U are linearly independent.
Lemma: The pivot columns of a matrix U in row-echelon form (or reduced row-echelon form) span the column space of U.
Example: U = . The column space of U is the subset of R5 consisting
of vectors whose fifth component is zero (because we are taking linear combinations of vectors whose fifth component is zero).
v1 v 2
If we look only at the pivot columns, we can always solve = v3 , by back v4 0 substitution so the pivot columns of U span the column space of U.
Theorem Basis for the Column Space of a Matrix .( Given an mn matrix A and its row-echelon (or reduced row-echelon) form U. Then the columns of A that correspond to the pivot columns of U form a basis for the column space of A.
Proof: by combining the two preceding lemmas, we know that the pivot columns of U form a basis for the column space of U. Let ai be the column vectors of A and ui be the column vectors of U, so A = [a1| …| an ] and U = [u1| …| un ]. We can factor PA = LU, so Pai = Lui for each column vector, –1 –1 ui = L Pai, and ai = P Lui. (essential idea)
p. 108 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. To show that the pivot columns of A are linearly independent, consider the equation 0 = , where the terms in the sum include only the pivot columns. Then c aii ipivots L–1P0 = L–1P , so 0 = 1 = c aii PLcaii c uii ipivots ipivots ipivots
Since the pivot columns of U are linearly independent, all of the ci (i pivots) are zero, so the pivot columns of A are also linearly independent.
To show that the pivot columns of A span the column space of A, consider any vector v in the n column space of A. Then v can be written as v = c aii so i1 n n n –1 –1 1 L Pv = L P c aii = PLcaii = c uii i1 i1 i1 –1 Since L Pv is a linear combination of the ui, it can be written as a linear combination of the pivot columns (which are a basis for the column space of U): –1 –1 –1 –1 1 L Pv = d uii so v = P L(L Pv) = P L = i LPdui = d aii ipivots ipivots ipivots Therefore, the pivot columns of A span the column space of A.
Taken together, the two parts of this proof show that the pivot columns of A form a basis for the column space of A.
Example: Find a basis for the column space of 4981 4 224433 A = 4 457637 . 9 5211376 40 3032
.(
p. 109 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations.
Theorem 4.15 Row Space and Column Space have the Same Dimension. .( If A is an mn matrix, then the row space and column space of A have the same dimension.
Proof: The dimension of the row space and the dimension of the column space are both equal to the number of pivots in the row-echelon form (or in the reduced row echelon form) of A.
.( The rank r = rank(A) of a matrix is the dimension of its row space, the dimension of its column space, and the number of pivots in the row-echelon form (or in the reduced row echelon form).
Theorem 4.16 Solutions to a Homogeneous System Ax = 0
If A is an mn matrix, then the set of all solutions to the homogeneous system of linear equations Ax = 0 is a subspace of Rn.
Proof: Because A is mn, x is n1, so the set of all solutions to the homogeneous system of linear equations Ax = 0 is a subset of Rn. We will call this subset N. N is not empty because it contains 0: A0 = 0. Now we must show …
.(
If A is an mn matrix, the subspace of Rn consisting of all solutions to the homogeneous system Ax = 0 is a called the nullspace of A and is denoted by N(A). We sometimes call N(A) the solution space of Ax = 0. The dimension of N(A) is called the nullity of A, and is sometimes written as (the Greek letter nu).
p. 110 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Theorem 4.17+ Solutions of a Homogeneous System Ax = 0
If A is an mn matrix of rank r and nullity , then the dimension of the solution space of Ax = 0, i.e. N(A), is = n – r. That is, n = rank(A) + nullity(A), or the number of variables = number of pivot variables + number of free variables.
Moreover, the solution set to the homogeneous system Ax = 0 is the same as the solution set to the reduced row-echelon form Rx = 0. This solution has the form
xh = t1x1 + t2x2 + … + tx
where the ti are free parameters, one for each free (non-pivot) variable, and {x1, x2, …, x} is a basis for N(A).
Proof Ax = 0 and Rx = 0 have the same solution set, i.e. N(A) = N(R), because A and R are row- equivalent. .( Ax = 0 iff En…E2E1Ax = En…E2E10 iff Rx = 0
Each column of R is multiplied by a component of x. R has r pivots, corresponding to r pivot variables in x. The remaining n – r variables are free variables, which we can replace with free parameters t1, t2, …, tn–r. Now we can solve r equations (on for each row of R that has a pivot, and is therefore not all zeroes) for r pivot variables by back-substitution. The set of these solutions xh is N(A). If we collect like terms in ti, and factor out the ti, we have
xh = t1x1 + t2x2 + … + tx
so N(A) is spanned by {x1, x2, …, x}. Moreover, {x1, x2, …, x} is linearly independent, because the solution to 0 = t1x1 + t2x2 + … + tx = xh is 0 = xfree(i) = ti and for each free variable and 0 = xpivot(j) = (some linear combination of the ti) for each pivot variable. Since all of the ti are 0, {x1, x2, …, x} is linearly independent, so {x1, x2, …, x} is a basis for N(A), and N(A) has dimension = nullity(A) = n – r. That is, the number of variables = number of pivot variable. + number of free variables.
Example: Find a basis for nullspace of 1 2 1 0 0 2 5 1 1 0 A = . 3 7 2 2 2 4 9 3 1 4
p. 111 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Solution: 1 0 3 0 4 0 0 1 1 0 2 0 The reduced row-echelon form is [ R | 0 ] = . 0 0 0 1 2 0 0 0 0 0 0 0
Now we need to solve the system pivots
x1 1 0 3 0 4 0 1x 3x 4x 0 x 1 3 5 0 1 1 0 2 2 0 1x 1x 2x 0 2 3 5 x3 = or the equivalent form 0 0 0 1 2 0 1x4 2x5 0 x4 0 0 0 0 0 0 0 0 x5
The two free (non-pivot) variables are x3 and x5. We can choose parameters x3 = s and x5 = t, and solve for the pivot variables x1, x2, and x4 using back substitution.
1x1 3s 4t 0 1x2 1s 2t 0 x4 = 2t, x2 = s – 2t, and x1 = –3s + 4t 1x4 2t 0
3s 4t 3 4 s 2t 1 2 That means that N(A) is the set of all vectors of the form = s = s 1 + t 0 2t 0 2 t 0 1
So { , } is a basis for N(A).
p. 112 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Example: Find a basis for nullspace of 1 4 2 1 A = . 0 1 1 1 2 8 4 2
Solution:
Now that we have solved Ax = 0, let us consider Ax = b, where b 0. We have already seen (Ch. 1.2) that Ax = b has an infinite number of solutions if there are more variables than equations (n > m). And since you can’t have more pivots than rows, m r, so n > r, or > 0.
Theorem 4.18 Solutions of a Nonhomogeneous System Ax = b
If xp is a particular solution to the nonhomogeneous equation Ax = b, then every solution of this system can be written in the form x = xp + xh, whereS* x h is a solution of the corresponding homogeneous system Ax = 0, i.e. xh is in N(A). The general solution of Ax = b, parameterized by the free parameters t1, t2, …, t is
x = xp + t1x1 + t2x2 + … + t x
Proof
Let x be any solution of Ax = b. Then (x – xp) is a solution of the homogeneous system Ax = 0, because
A(x – xp) = Ax – Axp = b – b = 0 .( If we let xp = x – xp, then x = xp + xh.
(For ease of calculation, we usually find xp by solving Ax = b when all of the free variables are set to zero.)
From Thm. 4.17+ we know that xh = t1x1 + t2x2 + … + tx.
p. 113 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Example: Find the general solution (the set of all solutions) of the system of linear system.
x1 2x2 x3 x4 5x5 0 5x 10x 3x 3x 55x 8 1 2 3 4 5 x1 2x2 2x3 3x4 5x5 14
1x1 2x2 x3 x4 15x5 2
1 2 1 1 5 0 1 2 0 0 5 1 5 10 3 3 55 8 0 0 1 0 6 2 Solution: rref 1 2 2 3 5 14 0 0 0 1 4 3 1 2 1 1 15 2 0 0 0 0 0 0
1x1 2(0) 5(0) 1
To find xp, set the free variables x2 = 0, x5 = 0. 1x3 6(0) 2
1x4 4(0) 3
1 0 .( So x4 = –3, x3 = 2, x1 = 1. xp = 2 3 0
1x1 2s 5t 0
To find xh, set x2 = s, x5 = t. Ax = 0 1x3 6t 0
1x4 4t 0
2s 5t 2 5 s 1 0 So x4 = –4t, x3 = –6t, x1 = –2s + 5t. xh = 6t = s 0 + t 6 4t 0 4 t 0 1
5 0 x = + s + t 6 4 1
p. 114 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Example: Find the general solution (the set of all solutions) of the system of linear system. x 3y 10z 18 2x 7y 32z 29
x 3y 14z 12 x y 2z 8
Solution:
Example: Find the general solution (the set of all solutions) of the system of linear system. 3x 8y 4z 19 6y 2z 4w 5
5x 22z w 29 x 2y 2z 8
Solution:
p. 115 Chapter 4: Vector Spaces. 4.6 Rank of a Matrix and Systems of Linear Equations. Theorem 4.19 Consistency of a System of Linear Equations
The system Ax = b is consistent if and only if b is in the column space of A. Proof Suppose A is mn, so x is n1 and b is m1. a a a x a a a 11 12 1n 1 11 S* 12 1n a a a x a a a Then Ax = 21 22 2n 2 = 21 x 22 x 2n x (see Section 2.1) 1 2 n am1 am2 amn xn am1 am2 amn
So Ax = b is if and only if b is a linear combination of the columns of A, i.e. b is in the column space of A.
Summary of Equivalent Conditions for Square Matrices
If A is an nn matrix, then the following conditions are equivalent.
1) A is invertible. 2) Ax = b has a unique solution for any n1 matrix b. 3) Ax = 0 has only the trivial solution. 4) A is row-equivalent to In. 5) det(A) 0 6) rank(A) = n 7) The n row vectors of A are linearly independent. 8) The n column vectors of A are linearly independent.
p. 116 Chapter 4: Vector Spaces. 4.7 Coordinates and Change of Basis.
4.7 Coordinates and Change of Basis. Objective: Represent coordinates in general n-dimensional spaces. Objective: Find a coordinate matrix relative to a basis in Rn. Objective: Find the transition matrix from the basis B to the basis B in Rn. Let B = {v1, v2, …, vn} be an ordered basis for a vector space V and let x be a vector in V such that
x = c1v1 + c2v2 + …+ cnvn
The scalars c1, c2, …, cn are called the coordinates of the vector x relative to (or with respect to) the basis B. The coordinate vector (or matrix) of x relative B is the column matrix in Rn whose components are the coordinates of x. S*
c1 c2 [x]B = cn
Example: The coordinate matrix of x = (x1, x2, x3) relative to the standard basis S = {e1, e2, e3} = {(1, 0, 0), (0, 1, 0), (0 0, 1)} in R3 is simply
x1 [x] = x because x = x e + x e + x e S 2 1 1 2 2 3 3 x3
Example: Find the coordinate matrix of p(x) = 2x2 –5x + 6 relative to the standard ordered basis 2 S = {1, x, x } in P2
6 Solution: p(x) = 6(1) – 5(x) + 2(x2) so [p] = S 5 6
28 Example: Find the coordinate matrix of A = relative to the standard ordered basis 71 01 10 00 00 S = { , , , } in M2,2 00 00 01 10
8 2 Solution: A = 8 + 2 + 1 + 7 so [A]S = 1 7
p. 117 Chapter 4: Vector Spaces. 4.7 Coordinates and Change of Basis. Example 1: Given that the coordinate matrix of x in R2 relative to the nonstandard ordered basis B = {v1, v2} = {(1, 0), (1, 2)} is 3 [x]B = 2 Find the coordinate matrix of x relative to the standard basis B = {u1, u2} = {(1, 0), (0, 1)} = S.
Solution: [x]B = means that x = 3v1 + 2v2 = 3(1, 0) + 2(1, 2) = (5, 4).
5 Now (5, 4) = 5(1, 0) + 4(0, 1) = 5u1 + 4u2 so [x]B=S = . 4
Another method is to take the coordinate matrices relative to B of 11 3 x = 3v1 + 2v2 = vv 21 to obtain [x]B=S = B vv ][][ B'2'1 = = . 20 2
This procedure is called a change of basis from B to B. Given [x]B, we found [x]B using an equation of the form
1 –1 [x]B = P [x]B where P = B vv ][][ B'2'1 .
The matrix P–1 is called the transition matrix from B to B.
Example 2: Given that the coordinate matrix of x in R3 relative to the standard ordered basis B = S = {e1, e2, e3} is 1 [x] = . B 2 1
Find the coordinate matrix of x relative to the nonstandard ordered basis B = {u1, u2, u3} = {(1, 0, 1), (0, –1, 2), (2, 3, –5)}.
p. 118 Chapter 4: Vector Spaces. 4.7 Coordinates and Change of Basis. Solution: Write x as a linear combination of the vectors in B:
c1 x = c u + c u + c u = uuu c = [x] = uuu][][][ 1 1 2 2 3 3 321 2 B 1 B 2 B 3 B c3 1 1 201 c1 201 5 Since B = S, 2 = 310 so c = 310 = 8 . 2 1 521 c3 521 2
Thus we have [x]B=S = = .
Moreover, we found that [x]B = [x]B
1 Comparing to [x]B = P [x]B after Example 1above, we must have
[x]B = P[x]B where P = .
The matrix P is called the transition matrix from B to B.
Given nonstandard bases B = {v1, v2, …, vn}, B = {u1, u2, …, un}, and the standard basis S = n {e1, e2, …, en} of R . Define matrices B = vv v ][][][ and (B) = uu u ][][][ . BS 1 S 2 S Sn BS 1 S 2 S Sn S* (B and B are square, invertible matrices.) B is the transition matrix from B to S; B is the –1 transition matrix from B to S. To get P , the transition matrix from B to B, first apply BBS, –1 then apply (B) SB. –1 –1 –1 –1 P = (B) B or (P )BB = (B) SB BBS The inverse is –1 –1 P = B B or PBB = (B )SB (B)BS
p. 119 Chapter 4: Vector Spaces. 4.7 Coordinates and Change of Basis.
2 Example: Given B = {v , v , v } = {(1, 0, 1), (1, 1, 0), (0, 1, 1)} and [x] = . Find [x] 1 2 3 B 3 S 1
Solution:
Example: Given B = {u1, u2, u3} = {(8, 11, 0), (7, 0, 10), (1, 4, 6)} and x =(3, 19, 2). Find [x]B
Solution:
Example: Given B = {v1, v2, v3} = {(1, 0, 2), (0, 1, 3), (1, 1, 1)}, 1 B = {u , u , u } = {(2, 1, 1), (1, 0, 0), (0, 2, 1)}, and [x] = . 1 2 3 B 2 1 –1 Find the transition matrix P from B to B the transition matrix P from B to B, and [x]B.
Solution:
p. 120 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn.
Chapter 5: Inner Product Spaces.
5.1 Length and Dot Product in Rn. Objective: Find the length of a vector and find a unit vector. Objective: Find the distance between two vectors. Objective: Find a dot product and the angle between two vectors, determine orthogonality, and verify the Cauchy-Schwarz Inequality, the triangle inequality, and the Pythagorean Theorem. Objective: Use a matrix product to represent a dot product.
From the Pythagorean Theorem, we know that the length, or norm, of a vector v in R2 is 2 2 3 ||v|| = 1 vv2 . Also from the Pythagorean Theorem, in R we have 2 ||v|| = 2 2 vvv2 = 2 2 vvv2 . 1 2 3 1 2 3
n S* 2 2 2 In R we define the length or norm of v to be ||v|| = 1 2 vvv n
Notice that |v|| 0 ||v|| = 0 if and only if v = 0.
A unit vector is a vector whose length is one. In R2 and R3 alternate notation is sometimes used for the standard unit vectors i = (1, 0), j = (0, 1) in R2 i = (1, 0, 0), j = (0, 1, 0) , k = (0, 0, 1) in R3
Two nonzero vectors u and v are parallel when one of them is a scalar multiple of the other: u = cv. If c > 0, then u and v have the same direction. If c < 0, then u and v have opposite directions.
p. 121 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn. Theorem 5.1 Length of a scalar multiple
Let v be a vector in Rn and let c be a scalar. Then ||cv|| = |c| ||v||, where |c| is the absolute value of c.
2 2 2 2 2 2 2 Proof: ||cv|| = 1 cvcv2 cv n )()()( = 1 2 vvvcn = |c| ||v||
Theorem 5.2 Unit Vector in the Direction of v
v If v is a nonzero vector in Rn, then the vector u = has length one and has the same direction v as v. u is called the unit vector in the direction of v.
1 1 Proof: Because v 0, we know that ||v|| > 0 so > 0. Then u = v so u has the same v v v direction as v. Moreover, ||u|| = = ||v|| = 1, so u is a unit vector. v
The process of finding the unit vector in the same direction as v is called normalizing v.
The distance between two vectors u and v is d(u, v) = ||u – v||. In Rn, 2 2 2 d(u, v) = 11 vuvu 22 vu nn )()()(
Notice that d(u, v) 0 d(u, v) = 0 if and only if u = v. d(u, v) = d(v, u)
Example: Let u = (–2, 3, 4), v = (1, 2, –2). Find the length of u, normalize v, and find the distance between u and v.
Solution by hand: ||u|| = 43)2( 222 = 29 or 5.38516
)2,2,1( 1 2 2 = = , , 22 )2(21 2 3 3 3
d(u, v) = 2 2 2(4()23()12( ))2 =
p. 122 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn.
61)3(222 = 46 = 6.78233
Solution using Mathematica:
u={-2,3,4} Out = {-2,3,4}
Norm[u] Out =
v={1,2,-2} Out = {1,2,-2}
Normalize[v] Out =
Norm[u-v] Out =
Solution using TI-89: [-2,3,4].U
[1,2,-2].V
MatrixNorms norm(U)
MatrixVector ops unitV(V)
MatrixNorms norm(U-V)
To find the angle between to vectors in Rn, we start with the Law of Cosines in R2. u – v ||u – v||2 = ||u||2 + ||v||2 – 2||u|| ||v|| cos( ) v 2 2 2 2 2 2 (u1 – v1) + (u2 – v2) = u1 + u2 + v1 + v2 – 2||u|| ||v|| cos( ) u
2 2 2 2 2 2 2 2 u1 –2u1v1 + v1 + u2 –2u2v2 + v2 = u1 + u2 + v1 + v2 – 2||u|| ||v|| cos( )
–2u1v1 –2u2v2 = – 2||u|| ||v|| cos( )
u1v1 + u2v2 = ||u|| ||v|| cos( )
This equation leads us to the following definitions.
n The dot product between two vectors u = (u1, u2, …, un) and v = (v1, v2, …, vn) in R
u•v = u1v1 + u2v2 + … + unvn
S* n vu The angle between two nonzero vectors u and v in R is = arccos , 0 . vu
p. 123 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn. Two vectors u and v in Rn are orthogonal (perpendicular) when u•v = 0. (Notice that the 0 vector is orthogonal to every vector.)
n We often represent a vectors u = (u1, u2, …, un) and v = (v1, v2, …, vn) in R by their coordinate matrices relative to the standard basis {e1, e2, …, en}
u1 v1 u v u = 2 and v = 2 u n vn T T We then write u v for u•v because u v = [u1v1 + u2v2 + … + unvn]. Warning: In our formulas, we will be sloppy and pretend that uTv is equal to the scalar u•v. However, Mathematica and the TI-89 know that uTv is a 11 matrix, and will not let you use it as a scalar.
Example: Using u = (–2, 3, 4), v = (1, 2, –2) from the previous example, find u•v and the angle between u and v.
Solution by hand: u•v = (–2)(1) + (3)(2) + (4)(–2) = –4 ||u|| = 43)2( 222 = 29 or 5.38516 ||v|| = = 22 )2(21 2 = 3 4 = arccos 1.821 or 104.3 3 29
Solution using Mathematica:
u.v Out = 4 period, not asterisk
ArcCos[u.v/Norm[u]/Norm[v]] Out =
Out = 1.82099
Solution using TI-89: MatrixVector ops U V dotP( , )
Matrix Vector ops dotP(u, v) MatrixNorms norm(u) MatrixNorms norm(v))
Matrix Vector ops dotP(u, v)
p. 124 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn. MatrixNorms norm(u) MatrixNorms norm(v))
To obtain a decimal approximation instead of an exact formula, use N[...] in Mathematica or instead of on the TI-89. You can change modes between degrees and radians on the TI-89 using Angle……… DEGREE
Theorem 5.3 Properties of the Dot Product
If u, v, and w are vectors in Rn and c is a scalar, then
1) u•v = v•u 2) u•(v + w) = u•v + u•w 3) c(u•v) = (cu)•v = u•(cv) 4) v•v = ||v||2 5) v•v 0 and v•v = 0 if and only if v = 0.
Proof of 5.3.2:
u•(v + w) = optional
Rn combined with the usual operations of vector addition, scalar multiplication, vector length, and the dot product is called Euclidean n-space.
(Optional aside: non-Euclidean spaces include elliptic space, which is curved, cannot be smoothly mapped Rn and a has a different definition of distance; hyperbolic space, which is curved and has different definition of distance; and Minkowski space, used in Einstein’s special optional relativity, which replaces the non-negative dot product with an “interval” that can be negative.)
p. 125 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn.
Theorem 5.4 The Cauchy-Schwarz Inequality
If u and v are vectors in Rn, then |u•v| ||u|| ||v||, where |u•v| is the absolute value of u•v. Proof If u = 0, then equality holds, because
|u•v| = |0•v| = |0| = 0, and ||u|| ||v|
= ||0|| ||v|| = 0 ||v|| = 0.
If u 0, then let t be any real number and consider the vector tu + v. Then
0 || tu + v ||2 = S* The left-hand side is a quadratic function f (t) = at2 + bt + c, where
a = , b = , c =
−푏±√푏2−4푎푐 The quadratic formula is 푡 = 2푎
Because f (t) 0, we have either one or zero roots, so 푏2 − 4푎푐 ≤ 0
Theorem 5.5 The Triangle Inequality
If u and v are vectors in Rn, then ||u + v|| ||u|| + ||v||. Proof ||u + v||2 = S*
p. 126 Chapter 5: Inner Product Spaces. 5.1 Length and Dot Product in Rn.
Theorem 5.6 The Pythagorean Theorem
If u and v are vectors in Rn, then u and v are orthogonal if and only if ||u + v||2 = ||u||2 + ||v||2. Proof By definition, u and v are orthogonal if and only if S* ||u + v||2 =
The conclusion follows from these two equations.
p. 127
Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces.
5.2 Inner Product Spaces. Objective: Determine whether a function defines an inner product. n Objective: Find the inner product of two vectors in R , Mm,n, Pn, and C[a,b]. Objective: Use the inner product to find angles between two vectors and determine whether two vectors are orthogonal. Objective: Find an orthogonal projection of a vector onto another vector in an inner product space.
The inner product is an extension of the concept of the dot product from Rn to a general vector space. The standard dot product, also called the Euclidean inner product, on Rn is written as u•v, while the general inner product on a vector space V is written as u, v.
Definition of Inner Product
Let u, v, and w be vectors in a vector space V, and let c be any scalar. An inner product on V is a function that associates a real number u, v with each pair of vectors u and v. and satisfies the * following axioms.
1) u, v = v, u Symmetric 2) u, v + w = u, v + u, w 3) cu, v = cu, v 4) v, v 0 and v, v = 0 if and only if v = 0. Nonnegative
A vector space V with an inner product is called an inner product space.
2 u1 v1 Example: Consider R . Let u = and v = . Instead of the Euclidean inner product u2 v2
T T 12 u v, let u, v = u v = 2u1v1 – u1v2 – u2v1 + 3u2v2. Show that u, v is an inner 31 product.
Solution:
T 1) v, u = v u = 2v1u1 – v1u2 – v2u1 + 3v2u2 = u, v
2) Let A = . We know that for matrix multiplication, u, v + w = uTA(v + w) =
(uTA)(v + w) = (uTA)v + (uTA)w = uTAv + uTAw = u, v + u, w.
3) We know that for matrix multiplication, cu, v = c(uT v) = (cuT) v =
= cu, v.
p. 129 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces.
2 2 2 2 2 2 2 2 2 4) v, v = 2v1 – 2v1v2 + 3v2 = v1 + 2v2 + v1 – 2v1v2 + v2 = v1 + 2v2 + (v1 – v2) 0, and v, v = if and only if v1 = 0, v2 = 0, and v1 – v2 = 0 if and only if v = 0.
ut vt u v Example: Consider R4. Let u = x and v = x . Instead of the Euclidean inner product u y v y u z v z c 2 000 0010 uTv, let u, v = uT v, where c is a positive constant (the speed of light), so 0100 1000 2 u, v = –c utvt + uxvx + uyvy + uzvz. Show that u, v is not an inner product. (Optional aside: u, v is called the spacetime interval.)
Solution: 1 2 2 2 2 2 0 4) v, v = –c vt + vx + vy + vz which can be less than zero, for example when v = . 0 0 Note: Axioms 2 and 3 are satisfied because of the rules of matrix multiplication. Axiom 1 is satisfied because the matrix is symmetric (AT = A).
Theorem 5.7 Properties of Inner Products
If u, v, and w are vectors in an inner product space V “over the real numbers” and c is a scalar, then
1) 0, v = v, 0 = 0 2) u + v, w = u, w + v, w 3) u, cv = cu, v
Proof of 5.7.1: 0, v = 0v, v = 0v, v = 0. By the symmetric property, v, 0 = 0, v = 0
Proof of 5.7.2: u + v, w = w, u + v Symmetry
= w, u + w, v Axiom 2 = u, w + v, w Symmetry optional Proof of 5.7.3: u, cv = cv, u Symmetry = cv, u Axiom 3 = cu, v Symmetry
p. 130 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces. Together with the inner product axioms, Thm 5.7 says that the inner product is linear in each of its arguments (we will study linear operators much more in Chapter 6.) Linearity is essentially the same as distributivity: Dot product: (2u + 3v)•(5x + 7y) = 10(u•x) + 14(u•y) + 15(v•x) + 21(v•y) General inner product: 2u + 3v, 5x + 7y = 10u, x + 14u, y + 15v, x + 21v, y
Let u and v be vectors in an inner product space V.
The length, or norm, of u is ||u|| = ,uu.
* The distance between two vectors u and v is d(u, v) = ||u – v||.
, vu The angle between two nonzero vectors u and v is given by = arccos , 0 . vu
Two vectors u and v in Rn are orthogonal (perpendicular) when u, v = 0. (Notice that the 0 vector is orthogonal to every vector.)
v If ||u|| = 1 then u is called a unit vector. If v is any nonzero vector, then u = is the unit vector v , vu in the direction of v. Notice that the definition of angle presumes that –1 1. This is vu guaranteed by the Cauchy-Schwarz Inequality (Theorem 5.8.1).
Properties of Length
1) ||u|| 0 2) ||u|| = 0 if and only if u = 0. 3) ||cu|| = |c| ||u||
These follow from the axioms in the definition of inner product and the definition of norm.
Properties of Distance
1) d(u, v) 0 2) d(u, v) = 0 if and only if u = v. 3) d(u, v) = d(v, u)
These follow from the axioms in the definition of inner product and the definition of distance.
p. 131 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces.
a11 a12 b11 b12 Example: Let A = and B = be two matrices in M2,2 and define the inner a21 a22 b21 b22
product A, B = a11b11 + a12b12 + a21b21 + a22b22. 5 3 1 0 0 1 4 4 Let I = , R = and L = 3 5 . 0 1 1 0 4 4
a) Find the norm of I = .
Solution:
b) Find the inner product of and angle between R and L. Are they orthogonal?
Solution: .(
c) Find the inner product of and angle between I and L. Are they orthogonal?
Solution:
p. 132 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces. Example: Let p and q be two polynomials in P and define the inner product 1 p, q = xqxp)()(dx 1
a) Find the norm of p(x) = 1. (Sometimes we will just say “Find the norm of 1.”)
Solution:
b) Find the inner product of p(x) = 1 and q(x) = x. Are they orthogonal?
Solution: .(
c) Find the angle between 1 and x2. Are they orthogonal?
Solution:
Theorem 5.8
Let u and v be vectors in an inner product space V.
1) Cauchy-Schwarz Inequality: | u, v | ||u|| ||v|| 2) Triangle inequality: ||u + v|| ||u|| + ||v|| 3) Pythagorean Theorem: u and v are orthogonal if and only if ||u + v||2 = ||u||2 + ||v||2.
The proofs are the same as the proofs of Theorems 5.4, 5.5, and 5.6, except that we replace u•v with u, v. A result such as the triangle inequality, which seems intuitive in R2 and R3, can be generalized through linear algebra to some less obvious statements about functions in C[a, b].
p. 133 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces. 4 Example: Consider f (x) = cos(x) and g(x) = (/2 – x)(/2 + x) in C[–/2, /2] with the inner 2 / 2 product f , g = f (x)g(x)dx . Verify the triangle inequality by direct calculation. / 2
Solution: The triangle inequality is || f + g || || f || + || g ||.
|| f ||2 =
|| g ||2 =
|| f + g ||2 =
|| f + g || =
|| f || + || g || =
u Orthogonal Projections: Let u and v be two vectors in R2. If v is nonzero, then u can be orthogonally projected onto v. This projection is denoted by
projv u. Because projv u is a scalar multiple of v, we can write v projv u v projv u = a v v (Recall that is the unit vector in the direction of v.) v u v u v Then a = ||u|| cos = ||u|| = u v v
u v u v projv u = v = v v 2 v v
p. 134 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces.
* Let u and v be vectors in a general inner product space V. The orthogonal projection of u onto v is 〈퐮, 퐯〉 proj 퐮 = 퐯 퐯 〈퐯, 퐯〉
Example: Use the Euclidean inner product find the orthogonal projection of u = (1, 0, 0) onto
v = (1, 1, 1) and of v onto u .
Solution: .( projv u =
projv u =
p. 135 Chapter 5: Inner Product Spaces. 5.2 Inner Product Spaces. Theorem 5.9 Orthogonal Projection and Distance u u Let u and v be two vectors in an inner product d(u, proj u) d(u, cv) space V such that v 0. Then v
, vu projv u v cv v d(u, projv u) < d(u, cv) whenever c , vv
That is to say, the orthogonal projection projv u is the vector parallel to v that is closest to u.
u Proof: Let b = , c , p = cv – bv, q = u – bv, and t = u – cv. q t = u – cv The p is orthogonal to q, because bv p, q = cv – bv, u – bv p = (c – b)v
= cv, u – cbv, v – bv, u + b2v, v , vu 2 = cu, v – c v, v – u, v + v, v , vv 2 2 , vu , vu 2 = cv, u – cv, u – + = 0. , vv , vv So the Pythagorean Theorem tells us ||t||2 = ||p||2 + ||q||2 > ||q||2 because p 0 because c b.
Therefore, d(u, projv u) = ||q|| < ||t|| = d(u, cv).
Example: Consider C[–, ] with the inner product f , g = xgxf )()( dx . Find the projection of f (x) = x onto g(x) = sin(x).
Solution: .( , gf projg f = xg )( . , gg
p. 136 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process.
5.3 Orthogonal Bases: Gram-Schmidt Process. Objective: Show that a set of vectors is orthogonal and forms an orthonormal basis. Objective: Represent a vector relative to an orthonormal basis. Objective: Apply the Gram-Schmidt orthonormalization process.
Although a vector space can have many different bases, some bases may be easier to work with 3 than others. For example, in R , we often use the standard basis S = {e1, e2, e3} = {(1, 0, 0), (0, 1, 0), (0 0, 1)}. Two properties that make the standard basis desirable are that
1) The vectors are normalized, i.e. each basis vector is a unit vector: ||ei|| = 1 for i = 1, 2, 3.
2) The vectors are mutually orthogonal: ei•ej = 0 whenever i j, i.e. e1•e2 = 0, e1•e3 = 0, e2•e3 = 0
A set S of vectors in an inner product space V, is called orthogonal when every pair of vectors in S is orthogonal. If, in addition, each vector in the set is a unit vector, then S is called orthonormal. S* If S = {v1, v2, …, vn} then these definitions can be written as
S is orthogonal if and only if vi•vj = 0 whenever i j.
S is orthonormal if and only if vi•vj = 0 whenever i j, and ||ei|| = 1 for i = 1, 2, …, n.
Example: Show that S = {(cos, sin, 0)}, (–sin, cos, 0), (0, 0, 1)} is an orthonormal set.
Solution: S is orthogonal, because
(cos, sin, 0) • (–sin, cos, 0) = – cos sin + sin cos = 0
(cos, sin, 0) • (0, 0, 1) = 0
and (–sin, cos, 0) • (0, 0, 1) = 0 S*
S is normalized, because
||(cos, sin, 0)|| = 2 0sincos 22 = 1
||(–sin, cos, 0)|| = 2 0cossin 22 = 1
p. 137 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process.
||(0, 0, 1)|| = 100222 = 1
Theorem 5.10 Orthogonal Sets are Linearly Independent
If S = {v1, v2, …, vn} is an orthogonal set of nonzero vectors in an inner product space V, then S is linearly independent.
Proof: suppose c1v1 + c2v2 + …+ cnvn = 0
Then (c1v1 + c2v2 + …+ cnvn), vi = 0, vi for each of the vi
We can distribute the inner product on the left-hand side to obtain
Because S is orthogonal, when j i,
so equation becomes
Because vi 0,
2 Example: Consider C[0, 2] with the inner product f , g = xgxf )()( dx . 0 Show that the set {1, cos(x), sin(x), cos(2x), sin(2x), …, cos(nx), sin(nx)} is a linearly independent orthogonal set.
Solution: We use software to integrate the following expressions, and we use these identities to
k 0)2sin( evaluate the results: whenever k is an integer k 1)2cos(
2 2 nx)cos( 11 1, sin(nx) = )sin(1 dxnx = = = 0 0 n 0 nn
2 nx)sin( 2 1, cos(nx) = )cos(1 dxnx = = 0 – 0 = 0 0 n 0
p. 138 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process. sin(mx), sin(nx) when m n is 2 2 xnm ))sin(( xnm ))sin(( mx )sin()sin(dxnx= = (0 – 0) – (0 – 0) = 0 nm)(2 nm)(2 0 0
cos(mx), cos(nx) when m n is 2 2 xnm ))sin(( xnm ))sin(( mx )cos()cos(dxnx= = (0 + 0) – (0 + 0) = 0 nm)(2 nm)(2 0 0
cos(mx), sin(nx) when m n is 2 2 xnm ))cos(( xnm ))cos(( mx )sin()cos( dxnx = = nm)(2 nm)(2 0 0 1 1 1 1 = 0 )(2 )(2 )(2 nmnmnmnm )(2
2 2 2 nx)(cos 1 1 cos(nx), sin(nx) = )sin()cos( dxnxnx = = = 0 0 2n 0 2 2nn
Corollary Orthogonal Bases S* If V is an inner product space of dimension n, then any orthogonal set of n nonzero vectors in V is a basis for V.
Example: Show that S = {(0, 1, 0, 1), (–1, –1, 2, 1), (2, –1, 0, 1), (2, 2, 3, –2)} is an orthogonal basis of R4.
Solution: S is orthogonal, because (0, 1, 0, 1)•(–1, –1, 2, 1) = 0 – 1 + 0 + 1 = 0 (0, 1, 0, 1)•(2, –1, 0, 1) = 0 –1 + 0 – 1 = 0S* (0, 1, 0, 1)•(2, 2, 3, –2) = 0 + 2 + 0 – 2 = 0 (–1, –1, 2, 1)•(2, –1, 0, 1) = –2 + 1 + 0 + 1 = 0 (–1, –1, 2, 1)•(2, 2, 3, –2) = –2 – 2 + 3 – 2 = 0 and (2, –1, 0, 1)•(2, 2, 3, –2) = 4 – 2+ 0 – 2 = 0 By the Corollary to Theorem 5.10, S is an orthogonal basis for R4.
p. 139 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process. Theorem 5.11 Coordinates Relative to an Orthonormal Basis
B V If = {v1, v2, …, vn} is an orthonormal basis for anS* inner product space , then the coordinate representation of a vector w relative to B is
w = w, v1v1 + w, v2 v2 + … +w, vn vn
Proof: Because B is a basis for V, there are unique scalars c1, c2, …, cn such that
w = c1v1 + c2v2 + …+ cnvn
Taking the inner product with vi of both sides of the equation gives
w, vi = (c1v1 + c2v2 + …+ cnvn), vi
We can distribute the inner product on the left-hand side to obtain
Because S is orthonormal, when
so equation becomes
so
The coordinates of w relative to an orthonormal basis B are called the Fourier coefficients of w relative to B. The corresponding coordinate matrix is
c1 , vw 1 c 2 , vw 2 [w]B = = cn , vw n
1 Note: contrast Thm 5.11 with [w] = vv v ][][][ [w] B 1 S 2 S Sn S from Section 4.7 for a general, non-orthonormal basis B.
p. 140 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process. Example: Find the coordinate matrix of w = (–2, 6, 5) relative to the orthonormal basis 1 1 2 2 22 2 2 1 B = {( , , 0), (– , , ), ( , – , )} of R3. 2 2 6 6 3 3 3 3
Solution: S*
so the coordinate matrix is
Given any basis for a vector space, we can construct an orthonormal basis using a procedure called the Gram-Schmidt Orthonormalization.
Theorem 5.12(Alt) (Alternative) Gram-Schmidt Orthonormalization Process
Let B = {v1, v2, …, vn} be a basis for an inner product space V.
Let B = {u1, u2, …, un}, where ui is given by
w1 w1 = v1, ...... u1 = w1
w 2 w2 = v2 – v2, u1u1, ...... u2 = S* w 2
w 3 w3 = v3 –v3, u1u1 – v3, u2u2,...... u3 = w 3
…
w n wn = vn – vn, u1u1 – vn, u2u2 – … – vn, u n–1un–1, ...... un = w n
Then B is an orthonormal basis for V. Moreover, span{ u1, u2, …, uk} = span{ v1, v2, …, vk} for k = 1, 2, …, n.
Proof: First, B is normalized, since all of the ui are unit vectors: ||ui||= 1. Second, observe that the terms of the form v , u u are projections onto u : v , u u = proj v j i i i j i i ui j
When we subtract from vj, the resultant vector is
orthogonal to ui
p. 141 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process. Now we prove by induction that B is also an orthonormal set.
If n =1, then there are no pairs of vectors in B, so there is nothing to prove.
w k If {u1, u2, …, uk–1} is an orthonormal set, then consider uk = where w k
wk = vk – vk, u1u1 – vk, u2u2 – … – vk, ujuj – … – vk, uk–1uk–1. Let j = 1, 2, …, k – 1. Then
1 uk, uj = (vk – vk, u1u1 – vk, u2u2 – … – vk, ujuj – … vk, uk–1uk–1), uj w n
= (vk, uj – vk, u1u1, uj – vk, u2u2, uj – … – vk, ujuj, uj – …
– vk, uk–1uk–1, uj)
= (vk, uj – vk, u1(0) – vk, u2(0) – … – vk, uj(1) – … – vk, uk–1(0))
= (vk, uj – vk, uj) = 0
So {u1, u2, …, uk} is also an orthonormal set. By mathematical induction, B is orthonormal, and since it has n vectors in an n-dimensional space, B is an orthonormal basis.
Example: Use the Gram-Schmidt orthonormalization process to construct an orthonormal basis from B = {(1, 2, 2), (–1, –1,0), (1, 0, 0)}.
Solution:
)2,2,1( 1 2 w1 = (1, 2, 2) u1 = = ( , , ) )2,2,1( 3 3
w2 = (–1, –1,0) – ((–1, –1,0)•( , , ))( , , ) = (–1, –1,0) – (–1)(1, 2, 2) S* 2 1 2 ),,( = (– , – , ) u = 3 3 3 = (– , – , ) 2 2 1 2 3 3 3 ),,( w3 =
u3 =
p. 142 Chapter 5: Inner Product Spaces. 5.3 Orthogonal Bases: Gram-Schmidt Process. Example: Use the Gram-Schmidt orthonormalization process to construct an orthonormal basis 1 2 from B = {1, x, x } in P2 with the inner product p(x), q(x) = xqxp)()(dx . 1
2 Solution: Let v1 = 1, v2 = x, and v3 = x . w1 = 1 1 u1 = 1 1 1 ||1||2 = 11dx = x = 2 1 1 1 u1 = 2
w2 =
u2 =
p. 143
Chapter 5: Inner Product Spaces. 5.4 Mathematical Models and Least Squares Analysis (Optional).
5.4 Mathematical Models and Least Squares Analysis (Optional). Objective: Find the orthogonal complement of a subspace. Objective: Solve a least squares problem.
In this section, we consider inconsistent systems of linear equations, and we find the “best approximation” to a solution.
Example: Given a table of data. We want to find the t y coefficients c0 and c1 of the line y = c0 + c1t that “best fits” these points. 1 0
Solution: The system of linear equations that we 2 1 have to solve comes from plugging the three points 3 3 (ti, yi) into the equation c0 + c1ti = yi
cc 10 01 11 0 11 0 c c cc 12 or 0 . Let A = , b = , and x = 0 10 21 1 21 1 c1 c1 cc 10 33 31 3 31 3
001 The system is inconsistent: reduced row-echelon form is . (We also knew that there 010 100 would be no solution because the three points in the graph are not collinear.) But what is the “best approximation”?
1 1 Recall that Ax = 1 2cc is always in the column space of A. But since Ax = b has no 0 1 1 3 solution, b is not in the column space of A. We want to find an x that gives the Ax that is closest to b.
Least Squares Problem: Given an mn matrix A and a vector b in Rm, find x in Rn, such that ||Ax – b||2 is minimized.
This gives the Ax that is closest to b. It is customary to minimize ||Ax – b||2 instead of ||Ax – b|| because ||Ax – b|| involves a square root. Intuitively, we see that Ax – b is orthogonal to the column space of A.
p. 145 Chapter 5: Inner Product Spaces. 5.4 Mathematical Models and Least Squares Analysis (Optional). To solve the least squares problem, it helps to use the concept of orthogonal subspaces.
Let S1 and S2 be two subspaces of an n-dimensional vector space V.
S1 and S2 are orthogonal when v1, v2 = 0 for all v1 in S1 and v2 in S2. The orthogonal complement of S1 is the set S1 = {u V: u, v= 0 for all vectors v S1}. S1 is pronounced “S1 perp.”
If every vector x V can be written uniquely as a sum of a vector s1 from S1 and a vector s2
from S2, x = s1 + s2, then V is the direct sum of S1 and S2, and we write V = S1 S2.
Theorem 5.13 Properties of Orthogonal Complements
Let S be a subspace of an n-dimensional vector space V. Then
1) dim(S) + dim(S) = n 2) V = S S 3) (S) = S
1 0 Examples: Consider R3. {0} = R3. (R3) = {0}. Let S = span( ), S = span( ), x 0 y 1 0 0 0 0 S = span( ), S = span( , ), and S = span( , ). Then S is orthogonal to S . z 0 xy 1 yz x y 1 0 Sy is orthogonal to Sz. Sxy is not orthogonal to Syz. Sxy = Sx Sy. (Sz) = Sxy, and (Sxy) = Sz.
Theorem 5.16 Fundamental Subspaces of a Matrix
Let A be an mn matrix. The notation R(A) means the column space of A (R for range). R(AT ) is the row space of A. N(A) is the nullspace of A.
1) R(AT ) and N(A) are orthogonal complements. That is: all vectors in the row space of A are orthogonal to all vectors in the nullspace of A; dim(R(AT )) + dim(N(A)) = n, and R(AT ) N(A) = Rn. 2) R(A) and N(AT ) are orthogonal complements. That is: all vectors in the column space of A are orthogonal to all vectors in the nullspace of AT; dim(R(A)) + dim(N(AT )) = m, and R(A) N(AT ) = Rm.
p. 146 Chapter 5: Inner Product Spaces. 5.4 Mathematical Models and Least Squares Analysis (Optional). Sketch of the proof: 1) N(A) ={x Rn: Ax = 0} shows that the nullspace is orthogonal to the row space, because the null vector x is orthogonal to each row of A. 1 row( of A x 0) 2 row( of A x 0) Ax = 0 is equivalent to row( of Am x 0) dim(R(AT )) + dim(N(A)) = n is rank + nullity = n, which we already know. The basis vectors of R(AT )) are linear independent of the basis of N(A) because they are orthogonal to the basis of N(A), so the union of a basis of R(AT )) and a basis of N(A) is a basis of Rn, so R(AT ) N(A) = Rn.
2) Replace A with AT in Part (1), and use (AT )T = A. When we replace A with AT, we must interchange m and n.
1 0 2 1 Example: Find the orthogonal complement of S = span( , ). 0 0 0 1 01 12 Solution: S is the column space R(A), where A = , so S = N(AT). 00 10 00021 02001 rref To find N(A), solve Ax = 0 01010 01010
so 1x1 + 0x2 + 0s – 2t = 0 and 0x1 + 1x2 + 0s + 1t = 0 so x2 = –t and x1 = 2t 2t 0 2 0 2 t 0 1 0 1 x = = s + t so S = N(AT) = span( , ) s 1 0 1 0 t 0 1 0 1
If S is a subspace of an n-dimensional vector space V and v is a vector in V, then v can be written uniquely as the sum of a vector from S and a vector from S v = s1 + s2, where s1 S and s2 S because V = S S . Then s1 S is the orthogonal projection of v onto S, written projS v
p. 147 Chapter 5: Inner Product Spaces. 5.4 Mathematical Models and Least Squares Analysis (Optional).
Theorem: v – projS v is orthogonal to every vector in S.
Proof: v can be written uniquely as v = + s2, where s2 = v – S . Since v – S, v – is orthogonal to every vector in S.
Theorem 5.15 Orthogonal Projection and Distance
Let S be a subspace of an n-dimensional vector space V and v be a vector in V. Then for all u S, u , ||v – || < ||v – u||.
In other words, is the vector in S that is closest to v.
Proof: Let all u S, u so v – u = (v – ) + ( – u) Now v – is orthogonal to ( – u), so by the Pythagorean Theorem, ||v – u||2 = ||v – ||2 + || – u||2 where || – u|| > 0 because u . Therefore, ||v – u||2 > ||v – ||2 ||v – u|| > ||v – ||
To solve the least squares problem, we need Ax – b (R(A)) = N(AT ), so AT(Ax – b) = 0, or
ATAx = ATb “Normal equations”
Solution to Least Squares Problem
p. 148 Chapter 5: Inner Product Spaces. 5.4 Mathematical Models and Least Squares Analysis (Optional). Example: Let’s finish the example from the t y beginning of this section. Given a table of data. We want to find the coefficients c0 and c1 of the line 1 0 y = c + c t that “best fits” these points. 0 1 2 1 Solution: 3 3
cc1001 11 0 11 0 c c cc 12 or 0 . Let A = , b = , and x = 0 10 21 1 21 1 c1 c1 cc1033 31 3 31 3
1 T T 3 6 c0 4 3 6 5/ 3 So A Ax = A b , i.e. = , so = = . 6 14 c1 11 6 14 3/ 2
3 5 Therefore, the least-squares regression line for the data is y = 2 t – 3
Example: Given the table of data. Find the coefficients 2 t y c0, c1, and c2 of the quadratic y = c0 + c1t + c2t that best fits these points. 0 2
Solution: 1 1.5 2 2.5 1c 0c 02 c 2 1 0 0 2 0 1 2 3 4 2 c0 1c 1c 1 c 1.5 1 1 1 1.5 0 1 2 so c = 1c 2c 22 c 2.5 1 2 4 1 2.5 0 1 2 2 c2 1c0 3c1 3 c2 4 1 3 9 4
2 1.95 1.5 A = , x = , and b = . ATAx = ATb so x = (ATA)–1ATb = 0.8 2.5 0.5 4
y = 1.95 – 0.8t + 0.5t2
p. 149
Chapter 5: Inner Product Spaces. 5.5 Applications of Inner Product Spaces (Optional).
5.5 Applications of Inner Product Spaces (Optional). Objective: Find the nth-order Fourier approximation of a function. Objective: Given a subspace with an orthogonal or orthonormal basis, find the projection of a vector onto that subspace. Objective: Find the cross product of two vectors in R3.
Recall from Thm. 5.15 that projS v is the unique vector in S that is closest to v.
Theorems 5.14 & 5.19 Projection onto a Subspace & Least Squares Approximation
If {u1, u2, …, un} is an orthonormal basis for a subspace S of a vector space V, and v V, then
= v, u1u1 + v, u2u2 + … + v, unun
If V is a space of functions, e.g. V = C[a, b] (so the ui are functions), then
projS f = f, u1u1 + f, u2u2 + … + f, unun
is the least-squares approximating function of f with respect to S.
Proof: we will show that = v, u1u1 + v, u2u2 + … + v, unun by showing that v can be written uniquely as the sum of a vector from S and a vector from S v = s1 + s2, where s1 S and s2 S . Then by definition, = s1.
Let s1 = v, u1u1 + v, u2u2 + … v, unun and s2 = v – v, u1u1 – v, u2u2 – … – v, unun
Then s1 S because it is a linear combination of vectors in S.
For each basis vector ui of S
ui, s2 = ui, (v – v, u1u1 – v, u2u2 – … – v, uiui – … – v, unun) = ui, v – v, u1ui, u1 – v, u2ui, u2 – … – v, uiui, ui – … – v, unui, un = ui, v – 0 – 0 – … – v, ui (1) – … – 0 = 0
Since all vectors in S are linear combinations of the {ui}, s2 is orthogonal to all vectors in S, so s2 S .
Example: we know from 5.3 that set {1, cos(x), sin(x), cos(2x), sin(2x), …, cos(nx), sin(nx)} is a 2 an orthogonal set using the inner product f , g = xgxf )()( dx . Let’s normalize the set. 0
p. 151 Chapter 5: Inner Product Spaces. 5.5 Applications of Inner Product Spaces (Optional).
2 ||1||2 = (1)(1)dx = 2. 0 2 2 2 x sin(nx)cos(nx) For n 0, ||cos(nx)|| = cos(nx)cos(nx)dx = = 0 2 2n x0 2 2 2 x sin(nx)cos(nx) and ||sin(nx)|| = sin(nx)sin(nx)dx = = 0 2 2n x0 So { 1 , 1 cos(x), 1 sin(x), cos(2x), sin(2x), …, cos(nx), sin(nx)} is an 2 orthonormal set.
The nth-order Fourier approximation of a function f on the interval [0, 2] is the projection of f onto span { , cos(x), sin(x), cos(2x), sin(2x), …, cos(nx), sin(nx)}. Fourier approximations are useful in modeling periodic functions such as sound waves, heart rhythms, and electrical signals.
Example: Find the second-order Fourier approximation to the periodic function
f (x) = – x for x [0, 2 ), and f (x + 2) = f (x)
Solution: g(x) = f, + f, cos(x) cos(x) + f, sin(x) sin(x) + f, cos(2x) cos(2x) + f, sin(2x) sin(2x)
Using software to evaluate the integrals, we find 2 f, 1 = 1 1 ( x)dx = 0 2 2 0
2 f, 1 cos(x) = 1 ( x)cos(x)dx = 0 0
2 f, 1 sin(x) = 1 ( x)sin(x)dx = 2 0
2 f, cos(2x) = 1 ( x)cos(2x)dx = 0 0
2 f, sin(2x) = 1 ( x)sin(2x)dx = 1 0 nd So the 2 order Fourier approximation is g2(x) = 2sin(x) + sin(2x)
p. 152 Chapter 5: Inner Product Spaces. 5.5 Applications of Inner Product Spaces (Optional). See http://www.falstad.com/fourier/e-sawtooth.html It is common notation with Fourier series to write the coefficients as 2 2 2 a = 1 xf)(dx a = 1 xf kx)cos()(dx b = 1 xf kx)sin()(dx 0 k k 0 0 0 so that the nth-order Fourier approximation is
a0 gn(x) = + a1cos(x) + b1sin(x) + a2cos(2x) + b2sin(2x) + … + ancos(nx) + bnsin(nx) 2
2 2 22 2 2 1 Example: Given that S = {(– , , ), ( , – , )} is an orthonormal set in R3. Find 6 6 3 3 3 3
the projection of v = (1, 0, 0) onto span{(– , , ), ( , – , )}.
Solution:
projS v = ((1, 0, 0)•(– , , ))(– , , )
+ ((1, 0, 0)• ( , – , ))( , – , )
1 1 = – (– , , ) + ( , – , ) = ( , – , 0) 2 2
3 In R , the cross product of two vectors u = (u1, u2, u3) = u1i + u2j + u3k and v = (v1, v2, v3) = v1i + v2j + v3k is kji
uv = (u2v3 – u3v2)i + (u3v1 – u1v3)j + (u1v2 – u2v1)k = = uuu 321
vvv 321 where the right-hand side is a “determinant” containing the vectors i, j, and k. The cross product is undefined for vectors in vector spaces other than R3.
Theorem 5.17 Algebraic Properties of the Cross Product
If u, v, and w are vectors in R3 and c is a scalar, then
p. 153 Chapter 5: Inner Product Spaces. 5.5 Applications of Inner Product Spaces (Optional).
1) uv = –vu uuu321 2) u(v + w) = uv + uw 6) u•(vw) = (uv)•w = vvv 3) c(uv) = (cu)v = u(cv) 321 www 4) u0 = 0u = 0 321 5) uu = 0
p. 154 Chapter 5: Inner Product Spaces. 5.5 Applications of Inner Product Spaces (Optional). Theorem 5.18 Geometric Properties of the Cross Product
If u, v, and w are nonzero vectors in R3, then
1) uv is orthogonal both u and v 2) The angle between u and v is given by ||uv|| = ||u|| ||v|| sin( ) 3) u and v are parallel if and only if uv = 0 4) The parallelogram having u and v as adjacent sides has an area of ||uv|| 5) The parallelepiped having u, v, and w as edges has a volume of ||u•(vw)||
Example: Find the area of the parallelogram with vertices at (5, 2, 0), (3, –6, 7), (7, –2, 8), and (5, –10, 15)
kji Solution: area = ||uv||. uv = 782 = –36 + 30j + 24k 842
||–36i + 30j + 24k || = ( 36 2 () 30 2 () 24)2 = 6 77 52.6498
Example: Find a vector orthogonal to u = (–4, 3, –4) and v = (4, 7, 1).
kji Solution: uv = 434 = 31i – 12j – 40k 174
p. 155
8.3 Polar Form and De Moivre's Theorem. (Optional)
8.3 Polar Form and De Moivre's Theorem. (Optional) Objective: Determine the polar form of a complex number and convert between the polar form and the standard form of a complex number. Objective: Determine the exponential polar form of a complex number and convert between the exponential polar form and the standard form of a complex number using Euler’s formula. Objective: Multiply and divide complex numbers in polar form and in exponential polar form. Objective: Use DeMoivre’s Theorem to find powers of numbers in polar form and in exponential polar form.
Multiplication, division, and powers of complex number are tedious to calculate in standard form. However, when complex numbers are written in polar form, multiplication, division, and powers become easy to calculate and interpret.
The polar form of a nonzero complex number z = a + bi is given by
z = r (cos + i sin ) where a = r cos, b = r sin, r = ba 22 = |z|, and tan = b/a.
The number r is the modulus of z and is the argument of z.
There are infinitely many choices for the argument, because cosine and sine have a period of 2. Usually, we choose the principal argument, which satisfies < 2.
The principal argument = Arg(z) of a nonzero complex number z = a + bi is given by
tan = b/a and – <
Example: Graph and find the polar form (using the Arg(z) = /2 principal argument) of
axis a) –3 b) –4 – 3i c) 2 – 2 i Arg(z) = tan–1(b/a) + Arg(z) = tan–1(b/a) nary
Solution: Imagi Real axis Arg(z) = Arg(z) = 0
Arg(z) = tan–1(b/a) – Arg(z) = tan–1(b/a)
a) Arg(z) = – /2
c)
b)
p. 157 8.3 Polar Form and De Moivre's Theorem. (Optional) a) –3 = 3(cos + i sin)
b) r = 2 )3()4(2 = 25 = 5
–1 3 Arg(z) = + tan ( 4 ) 3.3785 –4 – 3i 5(cos(3.3785) + i sin(3.3785))
c) r = 2 )2()2(2 = 4 = 2 Arg(z) = tan–1( 2 ) = 2 4 2 – 2 i = 2(cos( ) + i sin( ))
Example: Find the standard form of 10 (cos( 6 ) + i sin( 6 )) 3 1 Solution: 10 (cos( ) + i sin( )) = 10 i = 535i 6 2 2
Theorem. Euler’s Formula
cos + i sin = ei
where e is Euler’s number, e 2.71828
Proof: From calculus, we know that the Maclaurin series (the Taylor series expansion around zero) of a function f is
1 1 1 1 1 1 f (x) = f (0) + f (0) x + f (0) x2 + f (0) x3 + f (4)(0) x4 + f (5)(0) x5 + … !0 !1 !2 !3 !4 !5
Therefore, the Maclaurin series for ex is
1 1 1 1 1 1 1 x 1 2 3 4 5 6 xxxxxxxe 7 ... !1 !2 !3 !4 !5 !6 !7
Substitute x = i. Note that i2 = –1, i3 = –i, i4 = 1, i5 = i, etc. So
1 1 1 1 1 1 1 i 1 ie 2 i 3 4 i 5 6 i 7 ... !1 !2 !3 !4 !5 !6 !7
Therefore, the Maclaurin series for cos and sin are
1 1 1 01)cos( 2 0 4 0 6 0 ... !2 !4 !6
1 1 1 1 0)sin( 0 3 0 5 0 7 ... !1 !3 !5 !7
p. 158 8.3 Polar Form and De Moivre's Theorem. (Optional) Comparing the series for ei, cos, sin, we see that ei = cos + i sin
Example: Graph and find the polar form (using the principal argument) of a) –i b) 1 – 3 i c) 3 + 2i c)
Solution
a) –i = e–i /2 a) b) r = 2 )3(12 = 4 = 2 –1 = tan (– /1) = – /3 b) 1 – i = 2 e–i /3
c) r = 2322 = 13 3.606 = tan–1(2/3) 0.588 3 + 2i = 3.606 e0.588i d) Example: Graph and find the polar form (using the principal argument) of a) –i b) 1 – i c) 3 + 2i
Example: Find the standard form of 3e5i/4
5 5 2 2 23 23 5i/4 Solution: 3e = 3 (cos + i sin ) = 3 i = i 4 4 2 2 2 2
Theorem 8.4 Product and Quotient of Two Complex Nubmers
er i1 er i2 = err i 21 )( irir )sin(cos)sin(cos = rr isin()[cos( )] 1 2 21 11 221 2 2121 21
er i1 (cos ir )sin r 1 r1 i 21 )( 11 1 1 = e = [cos( 21 isin() 21 )] , z2 0 i2 2er r2 2 (cos 2 ir 2 )sin r2
Proof: The exponential formulas follow directly from the laws of exponents. The polar forms follow from the exponential polar forms and Euler’s formula.
Example: Sketch and simplify.
3ei i sin()[cos(3 )] i/3 i/2 a) 2e 3e = 2(cos 3 + i sin 3 )3(cos 2 + i sin 2 ) b) i 4/3 3 3 4e 4 i sin()[cos(4 4 )]
p. 159 8.3 Polar Form and De Moivre's Theorem. (Optional) Solution:
a) 2ei/33ei/2 5 /6 = /2 + /3 i(/3+/2) = (2)(3)e 6 3
= 6ei5/6 /3 2 /2 /3 2(cos 3 + i sin 3 )3(cos 2 + i sin 2 )
5 5 = 6[cos( 6 ) + i sin( 6 )]
/2 = 90, /3 = 60, = 150
3ei 3 3 b) = ei 4/3(()) = ei 4/7 4e i 4/3 4 4
3 3 7 /4 = –(–3 /4) = ei )4/72( = ei 4/ 4 4
isin()[cos(3 )] 7 /4 3 3 3 4 isin()[cos(4 4 )]
3 ¾ – /4 = 2 – 7 /4 7 isin()[cos( 7 )] 4 4 4
3 i 4/sin()4/[cos( )] 4 –3 /4
4
Theorem 8.5 DeMoivre’s Theorem
P (rei )n = r nein
[r (cos + i sin )]n = r n [cos(n ) + i sin(n )]
Proof: The exponential formula follows directly from the laws of exponents. The polar forms follow from the exponential polar forms and Euler’s formula.
p. 160 8.3 Polar Form and De Moivre's Theorem. (Optional) Example: Sketch and simplify.
a) ei2k/3 for k = –1, 0, 1, 2, 3, 4
b) eik/2 for k = –2, –1, 0, 1, 2, 3, 4, 5
Solution
i(–1)/3 1 3 –i/3 a) e = – 2 – 2 i = e
ei(0)/3 = 1 = e0
i(1)/3 3 i/3 e = – + 2 i = e
ei(2)/3 = – – i = e–i/3
ei(3)/3 = 1 = e0
ei(4)/3 = – + i = ei/3
b) ei(–2)/2 = e–i = –1 = ei
ei(–1)/2 = –i = e–i/2
ei(0)/2 = 1 = e0
ei(1)/2 = i = ei/2
ei(2)/2 = –1 = ei
ei(3)/2 = –i = e–i/2
ei(4)/2 = 1 = e0
ei(5)/2 = i = ei/2
p. 161
9781133110873_0803.qxp 3/10/12 6:54 AM Page 404
404 Chapter 8 Complex Vector Spaces
8.3 Polar Form and DeMoivre’s Theorem
Determine the polar form of a complex number, convert between the polar form and standard form of a complex number, and multiply and divide complex numbers in polar form. Use DeMoivre’s Theorem to find powers and roots of complex numbers in polar form.
POLAR FORM OF A COMPLEX NUMBER Imaginary axis At this point you can add, subtract, multiply, and divide complex numbers. However, there is still one basic procedure that is missing from the algebra of complex numbers. (a, b) To see this, consider the problem of finding the square root of a complex number such as i. When you use the four basic operations (addition, subtraction, multiplication, and division), there seems to be no reason to guess that r b 1 i Ίi . θ Ί2 Real a 0 axis That is, 1 i 2 i. Ί Standard Form: a + bi 2 θθ Polar Form: r(cos + i sin ) To work effectively with powers and roots of complex numbers, it is helpful to use a Figure 8.7 polar representation for complex numbers, as shown in Figure 8.7. Specifically, if a bi is a nonzero complex number, then let be the angle from the positive real axis to the radial line passing through the point ͑a, b͒ and let r be the modulus of a bi. This leads to the following. a r cos b r sin r Ίa2 b2 So,a bi ͑r cos ͒ ͑r sin ͒i, from which the polar form of a complex number REMARK is obtained. The polar form of z 0 is expressed as Definition of the Polar Form of a Complex Number z 0͑cos i sin ͒, where is any angle. The polar form of the nonzero complex number z a bi is given by z r͑cos i sin ͒ where a r cos , b r sin , r Ίa2 b2, and tan b͞a. The number r is the modulus of z and is the argument of z.
Because there are infinitely many choices for the argument, the polar form of a complex number is not unique. Normally, the values of that lie between and are used, although on occasion it is convenient to use other values. The value of that satisfies the inequality < Principal argument is called the principal argument and is denoted by Arg(z ). Two nonzero complex numbers in polar form are equal if and only if they have the same modulus and the same principal argument. 9781133110873_0803.qxp 3/10/12 6:54 AM Page 405
8.3 Polar Form and DeMoivre’s Theorem 405
Finding the Polar Form of a Complex Number
Find the polar form of each of the complex numbers. (Use the principal argument.) a.z 1 i b.z 2 3i c. z i SOLUTION a. Because a 1 and b 1, then r 2 12 ͑ 1͒2 2, which implies that r Ί2. From a r cos and b r sin , a 1 Ί2 b 1 Ί2 cos and sin . r Ί2 2 r Ί2 2 So, ͞4 and z Ί2 ΄cos i sin ΅. 4 4 b. Because a 2 and b 3, then r2 22 32 13, which implies that r Ί13. So, a 2 b 3 cos and sin r Ί13 r Ί13 and it follows that Ϸ 0.98. So, the polar form is
z Ϸ Ί13 ΄cos͑0.98͒ i sin͑0.98͒΅.
c. Because a 0 and b 1, it follows that r 1 and ͞2, so z 1cos i sin . 2 2 The polar forms derived in parts (a), (b), and (c) are depicted graphically in Figure 8.8.
Imaginary Imaginary axis axis
2 4 z = 2 + 3i 1 3 Real 2 θ axis −2 −1 2 − 1 1 z = 1 − i θ Real −2 axis 12 a.z Ί2 ΄cos i sin ΅ b. z Ϸ Ί13 ͓cos͑0.98͒ i sin͑0.98͔͒ 4 4
Imaginary axis
z = i 1
θ Real axis 1