<<

AMS526: Numerical Analysis I (Numerical ) Lecture 2: Orthogonal Vectors and Matrices; Vector Norms

Xiangmin Jiao

SUNY Stony Brook

Xiangmin Jiao Numerical Analysis I 1 / 11 Outline

1 Orthogonal Vectors and Matrices

2 Vector Norms

Xiangmin Jiao Numerical Analysis I 2 / 11 Euclidean length√ of u is the square root of the inner of u with itself, i.e., u∗u

Inner product of two unit vectors u and v is the cosine of the angle α u∗v between u and v, i.e., cos α = kukkv k Inner product is bilinear, in the sense that it is linear in each vertex separately:

∗ ∗ ∗ (u1 + u2) v = u1v + u2v ∗ ∗ ∗ u (v 1 + v 2) = u v 1 + u v 2 (αu)∗(βv) =αβ ¯ u∗v

Inner Product m Inner product () of two column vectors u, v ∈ C is ∗ Pm u v = i=1 u¯i vi In contrast, outer product of u and v is uv ∗ Note that is different

Xiangmin Jiao Numerical Analysis I 3 / 11 Inner product of two unit vectors u and v is the cosine of the angle α u∗v between u and v, i.e., cos α = kukkv k Inner product is bilinear, in the sense that it is linear in each vertex separately:

∗ ∗ ∗ (u1 + u2) v = u1v + u2v ∗ ∗ ∗ u (v 1 + v 2) = u v 1 + u v 2 (αu)∗(βv) =αβ ¯ u∗v

Inner Product m Inner product (dot product) of two column vectors u, v ∈ C is ∗ Pm u v = i=1 u¯i vi In contrast, outer product of u and v is uv ∗ Note that cross product is different

Euclidean length√ of u is the square root of the inner product of u with itself, i.e., u∗u

Xiangmin Jiao Numerical Analysis I 3 / 11 Inner product is bilinear, in the sense that it is linear in each vertex separately:

∗ ∗ ∗ (u1 + u2) v = u1v + u2v ∗ ∗ ∗ u (v 1 + v 2) = u v 1 + u v 2 (αu)∗(βv) =αβ ¯ u∗v

Inner Product m Inner product (dot product) of two column vectors u, v ∈ C is ∗ Pm u v = i=1 u¯i vi In contrast, outer product of u and v is uv ∗ Note that cross product is different

Euclidean length√ of u is the square root of the inner product of u with itself, i.e., u∗u

Inner product of two unit vectors u and v is the cosine of the angle α u∗v between u and v, i.e., cos α = kukkv k

Xiangmin Jiao Numerical Analysis I 3 / 11 Inner Product m Inner product (dot product) of two column vectors u, v ∈ C is ∗ Pm u v = i=1 u¯i vi In contrast, outer product of u and v is uv ∗ Note that cross product is different

Euclidean length√ of u is the square root of the inner product of u with itself, i.e., u∗u

Inner product of two unit vectors u and v is the cosine of the angle α u∗v between u and v, i.e., cos α = kukkv k Inner product is bilinear, in the sense that it is linear in each vertex separately:

∗ ∗ ∗ (u1 + u2) v = u1v + u2v ∗ ∗ ∗ u (v 1 + v 2) = u v 1 + u v 2 (αu)∗(βv) =αβ ¯ u∗v

Xiangmin Jiao Numerical Analysis I 3 / 11 Definition Two sets of vectors X and Y are orthogonal if every x ∈ X is orthogonal to every y ∈ Y .

Definition A set of nonzero vectors S is orthogonal if they are pairwise orthogonal. They are orthonormal if it is orthogonal and in addition each vector has unit Euclidean length.

Orthogonal Vectors

Definition A pair of vectors are orthogonal if x∗y = 0.

In other words, the angle between them is 90 degrees

Xiangmin Jiao Numerical Analysis I 4 / 11 Definition A set of nonzero vectors S is orthogonal if they are pairwise orthogonal. They are orthonormal if it is orthogonal and in addition each vector has unit Euclidean length.

Orthogonal Vectors

Definition A pair of vectors are orthogonal if x∗y = 0.

In other words, the angle between them is 90 degrees Definition Two sets of vectors X and Y are orthogonal if every x ∈ X is orthogonal to every y ∈ Y .

Xiangmin Jiao Numerical Analysis I 4 / 11 Orthogonal Vectors

Definition A pair of vectors are orthogonal if x∗y = 0.

In other words, the angle between them is 90 degrees Definition Two sets of vectors X and Y are orthogonal if every x ∈ X is orthogonal to every y ∈ Y .

Definition A set of nonzero vectors S is orthogonal if they are pairwise orthogonal. They are orthonormal if it is orthogonal and in addition each vector has unit Euclidean length.

Xiangmin Jiao Numerical Analysis I 4 / 11 Proof. Prove by contradiction. If a vector can be expressed as of the other vectors in the set, then it is orthogonal to itself.

Question: If the column vectors of an m × n A are orthogonal, what is the of A? Answer: n = min{m, n}. In other words, A has full rank.

Orthogonal Vectors

Theorem The vectors in an orthogonal set S are linearly independent.

Xiangmin Jiao Numerical Analysis I 5 / 11 Answer: n = min{m, n}. In other words, A has full rank.

Orthogonal Vectors

Theorem The vectors in an orthogonal set S are linearly independent.

Proof. Prove by contradiction. If a vector can be expressed as linear combination of the other vectors in the set, then it is orthogonal to itself.

Question: If the column vectors of an m × n matrix A are orthogonal, what is the rank of A?

Xiangmin Jiao Numerical Analysis I 5 / 11 Orthogonal Vectors

Theorem The vectors in an orthogonal set S are linearly independent.

Proof. Prove by contradiction. If a vector can be expressed as linear combination of the other vectors in the set, then it is orthogonal to itself.

Question: If the column vectors of an m × n matrix A are orthogonal, what is the rank of A? Answer: n = min{m, n}. In other words, A has full rank.

Xiangmin Jiao Numerical Analysis I 5 / 11 Pm ∗ Another way to express the condition is v = i=1(qi qi )v ∗ qi qi is an orthogonal projection matrix. Note that it is NOT an orthogonal matrix.

More generally, given an orthonormal set {q1, q2,..., qn} with n ≤ m, we have

n n X ∗ X ∗ ∗ v = + (qi v)qi = r + (qi qi )v and r qi = 0, 1 ≤ i ≤ n i=1 i=1

Let Q be composed of column vectors {q1, q2,..., qn}. ∗ Pn ∗ QQ = i=1(qi qi ) is an orthogonal projection matrix.

Components of Vector

m Given an orthonormal set {q1, q2,..., qm} forming a of C , vector v can be decomposed into orthogonal components as Pm ∗ v = i=1(qi v)qi

Xiangmin Jiao Numerical Analysis I 6 / 11 More generally, given an orthonormal set {q1, q2,..., qn} with n ≤ m, we have

n n X ∗ X ∗ ∗ v = r + (qi v)qi = r + (qi qi )v and r qi = 0, 1 ≤ i ≤ n i=1 i=1

Let Q be composed of column vectors {q1, q2,..., qn}. ∗ Pn ∗ QQ = i=1(qi qi ) is an orthogonal projection matrix.

Components of Vector

m Given an orthonormal set {q1, q2,..., qm} forming a basis of C , vector v can be decomposed into orthogonal components as Pm ∗ v = i=1(qi v)qi

Pm ∗ Another way to express the condition is v = i=1(qi qi )v ∗ qi qi is an orthogonal projection matrix. Note that it is NOT an orthogonal matrix.

Xiangmin Jiao Numerical Analysis I 6 / 11 Components of Vector

m Given an orthonormal set {q1, q2,..., qm} forming a basis of C , vector v can be decomposed into orthogonal components as Pm ∗ v = i=1(qi v)qi

Pm ∗ Another way to express the condition is v = i=1(qi qi )v ∗ qi qi is an orthogonal projection matrix. Note that it is NOT an orthogonal matrix.

More generally, given an orthonormal set {q1, q2,..., qn} with n ≤ m, we have

n n X ∗ X ∗ ∗ v = r + (qi v)qi = r + (qi qi )v and r qi = 0, 1 ≤ i ≤ n i=1 i=1

Let Q be composed of column vectors {q1, q2,..., qn}. ∗ Pn ∗ QQ = i=1(qi qi ) is an orthogonal projection matrix.

Xiangmin Jiao Numerical Analysis I 6 / 11 Question: What is the geometric meaning of multiplication by a unitary matrix? Answer: It preserves angles and Euclidean length. In the real case, multiplication by an orthogonal matrix Q is a rotation (if det(Q) = 1) or reflection (if det(Q) = −1).

Unitary Matrices

Definition A matrix is unitary if Q∗ = Q−1, i.e., if Q∗Q = QQ∗ = I .

In the real case, we say the matrix is orthogonal. Its column vectors are orthonormal. ∗ In other words, qi qj = δij , the .

Xiangmin Jiao Numerical Analysis I 7 / 11 Answer: It preserves angles and Euclidean length. In the real case, multiplication by an orthogonal matrix Q is a rotation (if det(Q) = 1) or reflection (if det(Q) = −1).

Unitary Matrices

Definition A matrix is unitary if Q∗ = Q−1, i.e., if Q∗Q = QQ∗ = I .

In the real case, we say the matrix is orthogonal. Its column vectors are orthonormal. ∗ In other words, qi qj = δij , the Kronecker delta. Question: What is the geometric meaning of multiplication by a unitary matrix?

Xiangmin Jiao Numerical Analysis I 7 / 11 Unitary Matrices

Definition A matrix is unitary if Q∗ = Q−1, i.e., if Q∗Q = QQ∗ = I .

In the real case, we say the matrix is orthogonal. Its column vectors are orthonormal. ∗ In other words, qi qj = δij , the Kronecker delta. Question: What is the geometric meaning of multiplication by a unitary matrix? Answer: It preserves angles and Euclidean length. In the real case, multiplication by an orthogonal matrix Q is a rotation (if det(Q) = 1) or reflection (if det(Q) = −1).

Xiangmin Jiao Numerical Analysis I 7 / 11 Outline

1 Orthogonal Vectors and Matrices

2 Vector Norms

Xiangmin Jiao Numerical Analysis I 8 / 11 Definition m A is a function k · k : C → R that assigns a real-valued length to each vector. It must satisfy the following conditions:

(1)kxk ≥ 0, and kxk = 0 only if x = 0, (2)kx + yk ≤ kxk + kyk, (3)kαxk = |α|kxk.

pPm 2 An example is Euclidean length (i.e, kxk = i=1 |xi | )

Definition of Norms

Norm captures “size” of vector or “distance” between vectors There are many different measures for “sizes” but a norm must satisfy some requirements:

Xiangmin Jiao Numerical Analysis I 9 / 11 pPm 2 An example is Euclidean length (i.e, kxk = i=1 |xi | )

Definition of Norms

Norm captures “size” of vector or “distance” between vectors There are many different measures for “sizes” but a norm must satisfy some requirements:

Definition m A norm is a function k · k : C → R that assigns a real-valued length to each vector. It must satisfy the following conditions:

(1)kxk ≥ 0, and kxk = 0 only if x = 0, (2)kx + yk ≤ kxk + kyk, (3)kαxk = |α|kxk.

Xiangmin Jiao Numerical Analysis I 9 / 11 Definition of Norms

Norm captures “size” of vector or “distance” between vectors There are many different measures for “sizes” but a norm must satisfy some requirements:

Definition m A norm is a function k · k : C → R that assigns a real-valued length to each vector. It must satisfy the following conditions:

(1)kxk ≥ 0, and kxk = 0 only if x = 0, (2)kx + yk ≤ kxk + kyk, (3)kαxk = |α|kxk.

pPm 2 An example is Euclidean length (i.e, kxk = i=1 |xi | )

Xiangmin Jiao Numerical Analysis I 9 / 11 Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Pm 1-norm: kxk1 = i=1 |xi |

∞-norm: kxk∞ . What is its value?

I Answer: kxk∞ = max1≤i≤m |xi | Why we require p ≥ 1? What happens if 0 ≤ p < 1?

p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Xiangmin Jiao Numerical Analysis I 10 / 11 Pm 1-norm: kxk1 = i=1 |xi |

∞-norm: kxk∞ . What is its value?

I Answer: kxk∞ = max1≤i≤m |xi | Why we require p ≥ 1? What happens if 0 ≤ p < 1?

p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Xiangmin Jiao Numerical Analysis I 10 / 11 ∞-norm: kxk∞ . What is its value?

I Answer: kxk∞ = max1≤i≤m |xi | Why we require p ≥ 1? What happens if 0 ≤ p < 1?

p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Pm 1-norm: kxk1 = i=1 |xi |

Xiangmin Jiao Numerical Analysis I 10 / 11 I Answer: kxk∞ = max1≤i≤m |xi | Why we require p ≥ 1? What happens if 0 ≤ p < 1?

p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Pm 1-norm: kxk1 = i=1 |xi |

∞-norm: kxk∞ . What is its value?

Xiangmin Jiao Numerical Analysis I 10 / 11 Why we require p ≥ 1? What happens if 0 ≤ p < 1?

p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Pm 1-norm: kxk1 = i=1 |xi |

∞-norm: kxk∞ . What is its value?

I Answer: kxk∞ = max1≤i≤m |xi |

Xiangmin Jiao Numerical Analysis I 10 / 11 p-norms

Euclidean length is a special case of p-norms, defined as

m !1/p X p kxkp = |xi | i=1 for 1 ≤ p ≤ ∞

Euclidean norm is 2-norm kxk2 (i.e., p = 2)

Pm 1-norm: kxk1 = i=1 |xi |

∞-norm: kxk∞ . What is its value?

I Answer: kxk∞ = max1≤i≤m |xi | Why we require p ≥ 1? What happens if 0 ≤ p < 1?

Xiangmin Jiao Numerical Analysis I 10 / 11 Can we further generalize it to allow W being arbitrary matrix? No. But we can allow W to be arbitrary nonsingular matrix.

Weighted p-norms

A generalization of p-norm is weighted p-norm, which assigns different weights (priorities) to different components.

I It is anisotropic instead of isotropic

Algebraically, kxkW = kW xk, where W is diagonal matrix with ith diagonal entry wi 6= 0 being weight for ith component In other words, m !1/p X p kxkW = |wi xi | i=1

What happens if we allow wi = 0?

Xiangmin Jiao Numerical Analysis I 11 / 11 Weighted p-norms

A generalization of p-norm is weighted p-norm, which assigns different weights (priorities) to different components.

I It is anisotropic instead of isotropic

Algebraically, kxkW = kW xk, where W is diagonal matrix with ith diagonal entry wi 6= 0 being weight for ith component In other words, m !1/p X p kxkW = |wi xi | i=1

What happens if we allow wi = 0?

Can we further generalize it to allow W being arbitrary matrix? No. But we can allow W to be arbitrary nonsingular matrix.

Xiangmin Jiao Numerical Analysis I 11 / 11