Rotation Matrices in Two, Three and Many Dimensions

Total Page:16

File Type:pdf, Size:1020Kb

Rotation Matrices in Two, Three and Many Dimensions Physics 116A Fall 2019 Rotation Matrices in two, three and many dimensions 1. Proper and improper rotation matrices in n dimensions A matrix is a representation of a linear transformation, which can be viewed as a machine that consumes a vector and spits out another vector. A rotation is a trans- formation with the property that the vector consumed by the machine and the vector spit out by the machine have the same length. That is, physically rotating a vector by an angle θ leaves the length of the vector unchanged. As a matrix equation, if R is a rotation matrix and ~v is a vector, then w~ = R~v , where w~ = ~v . (1) k k k k In eq. (1), the length of the vector ~v is denoted by ~v . Note that if ~v is a vector in k k an n-dimensional space, then in terms of the components of ~v = (v1, v2,...,vn) and w~ =(w1,w2,...,wn), eq. (1) is equivalent to n wi = Rijvj , (2) j=1 X where Rij are the matrix elements of the rotation matrix R. In addition, the length of the vector ~v is given by the n-dimensional generalization of the Pythagorean theorem, 2 2 2 ~v = v1 + v2 + ... + v , (3) k k n q with a similar formula for the length of w~ . Let us now ask the following question: what is the most general form of the matrix R such that the lengths of the vectors ~v and w~ are equal, as specified in eq. (1)? That is, what is the most general form of R in eq. (2) such that n n wiwi = vivi , (4) i=1 i=1 X X which is the condition that must be satisfied if ~v = w~ in light of eq. (3). To answer this question, we insert eq. (2) into the leftk handk sidek k of eq. (3). Here, we must be careful with indices and write: n n wiwi = Rijvj Rikvk . (5) j=1 k=1 X X The second time we write wi as a sum on the right hand side of eq. (5), one must use a different summation index (e.g., k), since the index j has already been used as a 1 summation index in the first sum. The best way to see this is to take a simple example, say with i = 1 and n = 2. Then eq. (5) yields 2 w1 =(R11v1 + R12v2)(R11v1 + R12v2) . (6) It would be wrong to write n wiwi = RijvjRijvj , WRONG!, j=1 X 2 2 2 2 2 which in our simple example with i = 1 and n = 2 would yield w1 = R11v1 + R12v2, thereby missing the cross terms that appear when multiplying out the two factors on the right hand side of eq. (6). Since the sums in eq. (5) consist of a finite number of terms, one can move the summation signs to the left and reorder the terms being multiplied to obtain, n n wiwi = RijRikvjvk . (7) j=1 k=1 X X Next, we sum over the free index i on both sides of eq. (7) and use eq. (4) to obtain, n n n n n wiwi = RijRikvjvk = vivi . (8) i=1 i=1 j=1 k=1 i=1 X X X X X One can now use the following trick. Recalling the definition of the Kronecker delta symbol, 1 , if j = k, δjk = (9) 0 , if j = k, ( 6 it follows that the right hand side of eq. (8) can be rewritten as n n n vivi = δjkvjvk . (10) i=1 j=1 k=1 X X X As this is one of the crucial steps in the analysis, you should convince yourself that eq. (10) is an identity (e.g., by examining the case of n = 2 and writing out the sums explicitly). One can now insert eq. (10) back into eq. (8) to obtain, n n n n n RijRikvjvk = δjkvjvk . (11) i=1 j=1 k=1 j=1 k=1 X X X X X Interchanging the order of summation, n n n n n RijRikvjvk = δjkvjvk . (12) j=1 k=1 i=1 j=1 k=1 X X X X X 2 Moving the right hand side above over to the left hand side of eq. (12) yields, n n n RijRik δjk vjvk =0 . (13) − j=1 k=1 " i=1 ! # X X X Let us denote the expression inside the square brackets as, n Ajk RijRik δjk , (14) ≡ i=1 ! − X in which case, eq. (13) is equivalent to n n Ajkvjvk =0 . (15) j=1 k=1 X X In the above analysis, I never specified a specific value for the vj. In other words, if one were to rotate any vector ~v, the end result would still be eq. (15). This observation implies that eq. (15) must be true for any choice of the vj. Consequently, it must be 1 true that Ajk = 0 for all j, k =1, 2,...,n. Inserting Ajk = 0 into the left hand side of eq. (14) yields out final result, n RijRik = δjk . (16) i=1 X To interpret eq. (16), I remind you that the jk element of the product of two n n × matrices is given by the formula [cf. eq. (9.3) in Chapter 3, p. 138 of Boas], n (BC)jk = BjiCik . (17) i=1 X Inserting B = RT and C = R in eq. (17) and using eq. (16) then yields the matrix equation RTR = I. In obtaining this result, we noted that the matrix elements of the identity matrix are Iij = δij, and the transpose of a matrix interchanges the rows and T columns, so that (R )ji = Rij. In conclusion, eq. (16) in matrix form is equivalent to the statement that RTR = I . (18) A matrix R that satisfies eq. (18) is called an orthogonal matrix. Multiplying both sides of eq. (18) on the right by R−1 and using RR−1 = I yields, RT = R−1 . (19) 1 For example, choose ~v = (1, 0, 0,..., 0) and insert this choice in eq. (15) to obtain A11 = 0. Now keep on choosing vectors with different components, and eventually you will see that Ajk = 0 for all j, k =1, 2,...,n. 3 which can also be used as the definition of an orthogonal matrix. Finally, multiplying both sides of eq. (19) on the left by R and again using RR−1 = I yields, RRT = I . (20) Hence, orthogonal matrices are matrices that satisfy R−1 = RT, or equivalently satisfy, RTR = RRT = I . (21) We conclude that rotation matrices must be real orthogonal matrices. Next, consider the following question. Can any real n n orthogonal matrix be identified as some rotation matrix in n-dimensional space× [which is the converse of statement following eq. (21)]? To gain some insight, let us take the determinant of both sides of eq. (18). Recall that det I = 1 and that for any n n matrices B and C, we have det(BC) = (det B)(det C). Finally, for any matrix B,× we have det(BT) = det B. Making use of these properties, eq. (18) yields, det(RTR) = (detRT)(detR) = (det R)2 = det I =1 . (22) It then follows that2 det R = 1 . (23) ± That is, some real orthogonal matrices have determinant equal to 1, whereas other real orthogonal matrices have determinant equal to 1. In these notes, I shall explore the explicit forms for rotation matrices in 2 and 3 dimensions.− Extending those results to n dimensions, one can show that any real n n orthogonal matrix with determinant equal to 1 can be identified as some rotation matrix× in n dimensions. A real orthogonal matrix with determinant equal to 1 can be identified as a product of a rotation and − a reflection. The above observation leads to the following nomenclature. A real orthogonal matrix R with det R = 1 provides a matrix representation of a proper rotation. The most general rotation matrix R represents a counterclockwise rotation by an angle θ about a fixed axis that is parallel to the unit vector nˆ.3 The rotation matrix operates on vectors to produce rotated vectors, while the coordinate axes are held fixed. In typical parlance, a rotation refers to a proper rotation. Thus, in the following sections of these notes we will often omit the adjective proper when referring to a proper rotation. A real orthogonal matrix R with det R = 1 provides a matrix representation of an improper rotation.4 To perform an improper− rotation requires mirrors. That is, the most general improper rotation matrix is a product of a proper rotation by an angle θ about some axis nˆ and a mirror reflection through a plane that passes through the origin and is perpendicular to nˆ. 2Note that eq. (23) implies that detR = 0. This result implies that the inverse of a rotation matrix, R−1 always exist. We implicitly used this6 fact to go from eq. (18) to eq. (19), so this step is now justified. 3The corresponding inverse matrix R−1 represents a clockwise rotation by an angle θ about a fixed axis that is parallel to the unit vector nˆ. With this observation, you can quickly check that RR−1 = I, as expected. 4To distinguish matrices representing improper rotations from those of proper rotations, we shall employ the notation R to denote an improper rotation matrix. 4 Finally, it is instructive to consider the following question: how many real indepen- dent parameters (which we shall denote by N below) describe a real orthogonal n n × matrix? To answer this question, we start with eq.
Recommended publications
  • Quaternions and Cli Ord Geometric Algebras
    Quaternions and Cliord Geometric Algebras Robert Benjamin Easter First Draft Edition (v1) (c) copyright 2015, Robert Benjamin Easter, all rights reserved. Preface As a rst rough draft that has been put together very quickly, this book is likely to contain errata and disorganization. The references list and inline citations are very incompete, so the reader should search around for more references. I do not claim to be the inventor of any of the mathematics found here. However, some parts of this book may be considered new in some sense and were in small parts my own original research. Much of the contents was originally written by me as contributions to a web encyclopedia project just for fun, but for various reasons was inappropriate in an encyclopedic volume. I did not originally intend to write this book. This is not a dissertation, nor did its development receive any funding or proper peer review. I oer this free book to the public, such as it is, in the hope it could be helpful to an interested reader. June 19, 2015 - Robert B. Easter. (v1) [email protected] 3 Table of contents Preface . 3 List of gures . 9 1 Quaternion Algebra . 11 1.1 The Quaternion Formula . 11 1.2 The Scalar and Vector Parts . 15 1.3 The Quaternion Product . 16 1.4 The Dot Product . 16 1.5 The Cross Product . 17 1.6 Conjugates . 18 1.7 Tensor or Magnitude . 20 1.8 Versors . 20 1.9 Biradials . 22 1.10 Quaternion Identities . 23 1.11 The Biradial b/a .
    [Show full text]
  • Lecture Notes: Qubit Representations and Rotations
    Phys 711 Topics in Particles & Fields | Spring 2013 | Lecture 1 | v0.3 Lecture notes: Qubit representations and rotations Jeffrey Yepez Department of Physics and Astronomy University of Hawai`i at Manoa Watanabe Hall, 2505 Correa Road Honolulu, Hawai`i 96822 E-mail: [email protected] www.phys.hawaii.edu/∼yepez (Dated: January 9, 2013) Contents mathematical object (an abstraction of a two-state quan- tum object) with a \one" state and a \zero" state: I. What is a qubit? 1 1 0 II. Time-dependent qubits states 2 jqi = αj0i + βj1i = α + β ; (1) 0 1 III. Qubit representations 2 A. Hilbert space representation 2 where α and β are complex numbers. These complex B. SU(2) and O(3) representations 2 numbers are called amplitudes. The basis states are or- IV. Rotation by similarity transformation 3 thonormal V. Rotation transformation in exponential form 5 h0j0i = h1j1i = 1 (2a) VI. Composition of qubit rotations 7 h0j1i = h1j0i = 0: (2b) A. Special case of equal angles 7 In general, the qubit jqi in (1) is said to be in a superpo- VII. Example composite rotation 7 sition state of the two logical basis states j0i and j1i. If References 9 α and β are complex, it would seem that a qubit should have four free real-valued parameters (two magnitudes and two phases): I. WHAT IS A QUBIT? iθ0 α φ0 e jqi = = iθ1 : (3) Let us begin by introducing some notation: β φ1 e 1 state (called \minus" on the Bloch sphere) Yet, for a qubit to contain only one classical bit of infor- 0 mation, the qubit need only be unimodular (normalized j1i = the alternate symbol is |−i 1 to unity) α∗α + β∗β = 1: (4) 0 state (called \plus" on the Bloch sphere) 1 Hence it lives on the complex unit circle, depicted on the j0i = the alternate symbol is j+i: 0 top of Figure 1.
    [Show full text]
  • Vectors, Matrices and Coordinate Transformations
    S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between them, physical laws can often be written in a simple form. Since we will making extensive use of vectors in Dynamics, we will summarize some of their important properties. Vectors For our purposes we will think of a vector as a mathematical representation of a physical entity which has both magnitude and direction in a 3D space. Examples of physical vectors are forces, moments, and velocities. Geometrically, a vector can be represented as arrows. The length of the arrow represents its magnitude. Unless indicated otherwise, we shall assume that parallel translation does not change a vector, and we shall call the vectors satisfying this property, free vectors. Thus, two vectors are equal if and only if they are parallel, point in the same direction, and have equal length. Vectors are usually typed in boldface and scalar quantities appear in lightface italic type, e.g. the vector quantity A has magnitude, or modulus, A = |A|. In handwritten text, vectors are often expressed using the −→ arrow, or underbar notation, e.g. A , A. Vector Algebra Here, we introduce a few useful operations which are defined for free vectors. Multiplication by a scalar If we multiply a vector A by a scalar α, the result is a vector B = αA, which has magnitude B = |α|A. The vector B, is parallel to A and points in the same direction if α > 0.
    [Show full text]
  • Theory of Angular Momentum and Spin
    Chapter 5 Theory of Angular Momentum and Spin Rotational symmetry transformations, the group SO(3) of the associated rotation matrices and the 1 corresponding transformation matrices of spin{ 2 states forming the group SU(2) occupy a very important position in physics. The reason is that these transformations and groups are closely tied to the properties of elementary particles, the building blocks of matter, but also to the properties of composite systems. Examples of the latter with particularly simple transformation properties are closed shell atoms, e.g., helium, neon, argon, the magic number nuclei like carbon, or the proton and the neutron made up of three quarks, all composite systems which appear spherical as far as their charge distribution is concerned. In this section we want to investigate how elementary and composite systems are described. To develop a systematic description of rotational properties of composite quantum systems the consideration of rotational transformations is the best starting point. As an illustration we will consider first rotational transformations acting on vectors ~r in 3-dimensional space, i.e., ~r R3, 2 we will then consider transformations of wavefunctions (~r) of single particles in R3, and finally N transformations of products of wavefunctions like j(~rj) which represent a system of N (spin- Qj=1 zero) particles in R3. We will also review below the well-known fact that spin states under rotations behave essentially identical to angular momentum states, i.e., we will find that the algebraic properties of operators governing spatial and spin rotation are identical and that the results derived for products of angular momentum states can be applied to products of spin states or a combination of angular momentum and spin states.
    [Show full text]
  • ROTATION THEORY These Notes Present a Brief Survey of Some
    ROTATION THEORY MICHALMISIUREWICZ Abstract. Rotation Theory has its roots in the theory of rotation numbers for circle homeomorphisms, developed by Poincar´e. It is particularly useful for the study and classification of periodic orbits of dynamical systems. It deals with ergodic averages and their limits, not only for almost all points, like in Ergodic Theory, but for all points. We present the general ideas of Rotation Theory and its applications to some classes of dynamical systems, like continuous circle maps homotopic to the identity, torus homeomorphisms homotopic to the identity, subshifts of finite type and continuous interval maps. These notes present a brief survey of some aspects of Rotation Theory. Deeper treatment of the subject would require writing a book. Thus, in particular: • Not all aspects of Rotation Theory are described here. A large part of it, very important and deserving a separate book, deals with homeomorphisms of an annulus, homotopic to the identity. Including this subject in the present notes would make them at least twice as long, so it is ignored, except for the bibliography. • What is called “proofs” are usually only short ideas of the proofs. The reader is advised either to try to fill in the details himself/herself or to find the full proofs in the literature. More complicated proofs are omitted entirely. • The stress is on the theory, not on the history. Therefore, references are not cited in the text. Instead, at the end of these notes there are lists of references dealing with the problems treated in various sections (or not treated at all).
    [Show full text]
  • The Devil of Rotations Is Afoot! (James Watt in 1781)
    The Devil of Rotations is Afoot! (James Watt in 1781) Leo Dorst Informatics Institute, University of Amsterdam XVII summer school, Santander, 2016 0 1 The ratio of vectors is an operator in 2D Given a and b, find a vector x that is to c what b is to a? So, solve x from: x : c = b : a: The answer is, by geometric product: x = (b=a) c kbk = cos(φ) − I sin(φ) c kak = ρ e−Iφ c; an operator on c! Here I is the unit-2-blade of the plane `from a to b' (so I2 = −1), ρ is the ratio of their norms, and φ is the angle between them. (Actually, it is better to think of Iφ as the angle.) Result not fully dependent on a and b, so better parametrize by ρ and Iφ. GAViewer: a = e1, label(a), b = e1+e2, label(b), c = -e1+2 e2, dynamicfx = (b/a) c,g 1 2 Another idea: rotation as multiple reflection Reflection in an origin plane with unit normal a x 7! x − 2(x · a) a=kak2 (classic LA): Now consider the dot product as the symmetric part of a more fundamental geometric product: 1 x · a = 2(x a + a x): Then rewrite (with linearity, associativity): x 7! x − (x a + a x) a=kak2 (GA product) = −a x a−1 with the geometric inverse of a vector: −1 2 FIG(7,1) a = a=kak . 2 3 Orthogonal Transformations as Products of Unit Vectors A reflection in two successive origin planes a and b: x 7! −b (−a x a−1) b−1 = (b a) x (b a)−1 So a rotation is represented by the geometric product of two vectors b a, also an element of the algebra.
    [Show full text]
  • Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
    Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation
    [Show full text]
  • 5 Spinor Calculus
    5 Spinor Calculus 5.1 From triads and Euler angles to spinors. A heuristic introduction. As mentioned already in Section 3.4.3, it is an obvious idea to enrich the Pauli algebra formalism by introducing the complex vector space V(2; C) on which the matrices operate. The two-component complex vectors are traditionally called spinors28. We wish to show that they give rise to a wide range of applications. In fact we shall introduce the spinor concept as a natural answer to a problem that arises in the context of rotational motion. In Section 3 we have considered rotations as operations performed on a vector space. Whereas this approach enabled us to give a group-theoretical definition of the magnetic field, a vector is not an appropriate construct to account for the rotation of an orientable object. The simplest mathematical model suitable for this purpose is a Cartesian (orthogonal) three-frame, briefly, a triad. The problem is to consider two triads with coinciding origins, and the rotation of the object frame is described with respect to the space frame. The triads are represented in terms of their respective unit vectors: the space frame as Σs(x^1; x^2; x^3) and the object frame as Σc(e^1; e^2; e^3). Here c stands for “corpus,” since o for “object” seems ambiguous. We choose the frames to be right-handed. These orientable objects are not pointlike, and their parametrization offers novel problems. In this sense we may refer to triads as “higher objects,” by contrast to points which are “lower objects.” The thought that comes most easily to mind is to consider the nine direction cosines e^i · x^k but this is impractical, because of the six relations connecting these parameters.
    [Show full text]
  • Coordinate Transformation
    Coordinate Transformation Coordinate Transformations In this chapter, we explore mappings – where a mapping is a function that "maps" one set to another, usually in a way that preserves at least some of the underlyign geometry of the sets. For example, a 2-dimensional coordinate transformation is a mapping of the form T (u; v) = x (u; v) ; y (u; v) h i The functions x (u; v) and y (u; v) are called the components of the transforma- tion. Moreover, the transformation T maps a set S in the uv-plane to a set T (S) in the xy-plane: If S is a region, then we use the components x = f (u; v) and y = g (u; v) to …nd the image of S under T (u; v) : EXAMPLE 1 Find T (S) when T (u; v) = uv; u2 v2 and S is the unit square in the uv-plane (i.e., S = [0; 1] [0; 1]). Solution: To do so, let’s determine the boundary of T (S) in the xy-plane. We use x = uv and y = u2 v2 to …nd the image of the lines bounding the unit square: Side of Square Result of T (u; v) Image in xy-plane v = 0; u in [0; 1] x = 0; y = u2; u in [0; 1] y-axis for 0 y 1 u = 1; v in [0; 1] x = v; y = 1 v2; v in [0; 1] y = 1 x2; x in[0; 1] v = 1; u in [0; 1] x = u; y = u2 1; u in [0; 1] y = x2 1; x in [0; 1] u = 0; u in [0; 1] x = 0; y = v2; v in [0; 1] y-axis for 1 y 0 1 As a result, T (S) is the region in the xy-plane bounded by x = 0; y = x2 1; and y = 1 x2: Linear transformations are coordinate transformations of the form T (u; v) = au + bv; cu + dv h i where a; b; c; and d are constants.
    [Show full text]
  • Unipotent Flows and Applications
    Clay Mathematics Proceedings Volume 10, 2010 Unipotent Flows and Applications Alex Eskin 1. General introduction 1.1. Values of indefinite quadratic forms at integral points. The Op- penheim Conjecture. Let X Q(x1; : : : ; xn) = aijxixj 1≤i≤j≤n be a quadratic form in n variables. We always assume that Q is indefinite so that (so that there exists p with 1 ≤ p < n so that after a linear change of variables, Q can be expresses as: Xp Xn ∗ 2 − 2 Qp(y1; : : : ; yn) = yi yi i=1 i=p+1 We should think of the coefficients aij of Q as real numbers (not necessarily rational or integer). One can still ask what will happen if one substitutes integers for the xi. It is easy to see that if Q is a multiple of a form with rational coefficients, then the set of values Q(Zn) is a discrete subset of R. Much deeper is the following conjecture: Conjecture 1.1 (Oppenheim, 1929). Suppose Q is not proportional to a ra- tional form and n ≥ 5. Then Q(Zn) is dense in the real line. This conjecture was extended by Davenport to n ≥ 3. Theorem 1.2 (Margulis, 1986). The Oppenheim Conjecture is true as long as n ≥ 3. Thus, if n ≥ 3 and Q is not proportional to a rational form, then Q(Zn) is dense in R. This theorem is a triumph of ergodic theory. Before Margulis, the Oppenheim Conjecture was attacked by analytic number theory methods. (In particular it was known for n ≥ 21, and for diagonal forms with n ≥ 5).
    [Show full text]
  • NASA Tit It-462 &I -- .7, F
    N ,! S’ A TECHNICAL, --NASA Tit it-462 &I REP0R.T .7,f ; .A. SPACE-TIME. TENsOR FQRMULATION _ ,~: ., : _. Fq. CONTINUUM MECHANICS IN ’ -. :- -. GENERAL. CURVILINEAR, : MOVING, : ;-_. \. AND DEFQRMING COORDIN@E. SYSTEMS _. Langley Research Center Ehnpton, - Va. 23665 NATIONAL AERONAUTICSAND SPACE ADhiINISTRATI0.N l WASHINGTON, D. C. DECEMBER1976 _. TECHLIBRARY KAFB, NM Illllll lllllIII lllll luII Ill Ill 1111 006857~ ~~_~- - ;. Report No. 2. Government Accession No. 3. Recipient’s Catalog No. NASA TR R-462 5. Report Date A SPACE-TIME TENSOR FORMULATION FOR CONTINUUM December 1976 MECHANICS IN GENERAL CURVILINEAR, MOVING, AND 6. Performing Organization Code DEFORMING COORDINATE SYSTEMS 7. Author(s) 9. Performing Orgamzation Report No. Lee M. Avis L-10552 10. Work Unit No. 9. Performing Organization Name and Address 1’76-30-31-01 NASA Langley Research Center 11. Contract or Grant No. Hampton, VA 23665 13. Type of Report and Period Covered 2. Sponsoring Agency Name and Address Technical Report National Aeronautics and Space Administration 14. Sponsoring Agency Code Washington, DC 20546 %. ~u~~l&~tary Notes 6. Abstract Tensor methods are used to express the continuum equations of motion in general curvilinear, moving, and deforming coordinate systems. The space-time tensor formula- tion is applicable to situations in which, for example, the boundaries move and deform. Placing a coordinate surface on such a boundary simplifies the boundary-condition treat- ment. The space-time tensor formulation is also applicable to coordinate systems with coordinate surfaces defined as surfaces of constant pressure, density, temperature, or any other scalar continuum field function. The vanishing of the function gradient com- ponents along the coordinate surfaces may simplify the set of governing equations.
    [Show full text]
  • Linear Operators Hsiu-Hau Lin [email protected] (Mar 25, 2010)
    Linear Operators Hsiu-Hau Lin [email protected] (Mar 25, 2010) The notes cover linear operators and discuss linear independence of func- tions (Boas 3.7-3.8). • Linear operators An operator maps one thing into another. For instance, the ordinary func- tions are operators mapping numbers to numbers. A linear operator satisfies the properties, O(A + B) = O(A) + O(B);O(kA) = kO(A); (1) where k is a number. As we learned before, a matrix maps one vector into another. One also notices that M(r1 + r2) = Mr1 + Mr2;M(kr) = kMr: Thus, matrices are linear operators. • Orthogonal matrix The length of a vector remains invariant under rotations, ! ! x0 x x0 y0 = x y M T M : y0 y The constraint can be elegantly written down as a matrix equation, M T M = MM T = 1: (2) In other words, M T = M −1. For matrices satisfy the above constraint, they are called orthogonal matrices. Note that, for orthogonal matrices, computing inverse is as simple as taking transpose { an extremely helpful property for calculations. From the product theorem for the determinant, we immediately come to the conclusion det M = ±1. In two dimensions, any 2 × 2 orthogonal matrix with determinant 1 corresponds to a rotation, while any 2 × 2 orthogonal HedgeHog's notes (March 24, 2010) 2 matrix with determinant −1 corresponds to a reflection about a line. Let's come back to our good old friend { the rotation matrix, cos θ − sin θ ! cos θ sin θ ! R(θ) = ;RT = : (3) sin θ cos θ − sin θ cos θ It is straightforward to check that RT R = RRT = 1.
    [Show full text]