17.Matrices and Matrix Transformations (SC)

Total Page:16

File Type:pdf, Size:1020Kb

17.Matrices and Matrix Transformations (SC) 17. MATRICES AND MATRIX TRANSFORMATIONS MATRICES ⎛ 2 1 3 0 ⎞ A matrix is a rectangular array of numbers (or M = ⎜ 5 7 −6 8 ⎟ 3 × 4 symbols) enclosed in brackets either curved or ⎜ ⎟ ⎜ 9 −2 6 −3 ⎟ square. The constituents of a matrix are called entries ⎝ ⎠ or elements. A matrix is usually named by a letter for ⎛ 4 5 7 ⎞ convenience. Some examples are shown below. ⎜ −2 3 6 ⎟ N = ⎜ ⎟ 4 × 3 ⎜ −7 1 0 ⎟ æö314 ⎝⎜ 9 −5 8 ⎠⎟ X = ç÷ èø270- Note that the orders 3 × 4 and 4 × 3 are NOT the æö31426- same. A = ç÷ èø00105- Row matrices ⎡ a b ⎤ F = ⎢ ⎥ ⎣ c d ⎦ If a matrix is composed only of one row, then it is called a row matrix (regardless of its number of ⎡ 4 1 2 ⎤ elements). The matrices �, � and � are row matrices, ⎢ ⎥ Y = ⎢ 3 −1 5 ⎥ ⎢ 6 −2 10 ⎥ J = ()413 L =(30- 42) ⎣ ⎦ K = 21 ⎛ 1 0 ⎞ () I = ⎝⎜ 0 1 ⎠⎟ Column matrices Rows and Columns If a matrix is composed of only one column, then it is called a column matrix (regardless of the number of The elements of a matrix are arranged in rows and elements). The matrices �, � and � are column columns. Elements that are written from left to right matrices. (horizontally) are called rows. Elements that are written from top to bottom (vertically) are called ⎛ ⎞ columns. The first row is called ‘row 1’, the second ⎛ ⎞ 6 4 ⎜ ⎟ ‘row 2’, and so on. The first column is called ⎛ 1 ⎞ ⎜ ⎟ 7 P = ⎜ ⎟ Q = 0 R=⎜ ⎟ ‘column 1, the second ‘column 2’, and so on. ⎝ 2 ⎠ ⎜ ⎟ ⎜ 8 ⎟ ⎝⎜ 3 ⎠⎟ ⎝⎜ 9 ⎠⎟ ⎛ 3 1 4 −1 ⎞ Row 1 ⎜ ⎟ Square matrices M = ⎜ 0 11 −5 8 ⎟ Row 2 ⎜ ⎟ ⎝ 2 1 6 −4 ⎠ Row 3 If a matrix has the same number of rows as the Col 1 Col 2 Col 3 Col 4 number of columns, then it is called square. For example, a matrix that has 6 rows and 6 columns is a ⎛ 3 1 4 −1 ⎞ square matrix. We may describe such a matrix as ⎜ ⎟ being square of order 6 or simply a 66´ matrix. An N = ⎜www.faspassmaths.com0 11 − 5 8 ⎟ mm´ matrix is a square matrix of order m. Q and S ⎜ 2 1 6 − 4 ⎟ ⎝ ⎠ are square matrices. Order of a matrix æö11- æö124 Q = ç÷ ç÷ èø42 S =ç÷--112 The order of a matrix is written as � × �, where m ç÷ represents the number of rows and n represents the 22´ èø0011 number of columns. A matrix of order 4 × 3, consists 33´ of 4 rows and 3 columns while a matrix of order 3 × 4 consists of 3 rows and 4 columns. 160 Copyright © 2019. Some Rights Reserved. www.faspassmaths.com ( The position of elements in a matrix Diagonal matrices Every single entry in a matrix has a specific position A diagonal matrix is a square matrix whose non- that can be uniquely described. In describing the diagonal elements are zero. T and V are diagonal position, we use the notation aij where the subscript, matrices. i, refers to the row number and the subscript j refers to the column number of the element. ⎛ 3 0 0 ⎞ Since each position is unique, no two entries can ⎛ 2 0 ⎞ V = ⎜ 0 −5 0 ⎟ T = ⎜ ⎟ ⎜ ⎟ have the same row and column number. The symbol ⎝ 0 5 ⎠ ⎜ 0 0 7 ⎟ a represents an element. ⎝ ⎠ a belongs to the ith row and the jth column. ij a belongs to the 4th row and the 2nd column. Zero matrix 42 a belongs to the 2nd row and the 4th column. 24 If all the elements of any matrix are zero(s), then the For the matrix, A, defined below, we can assign each matrix is called a zero matrix. Some examples are element its unique position shown below. Notice shown below. A zero matrix can be of any order. a ≠ a . For example, a ≠ a . ij ji 13 31 ⎡ 0 0 ⎤ ⎡ 0 0 0 ⎤ ⎡ 0 0 0 ⎤ ⎢ ⎥ ⎢ ⎥ ⎣ ⎦ 0 0 0 0 0 ⎛ 4 1 3 2 ⎞ ⎣ ⎦ ⎣ ⎦ ⎜ ⎟ A = ⎜ 0 −11 6 5 ⎟ ⎜ 5 10 2 1 ⎟ ⎝ − − ⎠ Operations on Matrices Col 1 Col 2 Col 3 Col 4 In performing operations on matrices, there are some restrictions. Unlike numbers, one cannot always add, ⎛ 4 1 3 2 ⎞ 11 12 13 14 Row 1 subtract or multiply any two matrices. In fact, a ⎜ ⎟ division of two matrices is not even possible. A = ⎜ 021 −1122 623 524 ⎟ Row 2 ⎜ ⎟ 5 −10 − 2 1 Row 3 ⎝ 31 32 33 34 ⎠ Addition Diagonal elements Matrices can only be added if they are of the same order. This is done by adding or subtracting Diagonal elements are the elements positioned along corresponding entries. The resulting matrix will also the diagonal of a square matrix. A and B are square be of the same order. matrices and have diagonal elements of the type aii . éù12- 1 éù042- A = êú B = êú Square matrix with three diagonal 43- 5 -163 elements ëû ëû 23´ 23´ æö3111 4 ç÷ A = 20 6 éù1211 12-- 1 13 é 011 42 12 13 ù ç÷22 AB+=êú+ê ú ç÷ 43-- 5 1 63 èø--11 5 833 ëû21 22 23ë 21 22 23 û www.faspassmaths.com23´´23 Square matrix with two diagonal éù()()10++() 2-- 4() 12+ elements AB+=êú11 12 13 41+()- ()() 36+ 53 + ëûêú()21 22 23 æö1411 23´ B = ç÷06 èø22 éù121- AB+=êú ëû398 23´ 161 Copyright © 2019. Some Rights Reserved. www.faspassmaths.com Subtraction Solution Matrices can only be subtracted if they are of the æöæö21 6 3 same order. This is done by subtracting (i) 33A ==ç÷ç÷ corresponding entries. The resulting matrix will also èøèø34 912 be of the same order. æöæö412--- 821 éù12- 1 (ii) -22B =-ç÷ç÷= A = èøèø073-- 0146 êú ëû43- 5 23´ éù042- B = Equal matrices êú-163 ëû 23´ Two matrices are equal if: 1. They are of the same order éù1211 12-- 1 13 é 011 42 12 13 ù AB-=êú-ê ú 2. Their corresponding entries are equal. ëû4321 22-- 5 23ë 1 21 63 22 23 û 23´´23 Both conditions must be satisfied before we can deduce that the matrices are equal. Conversely, if éù()()10-----() 2 4() 12 AB-=êú11 12 13 matrices are equal then we can deduce that they must 41--() ()() 36 - 53 - be of the same order and that their corresponding ëûêú()21 22 23 entries are equal. Consider the matrices S and T. They 23´ are of the same order and their corresponding entries éù163- are the same. Therefore, ST= . AB-=êú ëû532- æö13 æö 13 23´ ç÷ ç÷ ST=46- = 46- ç÷ ç÷ ç÷ ç÷ Rule for addition and subtraction of matrices èø20 èø 20 If two matrices have the same order, the matrices Example 2 are said to be conformable to addition and éù42x éù46 subtraction. We obtain the resulting matrix by Given that A = êú B = êú and adding or subtracting corresponding elements. ëû--y 2 ëû22- AB= , find the value of x and of y. Scalar multiplication Solution If AB= , then A and B are of the same order and If a matrix is multiplied by a scalar then each element their corresponding entries are the same. of the matrix is multiplied by the scalar. The resulting Equating corresponding entries: 26x = -y =2 matrix is of the same order. æöabc \x =3 \y =-2 A = ç÷ èødef www.faspassmaths.com æöka kb kc Example 3 kA = ç÷,k is a scalar èøkd ke kf æö12- æö01 Given that A = ç÷ and B = ç÷ and èø13 èø43 Example 1 æö12x æö21 (i) If A = ç÷, then calculate 3� AB-2 =ç÷y , find the value of x, y and z. èø34 ç÷-z èø2 æö412- (ii) If B = ç÷, then calculate −2� èø073- 162 Copyright © 2019. Some Rights Reserved. www.faspassmaths.com Solution We first need to obtain a single matrix for A – 2B. Order of A Order of B Order of C æö12x 2´3 3´2 2´2 AB-2 =ç÷y 3´4 4´2 3´2 ç÷-z èø2 4´2 2´3 4´3 4´3 3´2 4´2 æö12x æöæö12- 01 ç÷ ç÷ç÷-2 =y If two matrices cannot be multiplied, they are said to èøèø13 43ç÷-z èø2 be non-conformable to multiplication. For example, the product YX is not possible since the number of æö12x columns of Y (2) ≠ the number of rows of X, (4) æöæö12- 02ç÷ ç÷ç÷-=y èøèø13 86ç÷-z YX´ÞThis cannot be done. èø2 33´´2 4 æö12x æö14- ç÷ ç÷= y We will now illustrate how two matrices can be èø--73ç÷-z èø2 multiplied using the matrices A and B below. Equating corresponding entries æö1232 -42=x y -3 =-z æö14- 1 ç÷ -7 = A = ç÷ and B =ç÷---1111 24x =- 2 -z =-3 20- 3 èøç÷0462 y èø -4 =-7 z = 3 x = 2 We wish to determine the matrix AB. 2 y =-72 x =-2 () y =-14 First, check to see if this is possible. AB´= C 2424´´33 ´ The Identity or Unit matrix Based on the above rule, the matrix product �� exists If all the diagonal elements of a diagonal matrix are and the product, � will be a 2´4 matrix. equal to one, then it is called the unit or identity matrix and is denoted by U or I only. We write out the structure of the answer, that is, what C looks like. We set up a 2´4 matrix, using the A unit matrix of A unit matrix of order 3 symbol e, to represent an element.
Recommended publications
  • Parametrizations of K-Nonnegative Matrices
    Parametrizations of k-Nonnegative Matrices Anna Brosowsky, Neeraja Kulkarni, Alex Mason, Joe Suk, Ewin Tang∗ October 2, 2017 Abstract Totally nonnegative (positive) matrices are matrices whose minors are all nonnegative (positive). We generalize the notion of total nonnegativity, as follows. A k-nonnegative (resp. k-positive) matrix has all minors of size k or less nonnegative (resp. positive). We give a generating set for the semigroup of k-nonnegative matrices, as well as relations for certain special cases, i.e. the k = n − 1 and k = n − 2 unitriangular cases. In the above two cases, we find that the set of k-nonnegative matrices can be partitioned into cells, analogous to the Bruhat cells of totally nonnegative matrices, based on their factorizations into generators. We will show that these cells, like the Bruhat cells, are homeomorphic to open balls, and we prove some results about the topological structure of the closure of these cells, and in fact, in the latter case, the cells form a Bruhat-like CW complex. We also give a family of minimal k-positivity tests which form sub-cluster algebras of the total positivity test cluster algebra. We describe ways to jump between these tests, and give an alternate description of some tests as double wiring diagrams. 1 Introduction A totally nonnegative (respectively totally positive) matrix is a matrix whose minors are all nonnegative (respectively positive). Total positivity and nonnegativity are well-studied phenomena and arise in areas such as planar networks, combinatorics, dynamics, statistics and probability. The study of total positivity and total nonnegativity admit many varied applications, some of which are explored in “Totally Nonnegative Matrices” by Fallat and Johnson [5].
    [Show full text]
  • EXAM QUESTIONS (Part Two)
    Created by T. Madas MATRICES EXAM QUESTIONS (Part Two) Created by T. Madas Created by T. Madas Question 1 (**) Find the eigenvalues and the corresponding eigenvectors of the following 2× 2 matrix. 7 6 A = . 6 2 2 3 λ= −2, u = α , λ=11, u = β −3 2 Question 2 (**) A transformation in three dimensional space is defined by the following 3× 3 matrix, where x is a scalar constant. 2− 2 4 C =5x − 2 2 . −1 3 x Show that C is non singular for all values of x . FP1-N , proof Created by T. Madas Created by T. Madas Question 3 (**) The 2× 2 matrix A is given below. 1 8 A = . 8− 11 a) Find the eigenvalues of A . b) Determine an eigenvector for each of the corresponding eigenvalues of A . c) Find a 2× 2 matrix P , so that λ1 0 PT AP = , 0 λ2 where λ1 and λ2 are the eigenvalues of A , with λ1< λ 2 . 1 2 1 2 5 5 λ1= −15, λ 2 = 5 , u= , v = , P = − 2 1 − 2 1 5 5 Created by T. Madas Created by T. Madas Question 4 (**) Describe fully the transformation given by the following 3× 3 matrix. 0.28− 0.96 0 0.96 0.28 0 . 0 0 1 rotation in the z axis, anticlockwise, by arcsin() 0.96 Question 5 (**) A transformation in three dimensional space is defined by the following 3× 3 matrix, where k is a scalar constant. 1− 2 k A = k 2 0 . 2 3 1 Show that the transformation defined by A can be inverted for all values of k .
    [Show full text]
  • Two-Dimensional Geometric Transformations
    Two-Dimensional Geometric Transformations 1 Basic Two-Dimensional Geometric Transformations 2 Matrix Representations and Homogeneous Coordinates 3 Inverse Transformations 4 Two-Dimensional Composite Transformations 5 Other Two-Dimensional Transformations 6 Raster Methods for Geometric Transformations 7 OpenGL Raster Transformations 8 Transformations between Two-Dimensional Coordinate Systems 9 OpenGL Functions for Two-Dimensional Geometric Transformations 10 OpenGL Geometric-Transformation o far, we have seen how we can describe a scene in Programming Examples S 11 Summary terms of graphics primitives, such as line segments and fill areas, and the attributes associated with these primitives. Also, we have explored the scan-line algorithms for displaying output primitives on a raster device. Now, we take a look at transformation operations that we can apply to objects to reposition or resize them. These operations are also used in the viewing routines that convert a world-coordinate scene description to a display for an output device. In addition, they are used in a variety of other applications, such as computer-aided design (CAD) and computer animation. An architect, for example, creates a layout by arranging the orientation and size of the component parts of a design, and a computer animator develops a video sequence by moving the “camera” position or the objects in a scene along specified paths. Operations that are applied to the geometric description of an object to change its position, orientation, or size are called geometric transformations. Sometimes geometric transformations are also referred to as modeling transformations, but some graphics packages make a From Chapter 7 of Computer Graphics with OpenGL®, Fourth Edition, Donald Hearn, M.
    [Show full text]
  • Feature Matching and Heat Flow in Centro-Affine Geometry
    Symmetry, Integrability and Geometry: Methods and Applications SIGMA 16 (2020), 093, 22 pages Feature Matching and Heat Flow in Centro-Affine Geometry Peter J. OLVER y, Changzheng QU z and Yun YANG x y School of Mathematics, University of Minnesota, Minneapolis, MN 55455, USA E-mail: [email protected] URL: http://www.math.umn.edu/~olver/ z School of Mathematics and Statistics, Ningbo University, Ningbo 315211, P.R. China E-mail: [email protected] x Department of Mathematics, Northeastern University, Shenyang, 110819, P.R. China E-mail: [email protected] Received April 02, 2020, in final form September 14, 2020; Published online September 29, 2020 https://doi.org/10.3842/SIGMA.2020.093 Abstract. In this paper, we study the differential invariants and the invariant heat flow in centro-affine geometry, proving that the latter is equivalent to the inviscid Burgers' equa- tion. Furthermore, we apply the centro-affine invariants to develop an invariant algorithm to match features of objects appearing in images. We show that the resulting algorithm com- pares favorably with the widely applied scale-invariant feature transform (SIFT), speeded up robust features (SURF), and affine-SIFT (ASIFT) methods. Key words: centro-affine geometry; equivariant moving frames; heat flow; inviscid Burgers' equation; differential invariant; edge matching 2020 Mathematics Subject Classification: 53A15; 53A55 1 Introduction The main objective in this paper is to study differential invariants and invariant curve flows { in particular the heat flow { in centro-affine geometry. In addition, we will present some basic applications to feature matching in camera images of three-dimensional objects, comparing our method with other popular algorithms.
    [Show full text]
  • Section 2.4–2.5 Partitioned Matrices and LU Factorization
    Section 2.4{2.5 Partitioned Matrices and LU Factorization Gexin Yu [email protected] College of William and Mary Gexin Yu [email protected] Section 2.4{2.5 Partitioned Matrices and LU Factorization One approach to simplify the computation is to partition a matrix into blocks. 2 3 0 −1 5 9 −2 3 Ex: A = 4 −5 2 4 0 −3 1 5. −8 −6 3 1 7 −4 This partition can also be written as the following 2 × 3 block matrix: A A A A = 11 12 13 A21 A22 A23 3 0 −1 In the block form, we have blocks A = and so on. 11 −5 2 4 partition matrices into blocks In real world problems, systems can have huge numbers of equations and un-knowns. Standard computation techniques are inefficient in such cases, so we need to develop techniques which exploit the internal structure of the matrices. In most cases, the matrices of interest have lots of zeros. Gexin Yu [email protected] Section 2.4{2.5 Partitioned Matrices and LU Factorization 2 3 0 −1 5 9 −2 3 Ex: A = 4 −5 2 4 0 −3 1 5. −8 −6 3 1 7 −4 This partition can also be written as the following 2 × 3 block matrix: A A A A = 11 12 13 A21 A22 A23 3 0 −1 In the block form, we have blocks A = and so on. 11 −5 2 4 partition matrices into blocks In real world problems, systems can have huge numbers of equations and un-knowns.
    [Show full text]
  • Geometric Transformation Techniques for Digital I~Iages: a Survey
    GEOMETRIC TRANSFORMATION TECHNIQUES FOR DIGITAL I~IAGES: A SURVEY George Walberg Department of Computer Science Columbia University New York, NY 10027 [email protected] December 1988 Technical Report CUCS-390-88 ABSTRACT This survey presents a wide collection of algorithms for the geometric transformation of digital images. Efficient image transformation algorithms are critically important to the remote sensing, medical imaging, computer vision, and computer graphics communities. We review the growth of this field and compare all the described algorithms. Since this subject is interdisci­ plinary, emphasis is placed on the unification of the terminology, motivation, and contributions of each technique to yield a single coherent framework. This paper attempts to serve a dual role as a survey and a tutorial. It is comprehensive in scope and detailed in style. The primary focus centers on the three components that comprise all geometric transformations: spatial transformations, resampling, and antialiasing. In addition, considerable attention is directed to the dramatic progress made in the development of separable algorithms. The text is supplemented with numerous examples and an extensive bibliography. This work was supponed in part by NSF grant CDR-84-21402. 6.5.1 Pyramids 68 6.5.2 Summed-Area Tables 70 6.6 FREQUENCY CLAMPING 71 6.7 ANTIALIASED LINES AND TEXT 71 6.8 DISCUSSION 72 SECTION 7 SEPARABLE GEOMETRIC TRANSFORMATION ALGORITHMS 7.1 INTRODUCTION 73 7.1.1 Forward Mapping 73 7.1.2 Inverse Mapping 73 7.1.3 Separable Mapping 74 7.2 2-PASS TRANSFORMS 75 7.2.1 Catmull and Smith, 1980 75 7.2.1.1 First Pass 75 7.2.1.2 Second Pass 75 7.2.1.3 2-Pass Algorithm 77 7.2.1.4 An Example: Rotation 77 7.2.1.5 Bottleneck Problem 78 7.2.1.6 Foldover Problem 79 7.2.2 Fraser, Schowengerdt, and Briggs, 1985 80 7.2.3 Fant, 1986 81 7.2.4 Smith, 1987 82 7.3 ROTATION 83 7.3.1 Braccini and Marino, 1980 83 7.3.2 Weiman, 1980 84 7.3.3 Paeth, 1986/ Tanaka.
    [Show full text]
  • Transformation Groups in Non-Riemannian Geometry Charles Frances
    Transformation groups in non-Riemannian geometry Charles Frances To cite this version: Charles Frances. Transformation groups in non-Riemannian geometry. Sophus Lie and Felix Klein: The Erlangen Program and Its Impact in Mathematics and Physics, European Mathematical Society Publishing House, pp.191-216, 2015, 10.4171/148-1/7. hal-03195050 HAL Id: hal-03195050 https://hal.archives-ouvertes.fr/hal-03195050 Submitted on 10 Apr 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Transformation groups in non-Riemannian geometry Charles Frances Laboratoire de Math´ematiques,Universit´eParis-Sud 91405 ORSAY Cedex, France email: [email protected] 2000 Mathematics Subject Classification: 22F50, 53C05, 53C10, 53C50. Keywords: Transformation groups, rigid geometric structures, Cartan geometries. Contents 1 Introduction . 2 2 Rigid geometric structures . 5 2.1 Cartan geometries . 5 2.1.1 Classical examples of Cartan geometries . 7 2.1.2 Rigidity of Cartan geometries . 8 2.2 G-structures . 9 3 The conformal group of a Riemannian manifold . 12 3.1 The theorem of Obata and Ferrand . 12 3.2 Ideas of the proof of the Ferrand-Obata theorem . 13 3.3 Generalizations to rank one parabolic geometries .
    [Show full text]
  • Math 217: True False Practice Professor Karen Smith 1. a Square Matrix Is Invertible If and Only If Zero Is Not an Eigenvalue. Solution Note: True
    (c)2015 UM Math Dept licensed under a Creative Commons By-NC-SA 4.0 International License. Math 217: True False Practice Professor Karen Smith 1. A square matrix is invertible if and only if zero is not an eigenvalue. Solution note: True. Zero is an eigenvalue means that there is a non-zero element in the kernel. For a square matrix, being invertible is the same as having kernel zero. 2. If A and B are 2 × 2 matrices, both with eigenvalue 5, then AB also has eigenvalue 5. Solution note: False. This is silly. Let A = B = 5I2. Then the eigenvalues of AB are 25. 3. If A and B are 2 × 2 matrices, both with eigenvalue 5, then A + B also has eigenvalue 5. Solution note: False. This is silly. Let A = B = 5I2. Then the eigenvalues of A + B are 10. 4. A square matrix has determinant zero if and only if zero is an eigenvalue. Solution note: True. Both conditions are the same as the kernel being non-zero. 5. If B is the B-matrix of some linear transformation V !T V . Then for all ~v 2 V , we have B[~v]B = [T (~v)]B. Solution note: True. This is the definition of B-matrix. 21 2 33 T 6. Suppose 40 2 05 is the matrix of a transformation V ! V with respect to some basis 0 0 1 B = (f1; f2; f3). Then f1 is an eigenvector. Solution note: True. It has eigenvalue 1. The first column of the B-matrix is telling us that T (f1) = f1.
    [Show full text]
  • Geometric Transformation of Finite Element Methods: Theory and Applications
    GEOMETRIC TRANSFORMATION OF FINITE ELEMENT METHODS: THEORY AND APPLICATIONS MICHAEL HOLST AND MARTIN LICHT Abstract. We present a new technique to apply finite element methods to partial differential equations over curved domains. A change of variables along a coordinate transformation satisfying only low regularity assumptions can translate a Poisson problem over a curved physical domain to a Poisson prob- lem over a polyhedral parametric domain. This greatly simplifies both the geo- metric setting and the practical implementation, at the cost of having globally rough non-trivial coefficients and data in the parametric Poisson problem. Our main result is that a recently developed broken Bramble-Hilbert lemma is key in harnessing regularity in the physical problem to prove higher-order finite el- ement convergence rates for the parametric problem. Numerical experiments are given which confirm the predictions of our theory. 1. Introduction The computational theory of partial differential equations has been in a paradoxi- cal situation from its very inception: partial differential equations over domains with curved boundaries are of theoretical and practical interest, but numerical methods are generally conceived only for polyhedral domains. Overcoming this geomet- ric gap continues to inspire much ongoing research in computational mathematics for treating partial differential equations with geometric features. Computational methods for partial differential equations over curved domains commonly adhere to the philosophy of approximating the physical domain of the partial differential equation by a parametric domain. We mention isoparametric finite element meth- ods [2], surface finite element methods [10, 8, 3], or isogeometric analysis [14] as examples, which describe the parametric domain by a polyhedral mesh whose cells are piecewise distorted to approximate the physical domain closely or exactly.
    [Show full text]
  • MATH 9 HOMEWORK ASSIGNMENT GIVEN on FEBRUARY 3, 2019. CIRCLE INVERSION Circle Inversion Circle Inversion (Or Simply Inversion) I
    MATH 9 HOMEWORK ASSIGNMENT GIVEN ON FEBRUARY 3, 2019. CIRCLE INVERSION Circle inversion Circle inversion (or simply inversion) is a geometric transformation of the plane. One can think of the inversion as of reflection with respect to a circle which is analogous to a reflection with respect to a straight line. See wikipedia article on Inversive geometry. Definition of inversion Definition. Given a circle S with the center at point O and having a radius R consider the transformation of the plane taking each point P to a point P 0 such that (i) the image P 0 belongs to a ray OP (ii) the distance jOP 0j satisfies jOP j · jOP 0j = R2. Remark. Strictly speaking the inversion is the transformation not of the whole plane but a plane without point O. The point O does not have an image. It is convenient to think of an “infinitely remote" point O0 added to a euclidian plane so that O and O0 are mapped to each other by the inversion. Basic properties of inversion We formulated and proved some basic properties of the inversion. The most important properties are that the inversion • maps circles and straight lines onto circles and straight lines • preserves angles These properties are stated as homework problems below. Try to prove them indepen- dently. In the following we denote the center of inversion as O, the circle of inversion as S and we use prime to denote the images of various objects under inversion. Classwork Problems The properties presented in problems 1-3 are almost obvious.
    [Show full text]
  • Geometric Transformations
    Chapter 4 Geometric transformations • geometric transformations deal with “moving pixels around” !Fig. 4.1 Fig. 4.1: Some examples for geometric transformations: Rotations, mirroring, distortion, map projection. – shift – rotation – scaling – other distortions • Note: pixels of a digital image must be in a regular (Cartesian, equidistant) grid – i.e., pixel coordinates must be integer 22 Chapter 4. Geometric transformations • BUT: transformation might cause pixels to end up in a different grid – non-integer pixel coordinates !Fig. 4.2 Fig. 4.2: Pixels of rotated image (red) are between the positions of the original pixels (blue) – in other words, they have non-integer pixel coordinates. ) interpolation needed to get new pixels in a new regular grid with integer coordinates ) Geometric transformation of digital images = two separate operations: 1. spatial transformation: rotation, shift, scaling etc. 2. grey-level interpolation: f (x;y) : input image (original image) g(x;y) : output image (resulting image after transformation) x,y: integers (pixel coordinates) g(x;y) = f (x0;y0) = f (a(x;y);b(x;y)) (4.1) i.e., x0 = a(x;y) (4.2) y0 = b(x;y) (4.3) 4.1 Grey level interpolation “Grey level” = pixel value There are two ways to solve the interpolation problem: 1. forward-mapping approach (pixel carry-over): V0.9.93, April 19, 2017 page 23 Melsheimer/Spreen Digital Image Processing Summer 2017 • if a pixel from input image f maps to non-integer coordinates (between four integer pixel positions): divide its grey-level among these four pixels !Fig. 4.3 Fig. 4.3: Pixel carry-over (from Castleman, 1996, Fig.
    [Show full text]
  • Homogeneous Systems (1.5) Linear Independence and Dependence (1.7)
    Math 20F, 2015SS1 / TA: Jor-el Briones / Sec: A01 / Handout Page 1 of3 Homogeneous systems (1.5) Homogenous systems are linear systems in the form Ax = 0, where 0 is the 0 vector. Given a system Ax = b, suppose x = α + t1α1 + t2α2 + ::: + tkαk is a solution (in parametric form) to the system, for any values of t1; t2; :::; tk. Then α is a solution to the system Ax = b (seen by seeting t1 = ::: = tk = 0), and α1; α2; :::; αk are solutions to the homoegeneous system Ax = 0. Terms you should know The zero solution (trivial solution): The zero solution is the 0 vector (a vector with all entries being 0), which is always a solution to the homogeneous system Particular solution: Given a system Ax = b, suppose x = α + t1α1 + t2α2 + ::: + tkαk is a solution (in parametric form) to the system, α is the particular solution to the system. The other terms constitute the solution to the associated homogeneous system Ax = 0. Properties of a Homegeneous System 1. A homogeneous system is ALWAYS consistent, since the zero solution, aka the trivial solution, is always a solution to that system. 2. A homogeneous system with at least one free variable has infinitely many solutions. 3. A homogeneous system with more unknowns than equations has infinitely many solu- tions 4. If α1; α2; :::; αk are solutions to a homogeneous system, then ANY linear combination of α1; α2; :::; αk is also a solution to the homogeneous system Important theorems to know Theorem. (Chapter 1, Theorem 6) Given a consistent system Ax = b, suppose α is a solution to the system.
    [Show full text]