1 Welcome to the World of Linear Algebra: Vector Spaces

Total Page:16

File Type:pdf, Size:1020Kb

1 Welcome to the World of Linear Algebra: Vector Spaces Engineering Mathematics 1{Summer 2012 Linear Algebra The commercial message. Linearity is a key concept in mathematics and its applications.Linear objects are usually nice, smooth, clean, easy to handle. Nonlinear ones are slimy, temperamental, and full of dark corners in which monsters can hide. The world, unfortunately, has a strong tendency toward non-linearity, but there is no way that we can understand nonlinear phenomena without first having a good understanding of linear ones. In fact, one of the most common first approaches to a non-linear problem may be to approximate it by a linear one. 1 Welcome to the world of linear algebra: Vector Spaces Vector spaces, also known as a linear spaces, come in two flavors, real and complex. The main difference between them is what is meant by a scalar. When working with real vector spaces, a scalar is a real number. When working with complex vector spaces, a scalar is a complex number. The important thing is not to mix the two flavors. You either work exclusively with real vector spaces, or exclusively with complex ones. Well . ; nothing is that definite. There are times when one wants to go from one flavor to the other; but that should be done with care. One thing to remember, however, is that real numbers are also part of the complex universe; a real number is just a complex number with zero imaginary part. So when working with complex vector spaces, real numbers are also scalars because they are also complex numbers. So what is a vector space? We could give a vague definition saying it is an environment that is linear algebra friendly. But we need to be more precise. Linear operations are very basic ones, so as a better definition we can say that a vector space is any set of objects in which we have defined two operations: 1. An addition, so that if we have any two objects in that set, we know what their sum is. And that sum should be also an element of the set. 2.A scalar multiplication; it should be possible to multiply objects of the set by scalars to get (usually) new objects in the set. If by scalars we mean real numbers, then we have a real vector space; if we mean complex numbers, we have a complex vector space. There is some freedom in how these operations are defined, but only a little bit. The operations should have the usual properties we associate with sums and products. Before giving examples, we need to be a bit more precise. In general, I will use lower case boldface letters to stand for vectors, a vector being simply an object in a set that has been identified as a vector space, and regularly written lower case letters for scalars (real or complex numbers). If V is a vector space, if v; w are elements of V (in other words, vectors), then their sum is denoted by (written as) v + w, which also has to be in V . If a is a scalar, v an element of V , then av denotes the scalar product of a times v. It should be an element of V . Summary of the information so far. By scalar we mean a complex or a real number. We should be aware that if our scalars are real numbers, they should always be real; if complex, always complex. A vector (or linear) space is any set of objects in which we have defined a sum and a scalar product satisfying some (soon to be specified) properties. In this context, the elements of this set are called vectors. So we can say that a vector set is a bunch of vectors closed under addition and scalar multiplication, a vector is an element of a vector space. Finally!, here are the basic properties these operations have to satisfy. For a set V of objects to deserve to be called a vector space it is, in the first place necessary (as mentioned already several times) that if v, w 2 V , then we have defined v + w 2 V , and if a is a scalar, v 2 V , we have defined av 2 V . These operations must satisfy: 1. (Associativity of the sum) If v; w; z 2 V , then (v + w) + z = v + (w + z). This allows us to write the sum of any number of objects (vectors) without using parentheses: If v; w; y; z 2 V , it makes sense to write v + w + y + z. One person may compute this sum by adding first v + w, then y + z, finally adding these 1 WELCOME TO THE WORLD OF LINEAR ALGEBRA: VECTOR SPACES 2 two together. Another person may first compute w + y, then add v in front to get v + (w + y), finally add z to get (v + (w + y)) + z. The result is the same. 2. (Commutativity of the sum) If v; w 2 V , then v + w = w + v. 3. (Existence of 0) There exists a unique element of V , usually denoted by 0, such that v + 0 = v, for all v 2 V . 4. (Existence of an additive inverse) If v 2 V , there exists a unique element −v 2 V such that v + (−v) = 0. This allows us to define subtraction; one writes v − w for v + (−w). 5. (Distributivity 1) If v; w 2 V , a 2 R, then a(v + w) = av + aw. 6. (Distributivity 2) If a; b 2 R, and v 2 V , then (a + b)v = av + bv. 7. (Associativity 2) If a; b 2 R, v 2 V , then a(bv) = (ab)v = b(av). 8. (One is one) 1v = v for all v 2 V . (Here 1 is the scalar number 1.) These properties have consequences. The main consequence is that one operates with vectors as with numbers, as long as things make sense. So, for example, in general, one can't multiply a vector by a vector (there will be particular exceptions). But otherwise, what seems right is usually right. Here are some examples of what is true in any vector space V . Any scalar times the 0 vector is the zero vector. In symbols: a0 = 0. It is not one of the eight properties listed above, but it follows from them, It is also true that the scalar 0 times any vector is the zero vector: 0v = 0. One good thing is that all the linear spaces we will consider will be quite concrete, and all the properties of the operations quite evident. At least, so I hope. What this abstract part does is to provide a common framework for all these concrete sets of objects. Here are a number of examples. It might be a good idea to check that in every case the 8 properties hold or, at the very least, convince yourself that they hold. • Example 1. The stupid vector space. (But, since we are scientists and engineers, we cannot use the word stupid. We must be dignified. It is usually called the trivial vector space.) It is a silly example, but it needs to be seen. It is the absolute simplest case of a linear space. The space has a single vector, the 0 vector. Addition is quite easy; all you need to know is that 0 + 0 = 0. So is scalar multiplication: If a is a scalar, then a0 = 0. And −0 = 0. The trivial vector space can be either real or complex. The next set of examples consist of real vector spaces. • Example 2. The next vector space, just one degree above the previous one in complexity, is the set R of real numbers. Here the real numbers are forced to play a double role, have something like a double personality: On the one hand they are the numbers they always are, the scalars. But they also are the vectors. If a; b 2 R (I write them in boldface to emphasize that now they are acting as vectors), one defines a + b as usual. And if a 2 R and c is a scalar, then ca is just the usual product ca of c times a, but now interpreted as a vector. • Example 3. Our main examples of real vector spaces are the spaces known as Rn, where n is a positive integer. We already saw R1; it is just R. Now we will meet the next one: the space R2 consists of all pairs of real numbers; in symbols: 2 R = f(a; b): a; b are real numbersg: As we all know, we can identify the elements of R2 with points in the plane; once we fix a system of cartesian coordinates, we identify the pair (a; b) with the point of cartesian coordinates (a; b). Operations are defined in a more or less expected way: (a; b) + (c; d) = (a + c; b + d); (1) a(c; d) = (ac; ad): (2) Verifying associativity, commutativity, and distributivity, reduces to the usual associative, commutative, and distributive properties of the operations for real numbers. The 0 element is 0 = (0; 0); obviously (a; b) + 0 = (a; b) + (0; 0) = (a + 0; b + 0) = (a; b): 1 WELCOME TO THE WORLD OF LINEAR ALGEBRA: VECTOR SPACES 3 The additive inverse is also easy to identify, −(a; b) = (−a; −b): (a; b) + (−a; −b) = (a + (−a); b + (−b)) = (0; 0) = 0: These operations have geometric interpretations. I said that R2 can be identified with points of the plane. But another interpretation is to think of the elements of R2 as being vectors, we represent a pair (a; b) as an arrow beginning at(0; 0) and ending at the point of coordinates (a; b).
Recommended publications
  • Linear Algebra I
    Linear Algebra I Martin Otto Winter Term 2013/14 Contents 1 Introduction7 1.1 Motivating Examples.......................7 1.1.1 The two-dimensional real plane.............7 1.1.2 Three-dimensional real space............... 14 1.1.3 Systems of linear equations over Rn ........... 15 1.1.4 Linear spaces over Z2 ................... 21 1.2 Basics, Notation and Conventions................ 27 1.2.1 Sets............................ 27 1.2.2 Functions......................... 29 1.2.3 Relations......................... 34 1.2.4 Summations........................ 36 1.2.5 Propositional logic.................... 36 1.2.6 Some common proof patterns.............. 37 1.3 Algebraic Structures....................... 39 1.3.1 Binary operations on a set................ 39 1.3.2 Groups........................... 40 1.3.3 Rings and fields...................... 42 1.3.4 Aside: isomorphisms of algebraic structures...... 44 2 Vector Spaces 47 2.1 Vector spaces over arbitrary fields................ 47 2.1.1 The axioms........................ 48 2.1.2 Examples old and new.................. 50 2.2 Subspaces............................. 53 2.2.1 Linear subspaces..................... 53 2.2.2 Affine subspaces...................... 56 2.3 Aside: affine and linear spaces.................. 58 2.4 Linear dependence and independence.............. 60 3 4 Linear Algebra I | Martin Otto 2013 2.4.1 Linear combinations and spans............. 60 2.4.2 Linear (in)dependence.................. 62 2.5 Bases and dimension....................... 65 2.5.1 Bases............................ 65 2.5.2 Finite-dimensional vector spaces............. 66 2.5.3 Dimensions of linear and affine subspaces........ 71 2.5.4 Existence of bases..................... 72 2.6 Products, sums and quotients of spaces............. 73 2.6.1 Direct products...................... 73 2.6.2 Direct sums of subspaces................
    [Show full text]
  • Vector Spaces in Physics
    San Francisco State University Department of Physics and Astronomy August 6, 2015 Vector Spaces in Physics Notes for Ph 385: Introduction to Theoretical Physics I R. Bland TABLE OF CONTENTS Chapter I. Vectors A. The displacement vector. B. Vector addition. C. Vector products. 1. The scalar product. 2. The vector product. D. Vectors in terms of components. E. Algebraic properties of vectors. 1. Equality. 2. Vector Addition. 3. Multiplication of a vector by a scalar. 4. The zero vector. 5. The negative of a vector. 6. Subtraction of vectors. 7. Algebraic properties of vector addition. F. Properties of a vector space. G. Metric spaces and the scalar product. 1. The scalar product. 2. Definition of a metric space. H. The vector product. I. Dimensionality of a vector space and linear independence. J. Components in a rotated coordinate system. K. Other vector quantities. Chapter 2. The special symbols ij and ijk, the Einstein summation convention, and some group theory. A. The Kronecker delta symbol, ij B. The Einstein summation convention. C. The Levi-Civita totally antisymmetric tensor. Groups. The permutation group. The Levi-Civita symbol. D. The cross Product. E. The triple scalar product. F. The triple vector product. The epsilon killer. Chapter 3. Linear equations and matrices. A. Linear independence of vectors. B. Definition of a matrix. C. The transpose of a matrix. D. The trace of a matrix. E. Addition of matrices and multiplication of a matrix by a scalar. F. Matrix multiplication. G. Properties of matrix multiplication. H. The unit matrix I. Square matrices as members of a group.
    [Show full text]
  • An Attempt to Intuitively Introduce the Dot, Wedge, Cross, and Geometric Products
    An attempt to intuitively introduce the dot, wedge, cross, and geometric products Peeter Joot March 21, 2008 1 Motivation. Both the NFCM and GAFP books have axiomatic introductions of the gener- alized (vector, blade) dot and wedge products, but there are elements of both that I was unsatisfied with. Perhaps the biggest issue with both is that they aren’t presented in a dumb enough fashion. NFCM presents but does not prove the generalized dot and wedge product operations in terms of symmetric and antisymmetric sums, but it is really the grade operation that is fundamental. You need that to define the dot product of two bivectors for example. GAFP axiomatic presentation is much clearer, but the definition of general- ized wedge product as the totally antisymmetric sum is a bit strange when all the differential forms book give such a different definition. Here I collect some of my notes on how one starts with the geometric prod- uct action on colinear and perpendicular vectors and gets the familiar results for two and three vector products. I may not try to generalize this, but just want to see things presented in a fashion that makes sense to me. 2 Introduction. The aim of this document is to introduce a “new” powerful vector multiplica- tion operation, the geometric product, to a student with some traditional vector algebra background. The geometric product, also called the Clifford product 1, has remained a relatively obscure mathematical subject. This operation actually makes a great deal of vector manipulation simpler than possible with the traditional methods, and provides a way to naturally expresses many geometric concepts.
    [Show full text]
  • Dirac Notation 1 Vectors
    Physics 324, Fall 2001 Dirac Notation 1 Vectors 1.1 Inner product T Recall from linear algebra: we can represent a vector V as a column vector; then V y = (V )∗ is a row vector, and the inner product (another name for dot product) between two vectors is written as AyB = A∗B1 + A∗B2 + : (1) 1 2 ··· In conventional vector notation, the above is just A~∗ B~ . Note that the inner product of a vector with itself is positive definite; we can define the· norm of a vector to be V = pV V; (2) j j y which is a non-negative real number. (In conventional vector notation, this is V~ , which j j is the length of V~ ). 1.2 Basis vectors We can expand a vector in a set of basis vectors e^i , provided the set is complete, which means that the basis vectors span the whole vectorf space.g The basis is called orthonormal if they satisfy e^iye^j = δij (orthonormality); (3) and an orthonormal basis is complete if they satisfy e^ e^y = I (completeness); (4) X i i i where I is the unit matrix. (note that a column vectore ^i times a row vectore ^iy is a square matrix, following the usual definition of matrix multiplication). Assuming we have a complete orthonormal basis, we can write V = IV = e^ e^yV V e^ ;V (^eyV ) : (5) X i i X i i i i i ≡ i ≡ The Vi are complex numbers; we say that Vi are the components of V in the e^i basis.
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • History of Algebra and Its Implications for Teaching
    Maggio: History of Algebra and its Implications for Teaching History of Algebra and its Implications for Teaching Jaime Maggio Fall 2020 MA 398 Senior Seminar Mentor: Dr.Loth Published by DigitalCommons@SHU, 2021 1 Academic Festival, Event 31 [2021] Abstract Algebra can be described as a branch of mathematics concerned with finding the values of unknown quantities (letters and other general sym- bols) defined by the equations that they satisfy. Algebraic problems have survived in mathematical writings of the Egyptians and Babylonians. The ancient Greeks also contributed to the development of algebraic concepts. In this paper, we will discuss historically famous mathematicians from all over the world along with their key mathematical contributions. Mathe- matical proofs of ancient and modern discoveries will be presented. We will then consider the impacts of incorporating history into the teaching of mathematics courses as an educational technique. 1 https://digitalcommons.sacredheart.edu/acadfest/2021/all/31 2 Maggio: History of Algebra and its Implications for Teaching 1 Introduction In order to understand the way algebra is the way it is today, it is important to understand how it came about starting with its ancient origins. In a mod- ern sense, algebra can be described as a branch of mathematics concerned with finding the values of unknown quantities defined by the equations that they sat- isfy. Algebraic problems have survived in mathematical writings of the Egyp- tians and Babylonians. The ancient Greeks also contributed to the development of algebraic concepts, but these concepts had a heavier focus on geometry [1]. The combination of all of the discoveries of these great mathematicians shaped the way algebra is taught today.
    [Show full text]
  • 1 Sets and Set Notation. Definition 1 (Naive Definition of a Set)
    LINEAR ALGEBRA MATH 2700.006 SPRING 2013 (COHEN) LECTURE NOTES 1 Sets and Set Notation. Definition 1 (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most often name sets using capital letters, like A, B, X, Y , etc., while the elements of a set will usually be given lower-case letters, like x, y, z, v, etc. Two sets X and Y are called equal if X and Y consist of exactly the same elements. In this case we write X = Y . Example 1 (Examples of Sets). (1) Let X be the collection of all integers greater than or equal to 5 and strictly less than 10. Then X is a set, and we may write: X = f5; 6; 7; 8; 9g The above notation is an example of a set being described explicitly, i.e. just by listing out all of its elements. The set brackets {· · ·} indicate that we are talking about a set and not a number, sequence, or other mathematical object. (2) Let E be the set of all even natural numbers. We may write: E = f0; 2; 4; 6; 8; :::g This is an example of an explicity described set with infinitely many elements. The ellipsis (:::) in the above notation is used somewhat informally, but in this case its meaning, that we should \continue counting forever," is clear from the context. (3) Let Y be the collection of all real numbers greater than or equal to 5 and strictly less than 10. Recalling notation from previous math courses, we may write: Y = [5; 10) This is an example of using interval notation to describe a set.
    [Show full text]
  • Low-Level Image Processing with the Structure Multivector
    Low-Level Image Processing with the Structure Multivector Michael Felsberg Bericht Nr. 0202 Institut f¨ur Informatik und Praktische Mathematik der Christian-Albrechts-Universitat¨ zu Kiel Olshausenstr. 40 D – 24098 Kiel e-mail: [email protected] 12. Marz¨ 2002 Dieser Bericht enthalt¨ die Dissertation des Verfassers 1. Gutachter Prof. G. Sommer (Kiel) 2. Gutachter Prof. U. Heute (Kiel) 3. Gutachter Prof. J. J. Koenderink (Utrecht) Datum der mundlichen¨ Prufung:¨ 12.2.2002 To Regina ABSTRACT The present thesis deals with two-dimensional signal processing for computer vi- sion. The main topic is the development of a sophisticated generalization of the one-dimensional analytic signal to two dimensions. Motivated by the fundamental property of the latter, the invariance – equivariance constraint, and by its relation to complex analysis and potential theory, a two-dimensional approach is derived. This method is called the monogenic signal and it is based on the Riesz transform instead of the Hilbert transform. By means of this linear approach it is possible to estimate the local orientation and the local phase of signals which are projections of one-dimensional functions to two dimensions. For general two-dimensional signals, however, the monogenic signal has to be further extended, yielding the structure multivector. The latter approach combines the ideas of the structure tensor and the quaternionic analytic signal. A rich feature set can be extracted from the structure multivector, which contains measures for local amplitudes, the local anisotropy, the local orientation, and two local phases. Both, the monogenic signal and the struc- ture multivector are combined with an appropriate scale-space approach, resulting in generalized quadrature filters.
    [Show full text]
  • Relativistic Inversion, Invariance and Inter-Action
    S S symmetry Article Relativistic Inversion, Invariance and Inter-Action Martin B. van der Mark †,‡ and John G. Williamson *,‡ The Quantum Bicycle Society, 12 Crossburn Terrace, Troon KA1 07HB, Scotland, UK; [email protected] * Correspondence: [email protected] † Formerly of Philips Research, 5656 AE Eindhoven, The Netherlands. ‡ These authors contributed equally to this work. Abstract: A general formula for inversion in a relativistic Clifford–Dirac algebra has been derived. Identifying the base elements of the algebra as those of space and time, the first order differential equations over all quantities proves to encompass the Maxwell equations, leads to a natural extension incorporating rest mass and spin, and allows an integration with relativistic quantum mechanics. Although the algebra is not a division algebra, it parallels reality well: where division is undefined turns out to correspond to physical limits, such as that of the light cone. The divisor corresponds to invariants of dynamical significance, such as the invariant interval, the general invariant quantities in electromagnetism, and the basis set of quantities in the Dirac equation. It is speculated that the apparent 3-dimensionality of nature arises from a beautiful symmetry between the three-vector algebra and each of four sets of three derived spaces in the full 4-dimensional algebra. It is conjectured that elements of inversion may play a role in the interaction of fields and matter. Keywords: invariants; inversion; division; non-division algebra; Dirac algebra; Clifford algebra; geometric algebra; special relativity; photon interaction Citation: van der Mark, M.B.; 1. Introduction Williamson, J.G. Relativistic Inversion, Invariance and Inter-Action.
    [Show full text]
  • The Evolution of Equation-Solving: Linear, Quadratic, and Cubic
    California State University, San Bernardino CSUSB ScholarWorks Theses Digitization Project John M. Pfau Library 2006 The evolution of equation-solving: Linear, quadratic, and cubic Annabelle Louise Porter Follow this and additional works at: https://scholarworks.lib.csusb.edu/etd-project Part of the Mathematics Commons Recommended Citation Porter, Annabelle Louise, "The evolution of equation-solving: Linear, quadratic, and cubic" (2006). Theses Digitization Project. 3069. https://scholarworks.lib.csusb.edu/etd-project/3069 This Thesis is brought to you for free and open access by the John M. Pfau Library at CSUSB ScholarWorks. It has been accepted for inclusion in Theses Digitization Project by an authorized administrator of CSUSB ScholarWorks. For more information, please contact [email protected]. THE EVOLUTION OF EQUATION-SOLVING LINEAR, QUADRATIC, AND CUBIC A Project Presented to the Faculty of California State University, San Bernardino In Partial Fulfillment of the Requirements for the Degre Master of Arts in Teaching: Mathematics by Annabelle Louise Porter June 2006 THE EVOLUTION OF EQUATION-SOLVING: LINEAR, QUADRATIC, AND CUBIC A Project Presented to the Faculty of California State University, San Bernardino by Annabelle Louise Porter June 2006 Approved by: Shawnee McMurran, Committee Chair Date Laura Wallace, Committee Member , (Committee Member Peter Williams, Chair Davida Fischman Department of Mathematics MAT Coordinator Department of Mathematics ABSTRACT Algebra and algebraic thinking have been cornerstones of problem solving in many different cultures over time. Since ancient times, algebra has been used and developed in cultures around the world, and has undergone quite a bit of transformation. This paper is intended as a professional developmental tool to help secondary algebra teachers understand the concepts underlying the algorithms we use, how these algorithms developed, and why they work.
    [Show full text]
  • Mathematics for Earth Science
    Mathematics for Earth Science The module covers concepts such as: • Maths refresher • Fractions, Percentage and Ratios • Unit conversions • Calculating large and small numbers • Logarithms • Trigonometry • Linear relationships Mathematics for Earth Science Contents 1. Terms and Operations 2. Fractions 3. Converting decimals and fractions 4. Percentages 5. Ratios 6. Algebra Refresh 7. Power Operations 8. Scientific Notation 9. Units and Conversions 10. Logarithms 11. Trigonometry 12. Graphs and linear relationships 13. Answers 1. Terms and Operations Glossary , 2 , 3 & 17 are TERMS x4 + 2 + 3 = 17 is an EQUATION 17 is the SUM of + 2 + 3 4 4 4 is an EXPONENT + 2 + 3 = 17 4 3 is a CONSTANT 2 is a COEFFICIENT is a VARIABLE + is an OPERATOR +2 + 3 is an EXPRESSION 4 Equation: A mathematical sentence containing an equal sign. The equal sign demands that the expressions on either side are balanced and equal. Expression: An algebraic expression involves numbers, operation signs, brackets/parenthesis and variables that substitute numbers but does not include an equal sign. Operator: The operation (+ , ,× ,÷) which separates the terms. Term: Parts of an expression− separated by operators which could be a number, variable or product of numbers and variables. Eg. 2 , 3 & 17 Variable: A letter which represents an unknown number. Most common is , but can be any symbol. Constant: Terms that contain only numbers that always have the same value. Coefficient: A number that is partnered with a variable. The term 2 is a coefficient with variable. Between the coefficient and variable is a multiplication. Coefficients of 1 are not shown. Exponent: A value or base that is multiplied by itself a certain number of times.
    [Show full text]
  • 1 Vectors & Tensors
    1 Vectors & Tensors The mathematical modeling of the physical world requires knowledge of quite a few different mathematics subjects, such as Calculus, Differential Equations and Linear Algebra. These topics are usually encountered in fundamental mathematics courses. However, in a more thorough and in-depth treatment of mechanics, it is essential to describe the physical world using the concept of the tensor, and so we begin this book with a comprehensive chapter on the tensor. The chapter is divided into three parts. The first part covers vectors (§1.1-1.7). The second part is concerned with second, and higher-order, tensors (§1.8-1.15). The second part covers much of the same ground as done in the first part, mainly generalizing the vector concepts and expressions to tensors. The final part (§1.16-1.19) (not required in the vast majority of applications) is concerned with generalizing the earlier work to curvilinear coordinate systems. The first part comprises basic vector algebra, such as the dot product and the cross product; the mathematics of how the components of a vector transform between different coordinate systems; the symbolic, index and matrix notations for vectors; the differentiation of vectors, including the gradient, the divergence and the curl; the integration of vectors, including line, double, surface and volume integrals, and the integral theorems. The second part comprises the definition of the tensor (and a re-definition of the vector); dyads and dyadics; the manipulation of tensors; properties of tensors, such as the trace, transpose, norm, determinant and principal values; special tensors, such as the spherical, identity and orthogonal tensors; the transformation of tensor components between different coordinate systems; the calculus of tensors, including the gradient of vectors and higher order tensors and the divergence of higher order tensors and special fourth order tensors.
    [Show full text]