A Primer on Exterior Differential Calculus
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Introduction to Gauge Theory Arxiv:1910.10436V1 [Math.DG] 23
Introduction to Gauge Theory Andriy Haydys 23rd October 2019 This is lecture notes for a course given at the PCMI Summer School “Quantum Field The- ory and Manifold Invariants” (July 1 – July 5, 2019). I describe basics of gauge-theoretic approach to construction of invariants of manifolds. The main example considered here is the Seiberg–Witten gauge theory. However, I tried to present the material in a form, which is suitable for other gauge-theoretic invariants too. Contents 1 Introduction2 2 Bundles and connections4 2.1 Vector bundles . .4 2.1.1 Basic notions . .4 2.1.2 Operations on vector bundles . .5 2.1.3 Sections . .6 2.1.4 Covariant derivatives . .6 2.1.5 The curvature . .8 2.1.6 The gauge group . 10 2.2 Principal bundles . 11 2.2.1 The frame bundle and the structure group . 11 2.2.2 The associated vector bundle . 14 2.2.3 Connections on principal bundles . 16 2.2.4 The curvature of a connection on a principal bundle . 19 arXiv:1910.10436v1 [math.DG] 23 Oct 2019 2.2.5 The gauge group . 21 2.3 The Levi–Civita connection . 22 2.4 Classification of U(1) and SU(2) bundles . 23 2.4.1 Complex line bundles . 24 2.4.2 Quaternionic line bundles . 25 3 The Chern–Weil theory 26 3.1 The Chern–Weil theory . 26 3.1.1 The Chern classes . 28 3.2 The Chern–Simons functional . 30 3.3 The modui space of flat connections . 32 3.3.1 Parallel transport and holonomy . -
Laplacians in Geometric Analysis
LAPLACIANS IN GEOMETRIC ANALYSIS Syafiq Johar syafi[email protected] Contents 1 Trace Laplacian 1 1.1 Connections on Vector Bundles . .1 1.2 Local and Explicit Expressions . .2 1.3 Second Covariant Derivative . .3 1.4 Curvatures on Vector Bundles . .4 1.5 Trace Laplacian . .5 2 Harmonic Functions 6 2.1 Gradient and Divergence Operators . .7 2.2 Laplace-Beltrami Operator . .7 2.3 Harmonic Functions . .8 2.4 Harmonic Maps . .8 3 Hodge Laplacian 9 3.1 Exterior Derivatives . .9 3.2 Hodge Duals . 10 3.3 Hodge Laplacian . 12 4 Hodge Decomposition 13 4.1 De Rham Cohomology . 13 4.2 Hodge Decomposition Theorem . 14 5 Weitzenb¨ock and B¨ochner Formulas 15 5.1 Weitzenb¨ock Formula . 15 5.1.1 0-forms . 15 5.1.2 k-forms . 15 5.2 B¨ochner Formula . 17 1 Trace Laplacian In this section, we are going to present a notion of Laplacian that is regularly used in differential geometry, namely the trace Laplacian (also called the rough Laplacian or connection Laplacian). We recall the definition of connection on vector bundles which allows us to take the directional derivative of vector bundles. 1.1 Connections on Vector Bundles Definition 1.1 (Connection). Let M be a differentiable manifold and E a vector bundle over M. A connection or covariant derivative at a point p 2 M is a map D : Γ(E) ! Γ(T ∗M ⊗ E) 1 with the properties for any V; W 2 TpM; σ; τ 2 Γ(E) and f 2 C (M), we have that DV σ 2 Ep with the following properties: 1. -
1.2 Topological Tensor Calculus
PH211 Physical Mathematics Fall 2019 1.2 Topological tensor calculus 1.2.1 Tensor fields Finite displacements in Euclidean space can be represented by arrows and have a natural vector space structure, but finite displacements in more general curved spaces, such as on the surface of a sphere, do not. However, an infinitesimal neighborhood of a point in a smooth curved space1 looks like an infinitesimal neighborhood of Euclidean space, and infinitesimal displacements dx~ retain the vector space structure of displacements in Euclidean space. An infinitesimal neighborhood of a point can be infinitely rescaled to generate a finite vector space, called the tangent space, at the point. A vector lives in the tangent space of a point. Note that vectors do not stretch from one point to vector tangent space at p p space Figure 1.2.1: A vector in the tangent space of a point. another, and vectors at different points live in different tangent spaces and so cannot be added. For example, rescaling the infinitesimal displacement dx~ by dividing it by the in- finitesimal scalar dt gives the velocity dx~ ~v = (1.2.1) dt which is a vector. Similarly, we can picture the covector rφ as the infinitesimal contours of φ in a neighborhood of a point, infinitely rescaled to generate a finite covector in the point's cotangent space. More generally, infinitely rescaling the neighborhood of a point generates the tensor space and its algebra at the point. The tensor space contains the tangent and cotangent spaces as a vector subspaces. A tensor field is something that takes tensor values at every point in a space. -
A Some Basic Rules of Tensor Calculus
A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e. -
Tensor Calculus and Differential Geometry
Course Notes Tensor Calculus and Differential Geometry 2WAH0 Luc Florack March 10, 2021 Cover illustration: papyrus fragment from Euclid’s Elements of Geometry, Book II [8]. Contents Preface iii Notation 1 1 Prerequisites from Linear Algebra 3 2 Tensor Calculus 7 2.1 Vector Spaces and Bases . .7 2.2 Dual Vector Spaces and Dual Bases . .8 2.3 The Kronecker Tensor . 10 2.4 Inner Products . 11 2.5 Reciprocal Bases . 14 2.6 Bases, Dual Bases, Reciprocal Bases: Mutual Relations . 16 2.7 Examples of Vectors and Covectors . 17 2.8 Tensors . 18 2.8.1 Tensors in all Generality . 18 2.8.2 Tensors Subject to Symmetries . 22 2.8.3 Symmetry and Antisymmetry Preserving Product Operators . 24 2.8.4 Vector Spaces with an Oriented Volume . 31 2.8.5 Tensors on an Inner Product Space . 34 2.8.6 Tensor Transformations . 36 2.8.6.1 “Absolute Tensors” . 37 CONTENTS i 2.8.6.2 “Relative Tensors” . 38 2.8.6.3 “Pseudo Tensors” . 41 2.8.7 Contractions . 43 2.9 The Hodge Star Operator . 43 3 Differential Geometry 47 3.1 Euclidean Space: Cartesian and Curvilinear Coordinates . 47 3.2 Differentiable Manifolds . 48 3.3 Tangent Vectors . 49 3.4 Tangent and Cotangent Bundle . 50 3.5 Exterior Derivative . 51 3.6 Affine Connection . 52 3.7 Lie Derivative . 55 3.8 Torsion . 55 3.9 Levi-Civita Connection . 56 3.10 Geodesics . 57 3.11 Curvature . 58 3.12 Push-Forward and Pull-Back . 59 3.13 Examples . 60 3.13.1 Polar Coordinates in the Euclidean Plane . -
The Divergence Theorem Cartan's Formula II. for Any Smooth Vector
The Divergence Theorem Cartan’s Formula II. For any smooth vector field X and any smooth differential form ω, LX = iX d + diX . Lemma. Let x : U → Rn be a positively oriented chart on (M,G), with volume j ∂ form vM , and X = Pj X ∂xj . Then, we have U √ 1 n ∂( gXj) L v = di v = √ X , X M X M g ∂xj j=1 and √ 1 n ∂( gXj ) tr DX = √ X . g ∂xj j=1 Proof. (i) We have √ 1 n LX vM =diX vM = d(iX gdx ∧···∧dx ) n √ =dX(−1)j−1 gXjdx1 ∧···∧dxj ∧···∧dxn d j=1 n √ = X(−1)j−1d( gXj) ∧ dx1 ∧···∧dxj ∧···∧dxn d j=1 √ n ∂( gXj ) = X dx1 ∧···∧dxn ∂xj j=1 √ 1 n ∂( gXj ) =√ X v . g ∂xj M j=1 ∂X` ` k ∂ (ii) We have D∂/∂xj X = P` ∂xj + Pk Γkj X ∂x` , which implies ∂X` tr DX = X + X Γ` Xk. ∂x` k` ` k Since 1 Γ` = X g`r{∂ g + ∂ g − ∂ g } k` 2 k `r ` kr r k` r,` 1 == X g`r∂ g 2 k `r r,` √ 1 ∂ g ∂ g = k = √k , 2 g g √ √ ∂X` X` ∂ g 1 n ∂( gXj ) tr DX = X + √ k } = √ X . ∂x` g ∂x` g ∂xj ` j=1 Typeset by AMS-TEX 1 2 Corollary. Let (M,g) be an oriented Riemannian manifold. Then, for any X ∈ Γ(TM), d(iX dvg)=tr DX =(div X)dvg. Stokes’ Theorem. Let M be a smooth, oriented n-dimensional manifold with boundary. Let ω be a compactly supported smooth (n − 1)-form on M. -
Differential Forms Diff Geom II, WS 2015/16
J.M. Sullivan, TU Berlin B: Differential Forms Diff Geom II, WS 2015/16 B. DIFFERENTIAL FORMS instance, if S has k elements this gives a k-dimensional vector space with S as basis. We have already seen one-forms (covector fields) on a Given vector spaces V and W, let F be the free vector space over the set V × W. (This consists of formal sums manifold. In general, a k-form is a field of alternating k- P linear forms on the tangent spaces of a manifold. Forms ai(vi, wi) but ignores all the structure we have on the set are the natural objects for integration: a k-form can be in- V × W.) Now let R ⊂ F be the linear subspace spanned by tegrated over an oriented k-submanifold. We start with ten- all elements of the form: sor products and the exterior algebra of multivectors. (v + v0, w) − (v, w) − (v0, w), (v, w + w0) − (v, w) − (v, w0), (av, w) − a(v, w), (v, aw) − a(v, w). B1. Tensor products These correspond of course to the bilinearity conditions Recall that, if V, W and X are vector spaces, then a map we started with. The quotient vector space F/R will be the b: V × W → X is called bilinear if tensor product V ⊗ W. We have started with all possible v ⊗ w as generators and thrown in just enough relations to b(v + v0, w) = b(v, w) + b(v0, w), make the map (v, w) 7→ v ⊗ w be bilinear. b(v, w + w0) = b(v, w) + b(v, w0), The tensor product is commutative: there is a natural linear isomorphism V⊗W → W⊗V such that v⊗w 7→ w⊗v. -
Vector Calculus and Differential Forms with Applications To
Vector Calculus and Differential Forms with Applications to Electromagnetism Sean Roberson May 7, 2015 PREFACE This paper is written as a final project for a course in vector analysis, taught at Texas A&M University - San Antonio in the spring of 2015 as an independent study course. Students in mathematics, physics, engineering, and the sciences usually go through a sequence of three calculus courses before go- ing on to differential equations, real analysis, and linear algebra. In the third course, traditionally reserved for multivariable calculus, stu- dents usually learn how to differentiate functions of several variable and integrate over general domains in space. Very rarely, as was my case, will professors have time to cover the important integral theo- rems using vector functions: Green’s Theorem, Stokes’ Theorem, etc. In some universities, such as UCSD and Cornell, honors students are able to take an accelerated calculus sequence using the text Vector Cal- culus, Linear Algebra, and Differential Forms by John Hamal Hubbard and Barbara Burke Hubbard. Here, students learn multivariable cal- culus using linear algebra and real analysis, and then they generalize familiar integral theorems using the language of differential forms. This paper was written over the course of one semester, where the majority of the book was covered. Some details, such as orientation of manifolds, topology, and the foundation of the integral were skipped to save length. The paper should still be readable by a student with at least three semesters of calculus, one course in linear algebra, and one course in real analysis - all at the undergraduate level. -
Fast Higher-Order Functions for Tensor Calculus with Tensors and Subtensors
Fast Higher-Order Functions for Tensor Calculus with Tensors and Subtensors Cem Bassoy and Volker Schatz Fraunhofer IOSB, Ettlingen 76275, Germany, [email protected] Abstract. Tensors analysis has become a popular tool for solving prob- lems in computational neuroscience, pattern recognition and signal pro- cessing. Similar to the two-dimensional case, algorithms for multidimen- sional data consist of basic operations accessing only a subset of tensor data. With multiple offsets and step sizes, basic operations for subtensors require sophisticated implementations even for entrywise operations. In this work, we discuss the design and implementation of optimized higher-order functions that operate entrywise on tensors and subtensors with any non-hierarchical storage format and arbitrary number of dimen- sions. We propose recursive multi-index algorithms with reduced index computations and additional optimization techniques such as function inlining with partial template specialization. We show that single-index implementations of higher-order functions with subtensors introduce a runtime penalty of an order of magnitude than the recursive and iterative multi-index versions. Including data- and thread-level parallelization, our optimized implementations reach 68% of the maximum throughput of an Intel Core i9-7900X. In comparison with other libraries, the aver- age speedup of our optimized implementations is up to 5x for map-like and more than 9x for reduce-like operations. For symmetric tensors we measured an average speedup of up to 4x. 1 Introduction Many problems in computational neuroscience, pattern recognition, signal pro- cessing and data mining generate massive amounts of multidimensional data with high dimensionality [13,12,9]. Tensors provide a natural representation for massive multidimensional data [10,7]. -
6.10 the Generalized Stokes's Theorem
6.10 The generalized Stokes’s theorem 645 6.9.2 In the text we proved Proposition 6.9.7 in the special case where the mapping f is linear. Prove the general statement, where f is only assumed to be of class C1. n m 1 6.9.3 Let U R be open, f : U R of class C , and ξ a vector field on m ⊂ → R . a. Show that (f ∗Wξ)(x)=W . Exercise 6.9.3: The matrix [Df(x)]!ξ(x) adj(A) of part c is the adjoint ma- b. Let m = n. Show that if [Df(x)] is invertible, then trix of A. The equation in part b (f ∗Φ )(x) = det[Df(x)]Φ 1 . is unsatisfactory: it does not say ξ [Df(x)]− ξ(x) how to represent (f ∗Φ )(x) as the ξ c. Let m = n, let A be a square matrix, and let A be the matrix obtained flux of a vector field when the n n [j,i] × from A by erasing the jth row and the ith column. Let adj(A) be the matrix matrix [Df(x)] is not invertible. i+j whose (i, j)th entry is (adj(A))i,j =( 1) det A . Show that Part c deals with this situation. − [j,i] A(adj(A)) = (det A)I and f ∗Φξ(x)=Φadj([Df(x)])ξ(x) 6.10 The generalized Stokes’s theorem We worked hard to define the exterior derivative and to define orientation of manifolds and of boundaries. Now we are going to reap some rewards for our labor: a higher-dimensional analogue of the fundamental theorem of calculus, Stokes’s theorem. -
3. Introducing Riemannian Geometry
3. Introducing Riemannian Geometry We have yet to meet the star of the show. There is one object that we can place on a manifold whose importance dwarfs all others, at least when it comes to understanding gravity. This is the metric. The existence of a metric brings a whole host of new concepts to the table which, collectively, are called Riemannian geometry.Infact,strictlyspeakingwewillneeda slightly di↵erent kind of metric for our study of gravity, one which, like the Minkowski metric, has some strange minus signs. This is referred to as Lorentzian Geometry and a slightly better name for this section would be “Introducing Riemannian and Lorentzian Geometry”. However, for our immediate purposes the di↵erences are minor. The novelties of Lorentzian geometry will become more pronounced later in the course when we explore some of the physical consequences such as horizons. 3.1 The Metric In Section 1, we informally introduced the metric as a way to measure distances between points. It does, indeed, provide this service but it is not its initial purpose. Instead, the metric is an inner product on each vector space Tp(M). Definition:Ametric g is a (0, 2) tensor field that is: Symmetric: g(X, Y )=g(Y,X). • Non-Degenerate: If, for any p M, g(X, Y ) =0forallY T (M)thenX =0. • 2 p 2 p p With a choice of coordinates, we can write the metric as g = g (x) dxµ dx⌫ µ⌫ ⌦ The object g is often written as a line element ds2 and this expression is abbreviated as 2 µ ⌫ ds = gµ⌫(x) dx dx This is the form that we saw previously in (1.4). -
Hodge Theory
HODGE THEORY PETER S. PARK Abstract. This exposition of Hodge theory is a slightly retooled version of the author's Harvard minor thesis, advised by Professor Joe Harris. Contents 1. Introduction 1 2. Hodge Theory of Compact Oriented Riemannian Manifolds 2 2.1. Hodge star operator 2 2.2. The main theorem 3 2.3. Sobolev spaces 5 2.4. Elliptic theory 11 2.5. Proof of the main theorem 14 3. Hodge Theory of Compact K¨ahlerManifolds 17 3.1. Differential operators on complex manifolds 17 3.2. Differential operators on K¨ahlermanifolds 20 3.3. Bott{Chern cohomology and the @@-Lemma 25 3.4. Lefschetz decomposition and the Hodge index theorem 26 Acknowledgments 30 References 30 1. Introduction Our objective in this exposition is to state and prove the main theorems of Hodge theory. In Section 2, we first describe a key motivation behind the Hodge theory for compact, closed, oriented Riemannian manifolds: the observation that the differential forms that satisfy certain par- tial differential equations depending on the choice of Riemannian metric (forms in the kernel of the associated Laplacian operator, or harmonic forms) turn out to be precisely the norm-minimizing representatives of the de Rham cohomology classes. This naturally leads to the statement our first main theorem, the Hodge decomposition|for a given compact, closed, oriented Riemannian manifold|of the space of smooth k-forms into the image of the Laplacian and its kernel, the sub- space of harmonic forms. We then develop the analytic machinery|specifically, Sobolev spaces and the theory of elliptic differential operators|that we use to prove the aforementioned decom- position, which immediately yields as a corollary the phenomenon of Poincar´eduality.