Introduction to Differential Forms
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The Grassmann Manifold
The Grassmann Manifold 1. For vector spaces V and W denote by L(V; W ) the vector space of linear maps from V to W . Thus L(Rk; Rn) may be identified with the space Rk£n of k £ n matrices. An injective linear map u : Rk ! V is called a k-frame in V . The set k n GFk;n = fu 2 L(R ; R ): rank(u) = kg of k-frames in Rn is called the Stiefel manifold. Note that the special case k = n is the general linear group: k k GLk = fa 2 L(R ; R ) : det(a) 6= 0g: The set of all k-dimensional (vector) subspaces ¸ ½ Rn is called the Grassmann n manifold of k-planes in R and denoted by GRk;n or sometimes GRk;n(R) or n GRk(R ). Let k ¼ : GFk;n ! GRk;n; ¼(u) = u(R ) denote the map which assigns to each k-frame u the subspace u(Rk) it spans. ¡1 For ¸ 2 GRk;n the fiber (preimage) ¼ (¸) consists of those k-frames which form a basis for the subspace ¸, i.e. for any u 2 ¼¡1(¸) we have ¡1 ¼ (¸) = fu ± a : a 2 GLkg: Hence we can (and will) view GRk;n as the orbit space of the group action GFk;n £ GLk ! GFk;n :(u; a) 7! u ± a: The exercises below will prove the following n£k Theorem 2. The Stiefel manifold GFk;n is an open subset of the set R of all n £ k matrices. There is a unique differentiable structure on the Grassmann manifold GRk;n such that the map ¼ is a submersion. -
Topology and Physics 2019 - Lecture 2
Topology and Physics 2019 - lecture 2 Marcel Vonk February 12, 2019 2.1 Maxwell theory in differential form notation Maxwell's theory of electrodynamics is a great example of the usefulness of differential forms. A nice reference on this topic, though somewhat outdated when it comes to notation, is [1]. For notational simplicity, we will work in units where the speed of light, the vacuum permittivity and the vacuum permeability are all equal to 1: c = 0 = µ0 = 1. 2.1.1 The dual field strength In three dimensional space, Maxwell's electrodynamics describes the physics of the electric and magnetic fields E~ and B~ . These are three-dimensional vector fields, but the beauty of the theory becomes much more obvious if we (a) use a four-dimensional relativistic formulation, and (b) write it in terms of differential forms. For example, let us look at Maxwells two source-free, homogeneous equations: r · B = 0;@tB + r × E = 0: (2.1) That these equations have a relativistic flavor becomes clear if we write them out in com- ponents and organize the terms somewhat suggestively: x y z 0 + @xB + @yB + @zB = 0 x z y −@tB + 0 − @yE + @zE = 0 (2.2) y z x −@tB + @xE + 0 − @zE = 0 z y x −@tB − @xE + @yE + 0 = 0 Note that we also multiplied the last three equations by −1 to clarify the structure. All in all, we see that we have four equations (one for each space-time coordinate) which each contain terms in which the four coordinate derivatives act. Therefore, we may be tempted to write our set of equations in more \relativistic" notation as ^µν @µF = 0 (2.3) 1 with F^µν the coordinates of an antisymmetric two-tensor (i. -
Line, Surface and Volume Integrals
Line, Surface and Volume Integrals 1 Line integrals Z Z Z Á dr; a ¢ dr; a £ dr (1) C C C (Á is a scalar ¯eld and a is a vector ¯eld) We divide the path C joining the points A and B into N small line elements ¢rp, p = 1;:::;N. If (xp; yp; zp) is any point on the line element ¢rp, then the second type of line integral in Eq. (1) is de¯ned as Z XN a ¢ dr = lim a(xp; yp; zp) ¢ rp N!1 C p=1 where it is assumed that all j¢rpj ! 0 as N ! 1. 2 Evaluating line integrals The ¯rst type of line integral in Eq. (1) can be written as Z Z Z Á dr = i Á(x; y; z) dx + j Á(x; y; z) dy C CZ C +k Á(x; y; z) dz C The three integrals on the RHS are ordinary scalar integrals. The second and third line integrals in Eq. (1) can also be reduced to a set of scalar integrals by writing the vector ¯eld a in terms of its Cartesian components as a = axi + ayj + azk. Thus, Z Z a ¢ dr = (axi + ayj + azk) ¢ (dxi + dyj + dzk) C ZC = (axdx + aydy + azdz) ZC Z Z = axdx + aydy + azdz C C C 3 Some useful properties about line integrals: 1. Reversing the path of integration changes the sign of the integral. That is, Z B Z A a ¢ dr = ¡ a ¢ dr A B 2. If the path of integration is subdivided into smaller segments, then the sum of the separate line integrals along each segment is equal to the line integral along the whole path. -
LP THEORY of DIFFERENTIAL FORMS on MANIFOLDS This
TRANSACTIONSOF THE AMERICAN MATHEMATICALSOCIETY Volume 347, Number 6, June 1995 LP THEORY OF DIFFERENTIAL FORMS ON MANIFOLDS CHAD SCOTT Abstract. In this paper, we establish a Hodge-type decomposition for the LP space of differential forms on closed (i.e., compact, oriented, smooth) Rieman- nian manifolds. Critical to the proof of this result is establishing an LP es- timate which contains, as a special case, the L2 result referred to by Morrey as Gaffney's inequality. This inequality helps us show the equivalence of the usual definition of Sobolev space with a more geometric formulation which we provide in the case of differential forms on manifolds. We also prove the LP boundedness of Green's operator which we use in developing the LP theory of the Hodge decomposition. For the calculus of variations, we rigorously verify that the spaces of exact and coexact forms are closed in the LP norm. For nonlinear analysis, we demonstrate the existence and uniqueness of a solution to the /1-harmonic equation. 1. Introduction This paper contributes primarily to the development of the LP theory of dif- ferential forms on manifolds. The reader should be aware that for the duration of this paper, manifold will refer only to those which are Riemannian, compact, oriented, C°° smooth and without boundary. For p = 2, the LP theory is well understood and the L2-Hodge decomposition can be found in [M]. However, in the case p ^ 2, the LP theory has yet to be fully developed. Recent appli- cations of the LP theory of differential forms on W to both quasiconformal mappings and nonlinear elasticity continue to motivate interest in this subject. -
Laplacians in Geometric Analysis
LAPLACIANS IN GEOMETRIC ANALYSIS Syafiq Johar syafi[email protected] Contents 1 Trace Laplacian 1 1.1 Connections on Vector Bundles . .1 1.2 Local and Explicit Expressions . .2 1.3 Second Covariant Derivative . .3 1.4 Curvatures on Vector Bundles . .4 1.5 Trace Laplacian . .5 2 Harmonic Functions 6 2.1 Gradient and Divergence Operators . .7 2.2 Laplace-Beltrami Operator . .7 2.3 Harmonic Functions . .8 2.4 Harmonic Maps . .8 3 Hodge Laplacian 9 3.1 Exterior Derivatives . .9 3.2 Hodge Duals . 10 3.3 Hodge Laplacian . 12 4 Hodge Decomposition 13 4.1 De Rham Cohomology . 13 4.2 Hodge Decomposition Theorem . 14 5 Weitzenb¨ock and B¨ochner Formulas 15 5.1 Weitzenb¨ock Formula . 15 5.1.1 0-forms . 15 5.1.2 k-forms . 15 5.2 B¨ochner Formula . 17 1 Trace Laplacian In this section, we are going to present a notion of Laplacian that is regularly used in differential geometry, namely the trace Laplacian (also called the rough Laplacian or connection Laplacian). We recall the definition of connection on vector bundles which allows us to take the directional derivative of vector bundles. 1.1 Connections on Vector Bundles Definition 1.1 (Connection). Let M be a differentiable manifold and E a vector bundle over M. A connection or covariant derivative at a point p 2 M is a map D : Γ(E) ! Γ(T ∗M ⊗ E) 1 with the properties for any V; W 2 TpM; σ; τ 2 Γ(E) and f 2 C (M), we have that DV σ 2 Ep with the following properties: 1. -
NOTES on DIFFERENTIAL FORMS. PART 3: TENSORS 1. What Is A
NOTES ON DIFFERENTIAL FORMS. PART 3: TENSORS 1. What is a tensor? 1 n Let V be a finite-dimensional vector space. It could be R , it could be the tangent space to a manifold at a point, or it could just be an abstract vector space. A k-tensor is a map T : V × · · · × V ! R 2 (where there are k factors of V ) that is linear in each factor. That is, for fixed ~v2; : : : ;~vk, T (~v1;~v2; : : : ;~vk−1;~vk) is a linear function of ~v1, and for fixed ~v1;~v3; : : : ;~vk, T (~v1; : : : ;~vk) is a k ∗ linear function of ~v2, and so on. The space of k-tensors on V is denoted T (V ). Examples: n • If V = R , then the inner product P (~v; ~w) = ~v · ~w is a 2-tensor. For fixed ~v it's linear in ~w, and for fixed ~w it's linear in ~v. n • If V = R , D(~v1; : : : ;~vn) = det ~v1 ··· ~vn is an n-tensor. n • If V = R , T hree(~v) = \the 3rd entry of ~v" is a 1-tensor. • A 0-tensor is just a number. It requires no inputs at all to generate an output. Note that the definition of tensor says nothing about how things behave when you rotate vectors or permute their order. The inner product P stays the same when you swap the two vectors, but the determinant D changes sign when you swap two vectors. Both are tensors. For a 1-tensor like T hree, permuting the order of entries doesn't even make sense! ~ ~ Let fb1;:::; bng be a basis for V . -
1.2 Topological Tensor Calculus
PH211 Physical Mathematics Fall 2019 1.2 Topological tensor calculus 1.2.1 Tensor fields Finite displacements in Euclidean space can be represented by arrows and have a natural vector space structure, but finite displacements in more general curved spaces, such as on the surface of a sphere, do not. However, an infinitesimal neighborhood of a point in a smooth curved space1 looks like an infinitesimal neighborhood of Euclidean space, and infinitesimal displacements dx~ retain the vector space structure of displacements in Euclidean space. An infinitesimal neighborhood of a point can be infinitely rescaled to generate a finite vector space, called the tangent space, at the point. A vector lives in the tangent space of a point. Note that vectors do not stretch from one point to vector tangent space at p p space Figure 1.2.1: A vector in the tangent space of a point. another, and vectors at different points live in different tangent spaces and so cannot be added. For example, rescaling the infinitesimal displacement dx~ by dividing it by the in- finitesimal scalar dt gives the velocity dx~ ~v = (1.2.1) dt which is a vector. Similarly, we can picture the covector rφ as the infinitesimal contours of φ in a neighborhood of a point, infinitely rescaled to generate a finite covector in the point's cotangent space. More generally, infinitely rescaling the neighborhood of a point generates the tensor space and its algebra at the point. The tensor space contains the tangent and cotangent spaces as a vector subspaces. A tensor field is something that takes tensor values at every point in a space. -
Vector Bundles on Projective Space
Vector Bundles on Projective Space Takumi Murayama December 1, 2013 1 Preliminaries on vector bundles Let X be a (quasi-projective) variety over k. We follow [Sha13, Chap. 6, x1.2]. Definition. A family of vector spaces over X is a morphism of varieties π : E ! X −1 such that for each x 2 X, the fiber Ex := π (x) is isomorphic to a vector space r 0 0 Ak(x).A morphism of a family of vector spaces π : E ! X and π : E ! X is a morphism f : E ! E0 such that the following diagram commutes: f E E0 π π0 X 0 and the map fx : Ex ! Ex is linear over k(x). f is an isomorphism if fx is an isomorphism for all x. A vector bundle is a family of vector spaces that is locally trivial, i.e., for each x 2 X, there exists a neighborhood U 3 x such that there is an isomorphism ': π−1(U) !∼ U × Ar that is an isomorphism of families of vector spaces by the following diagram: −1 ∼ r π (U) ' U × A (1.1) π pr1 U −1 where pr1 denotes the first projection. We call π (U) ! U the restriction of the vector bundle π : E ! X onto U, denoted by EjU . r is locally constant, hence is constant on every irreducible component of X. If it is constant everywhere on X, we call r the rank of the vector bundle. 1 The following lemma tells us how local trivializations of a vector bundle glue together on the entire space X. -
The Mean Value Theorem Math 120 Calculus I Fall 2015
The Mean Value Theorem Math 120 Calculus I Fall 2015 The central theorem to much of differential calculus is the Mean Value Theorem, which we'll abbreviate MVT. It is the theoretical tool used to study the first and second derivatives. There is a nice logical sequence of connections here. It starts with the Extreme Value Theorem (EVT) that we looked at earlier when we studied the concept of continuity. It says that any function that is continuous on a closed interval takes on a maximum and a minimum value. A technical lemma. We begin our study with a technical lemma that allows us to relate 0 the derivative of a function at a point to values of the function nearby. Specifically, if f (x0) is positive, then for x nearby but smaller than x0 the values f(x) will be less than f(x0), but for x nearby but larger than x0, the values of f(x) will be larger than f(x0). This says something like f is an increasing function near x0, but not quite. An analogous statement 0 holds when f (x0) is negative. Proof. The proof of this lemma involves the definition of derivative and the definition of limits, but none of the proofs for the rest of the theorems here require that depth. 0 Suppose that f (x0) = p, some positive number. That means that f(x) − f(x ) lim 0 = p: x!x0 x − x0 f(x) − f(x0) So you can make arbitrarily close to p by taking x sufficiently close to x0. -
2 Hilbert Spaces You Should Have Seen Some Examples Last Semester
2 Hilbert spaces You should have seen some examples last semester. The simplest (finite-dimensional) ex- C • A complex Hilbert space H is a complete normed space over whose norm is derived from an ample is Cn with its standard inner product. It’s worth recalling from linear algebra that if V is inner product. That is, we assume that there is a sesquilinear form ( , ): H H C, linear in · · × → an n-dimensional (complex) vector space, then from any set of n linearly independent vectors we the first variable and conjugate linear in the second, such that can manufacture an orthonormal basis e1, e2,..., en using the Gram-Schmidt process. In terms of this basis we can write any v V in the form (f ,д) = (д, f ), ∈ v = a e , a = (v, e ) (f , f ) 0 f H, and (f , f ) = 0 = f = 0. i i i i ≥ ∀ ∈ ⇒ ∑ The norm and inner product are related by which can be derived by taking the inner product of the equation v = aiei with ei. We also have n ∑ (f , f ) = f 2. ∥ ∥ v 2 = a 2. ∥ ∥ | i | i=1 We will always assume that H is separable (has a countable dense subset). ∑ Standard infinite-dimensional examples are l2(N) or l2(Z), the space of square-summable As usual for a normed space, the distance on H is given byd(f ,д) = f д = (f д, f д). • • ∥ − ∥ − − sequences, and L2(Ω) where Ω is a measurable subset of Rn. The Cauchy-Schwarz and triangle inequalities, • √ (f ,д) f д , f + д f + д , | | ≤ ∥ ∥∥ ∥ ∥ ∥ ≤ ∥ ∥ ∥ ∥ 2.1 Orthogonality can be derived fairly easily from the inner product. -
AP Calculus AB Topic List 1. Limits Algebraically 2. Limits Graphically 3
AP Calculus AB Topic List 1. Limits algebraically 2. Limits graphically 3. Limits at infinity 4. Asymptotes 5. Continuity 6. Intermediate value theorem 7. Differentiability 8. Limit definition of a derivative 9. Average rate of change (approximate slope) 10. Tangent lines 11. Derivatives rules and special functions 12. Chain Rule 13. Application of chain rule 14. Derivatives of generic functions using chain rule 15. Implicit differentiation 16. Related rates 17. Derivatives of inverses 18. Logarithmic differentiation 19. Determine function behavior (increasing, decreasing, concavity) given a function 20. Determine function behavior (increasing, decreasing, concavity) given a derivative graph 21. Interpret first and second derivative values in a table 22. Determining if tangent line approximations are over or under estimates 23. Finding critical points and determining if they are relative maximum, relative minimum, or neither 24. Second derivative test for relative maximum or minimum 25. Finding inflection points 26. Finding and justifying critical points from a derivative graph 27. Absolute maximum and minimum 28. Application of maximum and minimum 29. Motion derivatives 30. Vertical motion 31. Mean value theorem 32. Approximating area with rectangles and trapezoids given a function 33. Approximating area with rectangles and trapezoids given a table of values 34. Determining if area approximations are over or under estimates 35. Finding definite integrals graphically 36. Finding definite integrals using given integral values 37. Indefinite integrals with power rule or special derivatives 38. Integration with u-substitution 39. Evaluating definite integrals 40. Definite integrals with u-substitution 41. Solving initial value problems (separable differential equations) 42. Creating a slope field 43. -
Chapter 8 Change of Variables, Parametrizations, Surface Integrals
Chapter 8 Change of Variables, Parametrizations, Surface Integrals x0. The transformation formula In evaluating any integral, if the integral depends on an auxiliary function of the variables involved, it is often a good idea to change variables and try to simplify the integral. The formula which allows one to pass from the original integral to the new one is called the transformation formula (or change of variables formula). It should be noted that certain conditions need to be met before one can achieve this, and we begin by reviewing the one variable situation. Let D be an open interval, say (a; b); in R , and let ' : D! R be a 1-1 , C1 mapping (function) such that '0 6= 0 on D: Put D¤ = '(D): By the hypothesis on '; it's either increasing or decreasing everywhere on D: In the former case D¤ = ('(a);'(b)); and in the latter case, D¤ = ('(b);'(a)): Now suppose we have to evaluate the integral Zb I = f('(u))'0(u) du; a for a nice function f: Now put x = '(u); so that dx = '0(u) du: This change of variable allows us to express the integral as Z'(b) Z I = f(x) dx = sgn('0) f(x) dx; '(a) D¤ where sgn('0) denotes the sign of '0 on D: We then get the transformation formula Z Z f('(u))j'0(u)j du = f(x) dx D D¤ This generalizes to higher dimensions as follows: Theorem Let D be a bounded open set in Rn;' : D! Rn a C1, 1-1 mapping whose Jacobian determinant det(D') is everywhere non-vanishing on D; D¤ = '(D); and f an integrable function on D¤: Then we have the transformation formula Z Z Z Z ¢ ¢ ¢ f('(u))j det D'(u)j du1::: dun = ¢ ¢ ¢ f(x) dx1::: dxn: D D¤ 1 Of course, when n = 1; det D'(u) is simply '0(u); and we recover the old formula.