Lecture 13: More on Gradient; the Operator 'Del' 13. 1. Examples Of

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 13: More on Gradient; the Operator 'Del' 13. 1. Examples Of Lecture 13: More on Gradient; the Operator `Del' 13. 1. Examples of the Gradient in Physical Laws Gravitational force due to Earth: Consider the potential energy of a particle of mass m, a height z above the Earth's surface V = mgz. Then the force due to gravity can be written as F = V = mg e . −∇ − 3 Gravitational attraction: Now consider the gravitational force on a mass m at r due to a 2 mass m0 at the origin i.e. Newton's law of Gravitation: F = (Gmm0=r )^r. We can write this as − F = V −∇ where the potential energy V = Gmm0=r (see p.51 for how to calculate (1=r)). − r In these two examples we see that force acts down the potential energy gradient. Diffusion: In Physics 2 you may have encountered the idea of diffusion: for example in a gas the molecular motion effectively smoothes out the density. This can be described by the current of particles j(r) being proportional to the gradient of the density (Fick's Law) j(r) = D n(r) − r i.e. the diffusion current flows down the concentration gradient. 13. 2. Examples on gradient Last lecture some examples using `xyz' notation were given. Here we do some exercises with suffix notation (in the lecture we will repeat using `xyz'). As usual suffix notation is most convenient for proving more complicated identities. 2 2 2 2 1. Let φ(r) = r = x1 + x2 + x3, then @ @ @ 2 2 2 φ(r) = e1 + e2 + e3 (x1 + x2 + x3) r @x1 @x2 @x3 = 2x1 e1 + 2x2 e2 + 2x3 e3 = 2r In suffix notation 2 @ r = ei (xjxj) = ei (δijxj + xjδij) = ei 2xi = 2r r @xi In the above we have used the important fact @xi = δij @xj The level surfaces of r2 are spheres centred on the origin, and the gradient of r2 at r points radially outward with magnitude 2r. 49 2. Let φ = a r where a is a constant vector. · @ (a r) = ei (ajxj) = ei ajδij = a r · @xi This is not surprising, since the level surfaces a r = c are planes orthogonal to a. · 3. Let φ(r) = r = x2 + x2 + x2 = (x x )1=2 p 1 2 3 j j @ 1=2 r = ei (xjxj) r @xi 1 1=2 @ = ei (xjxj)− (xkxk) (chain rule) 2 @xi 1 = e 2 x i 2r i 1 = r =r ^ r The gradient of the length of the position vector is the unit vector pointing radially outwards from the origin. It is normal to the level surfaces which are spheres centered on the origin. 13. 3. Identities for gradients If φ(r) and (r) are real scalar fields, then: 1. Distributive law φ(r) + (r) = φ(r) + (r) r r r Proof: @ φ(r) + (r) = ei φ(r) + (r) = φ(r) + (r) r @xi r r 2. Product rule φ(r) (r) = (r) φ(r) + φ(r) (r) r r r Proof: @ φ(r) (r) = ei φ(r) (r) r @xi @φ(r) @ (r) = ei (r) + φ(r) @xi @xi = (r) φ(r) + φ(r) (r) r r 3. Chain rule: If F (φ(r)) is a scalar field, then 50 @F (φ) F (φ(r)) = φ(r) r @φ r Proof: @ F (φ(r)) = ei F (φ(r)) r @xi @F (φ) @φ(r) @F (φ) = ei = φ(r) @φ @xi @φ r Example of Chain rule: If φ(r) = r @F (r) @F (r) 1 F (r) = r = r r @r r @r r where we used a result from 13.2. If F (φ(r)) = φ(r)n = rn we find that n n 1 1 n 2 (r ) = (n r − ) r = (n r − ) r r r In particular 1 r = r r −r3 13. 4. Transformation of the gradient Here we prove the claim that the gradient actually is a vector (so far we assumed it was!). Let the point P have coordinates xi in the ei basis and the same point P have coordinates x0 in the e 0 basis i.e. we consider the vector transformation law xi x0 = λ xj. i i ! i ij φ(r) is a scalar if it depends only on the physical point P and not on the coordinates xi or xi0 used to specify P . The value of φ at P is invariant under a change of basis λ (but the function may look different). φ(x1; x2; x3) φ0(x0 ; x0 ; x0 ) = φ(x1; x2; x3) ! 1 2 3 Now consider φ in the new (primed) basis, its components are r @ φ0(x10 ; x20 ; x30 ) @xi0 Using the chain rule, we obtain @ @xj @ φ0(x10 ; x20 ; x30 ) = φ(x1; x2; x3) : @xi0 @xi0 @xj Since xj = λkj xk0 (inverse vector transformation law) @x j @xk0 = λkj = λkj δik = λij : @xi0 @xi0 51 Hence @ @ φ(x10 ; x20 ; x30 ) = λij φ(x1; x2; x3) : @xi0 @xj which shows that the components of φ respect the vector transformation law. Thus φ(r) transforms as a vector field as claimed.r r 13. 5. The Operator `Del' We can think of the vector operator (confusingly pronounced \del") acting on the scalar field φ(r) to produce the vector fieldr φ(r). r @ @ @ @ In Cartesians: = ei = e1 + e2 + e3 r @xi @x1 @x2 @x3 We call an `operator' since it operates on something to its right. It is a vector operator since it hasr vector transformation properties. (More precisely it is a linear differential vector operator!) We have seen how acts on a scalar field to produce a vector field. We can make products of the vector operatorr with other vector quantities to produce new operators and fields in the same way as we couldr make scalar and vector products of two vectors. For example, recall that the directional derivative of φ in direction ^s was given by ^s φ. Generally, we can interpret A as a scalar operator: · r · r @ A = Ai · r @xi i.e. A acts on a scalar field to its right to produce another scalar field · r @φ(r) @φ(r) @φ(r) @φ(r) (A ) φ(r) = Ai = A1 + A2 + A3 · r @xi @x1 @x2 @x3 Actually we can also act with this operator on a vector field to get another vector field. @ @ (A ) V (r) = Ai V (r) = Ai Vj(r) ej · r @xi @xi = e (A ) V1(r) + e (A ) V2(r) + e (A ) V3(r) 1 · r 2 · r 3 · r The alternative expression A V (r) is undefined because V (r) doesn't make sense. · r r N.B. Great care is required with the order in products since, in general, products involving operators are not commutative. For example A = A r · 6 · r A is a scalar differential operator whereas · r @A A = i gives a scalar field called the divergence ofA r · @xi 52.
Recommended publications
  • A Brief but Important Note on the Product Rule
    A brief but important note on the product rule Peter Merrotsy The University of Western Australia [email protected] The product rule in the Australian Curriculum he product rule refers to the derivative of the product of two functions Texpressed in terms of the functions and their derivatives. This result first naturally appears in the subject Mathematical Methods in the senior secondary Australian Curriculum (Australian Curriculum, Assessment and Reporting Authority [ACARA], n.d.b). In the curriculum content, it is mentioned by name only (Unit 3, Topic 1, ACMMM104). Elsewhere (Glossary, p. 6), detail is given in the form: If h(x) = f(x) g(x), then h'(x) = f(x) g'(x) + f '(x) g(x) (1) or, in Leibniz notation, d dv du ()u v = u + v (2) dx dx dx presupposing, of course, that both u and v are functions of x. In the Australian Capital Territory (ACT Board of Senior Secondary Studies, 2014, pp. 28, 43), Tasmania (Tasmanian Qualifications Authority, 2014, p. 21) and Western Australia (School Curriculum and Standards Authority, 2014, WACE, Mathematical Methods Year 12, pp. 9, 22), these statements of the Australian Senior Mathematics Journal vol. 30 no. 2 product rule have been adopted without further commentary. Elsewhere, there are varying attempts at elaboration. The SACE Board of South Australia (2015, p. 29) refers simply to “differentiating functions of the form , using the product rule.” In Queensland (Queensland Studies Authority, 2008, p. 14), a slight shift in notation is suggested by referring to rules for differentiation, including: d []f (t)g(t) (product rule) (3) dt At first, the Victorian Curriculum and Assessment Authority (2015, p.
    [Show full text]
  • Linear Differential Equations in N Th-Order ;
    Ordinary Differential Equations Second Course Babylon University 2016-2017 Linear Differential Equations in n th-order ; An nth-order linear differential equation has the form ; …. (1) Where the coefficients ;(j=0,1,2,…,n-1) ,and depends solely on the variable (not on or any derivative of ) If then Eq.(1) is homogenous ,if not then is nonhomogeneous . …… 2 A linear differential equation has constant coefficients if all the coefficients are constants ,if one or more is not constant then has variable coefficients . Examples on linear differential equations ; first order nonhomogeneous second order homogeneous third order nonhomogeneous fifth order homogeneous second order nonhomogeneous Theorem 1: Consider the initial value problem given by the linear differential equation(1)and the n initial Conditions; … …(3) Define the differential operator L(y) by …. (4) Where ;(j=0,1,2,…,n-1) is continuous on some interval of interest then L(y) = and a linear homogeneous differential equation written as; L(y) =0 Definition: Linearly Independent Solution; A set of functions … is linearly dependent on if there exists constants … not all zero ,such that …. (5) on 1 Ordinary Differential Equations Second Course Babylon University 2016-2017 A set of functions is linearly dependent if there exists another set of constants not all zero, that is satisfy eq.(5). If the only solution to eq.(5) is ,then the set of functions is linearly independent on The functions are linearly independent Theorem 2: If the functions y1 ,y2, …, yn are solutions for differential
    [Show full text]
  • Symmetry and Tensors Rotations and Tensors
    Symmetry and tensors Rotations and tensors A rotation of a 3-vector is accomplished by an orthogonal transformation. Represented as a matrix, A, we replace each vector, v, by a rotated vector, v0, given by multiplying by A, 0 v = Av In index notation, 0 X vm = Amnvn n Since a rotation must preserve lengths of vectors, we require 02 X 0 0 X 2 v = vmvm = vmvm = v m m Therefore, X X 0 0 vmvm = vmvm m m ! ! X X X = Amnvn Amkvk m n k ! X X = AmnAmk vnvk k;n m Since xn is arbitrary, this is true if and only if X AmnAmk = δnk m t which we can rewrite using the transpose, Amn = Anm, as X t AnmAmk = δnk m In matrix notation, this is t A A = I where I is the identity matrix. This is equivalent to At = A−1. Multi-index objects such as matrices, Mmn, or the Levi-Civita tensor, "ijk, have definite transformation properties under rotations. We call an object a (rotational) tensor if each index transforms in the same way as a vector. An object with no indices, that is, a function, does not transform at all and is called a scalar. 0 A matrix Mmn is a (second rank) tensor if and only if, when we rotate vectors v to v , its new components are given by 0 X Mmn = AmjAnkMjk jk This is what we expect if we imagine Mmn to be built out of vectors as Mmn = umvn, for example. In the same way, we see that the Levi-Civita tensor transforms as 0 X "ijk = AilAjmAkn"lmn lmn 1 Recall that "ijk, because it is totally antisymmetric, is completely determined by only one of its components, say, "123.
    [Show full text]
  • The Wronskian
    Jim Lambers MAT 285 Spring Semester 2012-13 Lecture 16 Notes These notes correspond to Section 3.2 in the text. The Wronskian Now that we know how to solve a linear second-order homogeneous ODE y00 + p(t)y0 + q(t)y = 0 in certain cases, we establish some theory about general equations of this form. First, we introduce some notation. We define a second-order linear differential operator L by L[y] = y00 + p(t)y0 + q(t)y: Then, a initial value problem with a second-order homogeneous linear ODE can be stated as 0 L[y] = 0; y(t0) = y0; y (t0) = z0: We state a result concerning existence and uniqueness of solutions to such ODE, analogous to the Existence-Uniqueness Theorem for first-order ODE. Theorem (Existence-Uniqueness) The initial value problem 00 0 0 L[y] = y + p(t)y + q(t)y = g(t); y(t0) = y0; y (t0) = z0 has a unique solution on an open interval I containing the point t0 if p, q and g are continuous on I. The solution is twice differentiable on I. Now, suppose that we have obtained two solutions y1 and y2 of the equation L[y] = 0. Then 00 0 00 0 y1 + p(t)y1 + q(t)y1 = 0; y2 + p(t)y2 + q(t)y2 = 0: Let y = c1y1 + c2y2, where c1 and c2 are constants. Then L[y] = L[c1y1 + c2y2] 00 0 = (c1y1 + c2y2) + p(t)(c1y1 + c2y2) + q(t)(c1y1 + c2y2) 00 00 0 0 = c1y1 + c2y2 + c1p(t)y1 + c2p(t)y2 + c1q(t)y1 + c2q(t)y2 00 0 00 0 = c1(y1 + p(t)y1 + q(t)y1) + c2(y2 + p(t)y2 + q(t)y2) = 0: We have just established the following theorem.
    [Show full text]
  • A Brief Tour of Vector Calculus
    A BRIEF TOUR OF VECTOR CALCULUS A. HAVENS Contents 0 Prelude ii 1 Directional Derivatives, the Gradient and the Del Operator 1 1.1 Conceptual Review: Directional Derivatives and the Gradient........... 1 1.2 The Gradient as a Vector Field............................ 5 1.3 The Gradient Flow and Critical Points ....................... 10 1.4 The Del Operator and the Gradient in Other Coordinates*............ 17 1.5 Problems........................................ 21 2 Vector Fields in Low Dimensions 26 2 3 2.1 General Vector Fields in Domains of R and R . 26 2.2 Flows and Integral Curves .............................. 31 2.3 Conservative Vector Fields and Potentials...................... 32 2.4 Vector Fields from Frames*.............................. 37 2.5 Divergence, Curl, Jacobians, and the Laplacian................... 41 2.6 Parametrized Surfaces and Coordinate Vector Fields*............... 48 2.7 Tangent Vectors, Normal Vectors, and Orientations*................ 52 2.8 Problems........................................ 58 3 Line Integrals 66 3.1 Defining Scalar Line Integrals............................. 66 3.2 Line Integrals in Vector Fields ............................ 75 3.3 Work in a Force Field................................. 78 3.4 The Fundamental Theorem of Line Integrals .................... 79 3.5 Motion in Conservative Force Fields Conserves Energy .............. 81 3.6 Path Independence and Corollaries of the Fundamental Theorem......... 82 3.7 Green's Theorem.................................... 84 3.8 Problems........................................ 89 4 Surface Integrals, Flux, and Fundamental Theorems 93 4.1 Surface Integrals of Scalar Fields........................... 93 4.2 Flux........................................... 96 4.3 The Gradient, Divergence, and Curl Operators Via Limits* . 103 4.4 The Stokes-Kelvin Theorem..............................108 4.5 The Divergence Theorem ...............................112 4.6 Problems........................................114 List of Figures 117 i 11/14/19 Multivariate Calculus: Vector Calculus Havens 0.
    [Show full text]
  • A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- Sity, Pomona, CA 91768
    A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- sity, Pomona, CA 91768 In a recent calculus course, I introduced the technique of Integration by Parts as an integration rule corresponding to the Product Rule for differentiation. I showed my students the standard derivation of the Integration by Parts formula as presented in [1]: By the Product Rule, if f (x) and g(x) are differentiable functions, then d f (x)g(x) = f (x)g(x) + g(x) f (x). dx Integrating on both sides of this equation, f (x)g(x) + g(x) f (x) dx = f (x)g(x), which may be rearranged to obtain f (x)g(x) dx = f (x)g(x) − g(x) f (x) dx. Letting U = f (x) and V = g(x) and observing that dU = f (x) dx and dV = g(x) dx, we obtain the familiar Integration by Parts formula UdV= UV − VdU. (1) My student Victor asked if we could do a similar thing with the Quotient Rule. While the other students thought this was a crazy idea, I was intrigued. Below, I derive a Quotient Rule Integration by Parts formula, apply the resulting integration formula to an example, and discuss reasons why this formula does not appear in calculus texts. By the Quotient Rule, if f (x) and g(x) are differentiable functions, then ( ) ( ) ( ) − ( ) ( ) d f x = g x f x f x g x . dx g(x) [g(x)]2 Integrating both sides of this equation, we get f (x) g(x) f (x) − f (x)g(x) = dx.
    [Show full text]
  • 18.102 Introduction to Functional Analysis Spring 2009
    MIT OpenCourseWare http://ocw.mit.edu 18.102 Introduction to Functional Analysis Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 108 LECTURE NOTES FOR 18.102, SPRING 2009 Lecture 19. Thursday, April 16 I am heading towards the spectral theory of self-adjoint compact operators. This is rather similar to the spectral theory of self-adjoint matrices and has many useful applications. There is a very effective spectral theory of general bounded but self- adjoint operators but I do not expect to have time to do this. There is also a pretty satisfactory spectral theory of non-selfadjoint compact operators, which it is more likely I will get to. There is no satisfactory spectral theory for general non-compact and non-self-adjoint operators as you can easily see from examples (such as the shift operator). In some sense compact operators are ‘small’ and rather like finite rank operators. If you accept this, then you will want to say that an operator such as (19.1) Id −K; K 2 K(H) is ‘big’. We are quite interested in this operator because of spectral theory. To say that λ 2 C is an eigenvalue of K is to say that there is a non-trivial solution of (19.2) Ku − λu = 0 where non-trivial means other than than the solution u = 0 which always exists. If λ =6 0 we can divide by λ and we are looking for solutions of −1 (19.3) (Id −λ K)u = 0 −1 which is just (19.1) for another compact operator, namely λ K: What are properties of Id −K which migh show it to be ‘big? Here are three: Proposition 26.
    [Show full text]
  • Curl, Divergence and Laplacian
    Curl, Divergence and Laplacian What to know: 1. The definition of curl and it two properties, that is, theorem 1, and be able to predict qualitatively how the curl of a vector field behaves from a picture. 2. The definition of divergence and it two properties, that is, if div F~ 6= 0 then F~ can't be written as the curl of another field, and be able to tell a vector field of clearly nonzero,positive or negative divergence from the picture. 3. Know the definition of the Laplace operator 4. Know what kind of objects those operator take as input and what they give as output. The curl operator Let's look at two plots of vector fields: Figure 1: The vector field Figure 2: The vector field h−y; x; 0i: h1; 1; 0i We can observe that the second one looks like it is rotating around the z axis. We'd like to be able to predict this kind of behavior without having to look at a picture. We also promised to find a criterion that checks whether a vector field is conservative in R3. Both of those goals are accomplished using a tool called the curl operator, even though neither of those two properties is exactly obvious from the definition we'll give. Definition 1. Let F~ = hP; Q; Ri be a vector field in R3, where P , Q and R are continuously differentiable. We define the curl operator: @R @Q @P @R @Q @P curl F~ = − ~i + − ~j + − ~k: (1) @y @z @z @x @x @y Remarks: 1.
    [Show full text]
  • Numerical Operator Calculus in Higher Dimensions
    Numerical operator calculus in higher dimensions Gregory Beylkin* and Martin J. Mohlenkamp Applied Mathematics, University of Colorado, Boulder, CO 80309 Communicated by Ronald R. Coifman, Yale University, New Haven, CT, May 31, 2002 (received for review July 31, 2001) When an algorithm in dimension one is extended to dimension d, equation using only one-dimensional operations and thus avoid- in nearly every case its computational cost is taken to the power d. ing the exponential dependence on d. However, if the best This fundamental difficulty is the single greatest impediment to approximate solution of the form (Eq. 1) is not good enough, solving many important problems and has been dubbed the curse there is no way to improve the accuracy. of dimensionality. For numerical analysis in dimension d,we The natural extension of Eq. 1 is the form propose to use a representation for vectors and matrices that generalizes separation of variables while allowing controlled ac- r ͑ ͒ ϭ ͸ ␾l ͑ ͒ ␾l ͑ ͒ curacy. Basic linear algebra operations can be performed in this f x1,...,xd sl 1 x1 ··· d xd . [2] representation using one-dimensional operations, thus bypassing lϭ1 the exponential scaling with respect to the dimension. Although not all operators and algorithms may be compatible with this The key quantity in Eq. 2 is r, which we call the separation rank. representation, we believe that many of the most important ones By increasing r, the approximate solution can be made as are. We prove that the multiparticle Schro¨dinger operator, as well accurate as desired.
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • Matrix Calculus
    Appendix D Matrix Calculus From too much study, and from extreme passion, cometh madnesse. Isaac Newton [205, §5] − D.1 Gradient, Directional derivative, Taylor series D.1.1 Gradients Gradient of a differentiable real function f(x) : RK R with respect to its vector argument is defined uniquely in terms of partial derivatives→ ∂f(x) ∂x1 ∂f(x) , ∂x2 RK f(x) . (2053) ∇ . ∈ . ∂f(x) ∂xK while the second-order gradient of the twice differentiable real function with respect to its vector argument is traditionally called the Hessian; 2 2 2 ∂ f(x) ∂ f(x) ∂ f(x) 2 ∂x1 ∂x1∂x2 ··· ∂x1∂xK 2 2 2 ∂ f(x) ∂ f(x) ∂ f(x) 2 2 K f(x) , ∂x2∂x1 ∂x2 ··· ∂x2∂xK S (2054) ∇ . ∈ . .. 2 2 2 ∂ f(x) ∂ f(x) ∂ f(x) 2 ∂xK ∂x1 ∂xK ∂x2 ∂x ··· K interpreted ∂f(x) ∂f(x) 2 ∂ ∂ 2 ∂ f(x) ∂x1 ∂x2 ∂ f(x) = = = (2055) ∂x1∂x2 ³∂x2 ´ ³∂x1 ´ ∂x2∂x1 Dattorro, Convex Optimization Euclidean Distance Geometry, Mεβoo, 2005, v2020.02.29. 599 600 APPENDIX D. MATRIX CALCULUS The gradient of vector-valued function v(x) : R RN on real domain is a row vector → v(x) , ∂v1(x) ∂v2(x) ∂vN (x) RN (2056) ∇ ∂x ∂x ··· ∂x ∈ h i while the second-order gradient is 2 2 2 2 , ∂ v1(x) ∂ v2(x) ∂ vN (x) RN v(x) 2 2 2 (2057) ∇ ∂x ∂x ··· ∂x ∈ h i Gradient of vector-valued function h(x) : RK RN on vector domain is → ∂h1(x) ∂h2(x) ∂hN (x) ∂x1 ∂x1 ··· ∂x1 ∂h1(x) ∂h2(x) ∂hN (x) h(x) , ∂x2 ∂x2 ··· ∂x2 ∇ .
    [Show full text]
  • An Introduction to Pseudo-Differential Operators
    An introduction to pseudo-differential operators Jean-Marc Bouclet1 Universit´ede Toulouse 3 Institut de Math´ematiquesde Toulouse [email protected] 2 Contents 1 Background on analysis on manifolds 7 2 The Weyl law: statement of the problem 13 3 Pseudodifferential calculus 19 3.1 The Fourier transform . 19 3.2 Definition of pseudo-differential operators . 21 3.3 Symbolic calculus . 24 3.4 Proofs . 27 4 Some tools of spectral theory 41 4.1 Hilbert-Schmidt operators . 41 4.2 Trace class operators . 44 4.3 Functional calculus via the Helffer-Sj¨ostrandformula . 50 5 L2 bounds for pseudo-differential operators 55 5.1 L2 estimates . 55 5.2 Hilbert-Schmidt estimates . 60 5.3 Trace class estimates . 61 6 Elliptic parametrix and applications 65 n 6.1 Parametrix on R ................................ 65 6.2 Localization of the parametrix . 71 7 Proof of the Weyl law 75 7.1 The resolvent of the Laplacian on a compact manifold . 75 7.2 Diagonalization of ∆g .............................. 78 7.3 Proof of the Weyl law . 81 A Proof of the Peetre Theorem 85 3 4 CONTENTS Introduction The spirit of these notes is to use the famous Weyl law (on the asymptotic distribution of eigenvalues of the Laplace operator on a compact manifold) as a case study to introduce and illustrate one of the many applications of the pseudo-differential calculus. The material presented here corresponds to a 24 hours course taught in Toulouse in 2012 and 2013. We introduce all tools required to give a complete proof of the Weyl law, mainly the semiclassical pseudo-differential calculus, and then of course prove it! The price to pay is that we avoid presenting many classical concepts or results which are not necessary for our purpose (such as Borel summations, principal symbols, invariance by diffeomorphism or the G˚ardinginequality).
    [Show full text]