Chapter 2 Differentiation in Higher Dimensions

Total Page:16

File Type:pdf, Size:1020Kb

Chapter 2 Differentiation in Higher Dimensions Chapter 2 Di®erentiation in higher dimensions 2.1 The Total Derivative Recall that if f : R ! R is a 1-variable function, and a 2 R, we say that f is di®erentiable f(a+h)¡f(a) 0 at x = a if and only if the ratio h tends to a ¯nite limit, denoted f (a), as h tends to 0. There are two possible ways to generalize this for vector ¯elds f : D! Rm; D ⊆ Rn; for points a in the interior D0 of D. (The interior of a set X is de¯ned to be the subset X0 obtained by removing all the boundary points. Since every point of X0 is an interior point, it is open.) The reader seeing this material for the ¯rst time will be well advised to stick to vector ¯elds f with domain all of Rn in the beginning. Even in the one dimensional case, if a function is de¯ned on a closed interval [a; b], say, then one can properly speak of di®erentiability only at points in the open interval (a; b). The ¯rst thing one might do is to ¯x a vector v in Rn and saythat f is di®erentiable along v i® the following limit makes sense: 1 lim (f(a + hv) ¡ f(a)) : h!0 h When it does, we write f 0(a; v) for the limit. Note that this de¯nition makes sense because a is an interior point. Indeed, under this hypothesis, D contains a basic open set U containing a, and so a + hv will, for small enough h, fall into U, allowing us to speak of f(a + hv). This 1 derivative behaves exactly like the one variable derivative and has analogous properties. For example, we have the following Theorem 1 (Mean Value Theorem for scalar ¯elds) Suppose f is a scalar ¯eld. Assume 0 f (a + tv; v) exists for all 0 · t · 1. Then there is a to with 0 · to · 1 for which 0 f(a + v) ¡ f(a) = f (a + t0v; v). Proof. Put Á(t) = f(a + tv). By hypothesis, Á is di®erentiable at every t in [0; 1], and 0 0 Á (t) = f (a + tv; v). By the one variable mean value theorem, there exists a t0 such that 0 Á (t0) is Á(1) ¡ Á(0), which equals f(a + v) ¡ f(a). Done. When v is a unit vector, f 0(a; v) is called the directional derivative of f at a in the direction of v. The disadvantage of this construction is that it forces us to study the change of f in one direction at a time. So we revisit the one-dimensional de¯nition and note that the condition 0 for di®erentiabilityµ there is equivalent¶ to requiring that there exists a constant c (= f (a)), f(a + h) ¡ f(a) ¡ ch such that lim = 0. If we put L(h) = f 0(a)h, then L : R ! R is h!0 h clearly a linear map. We generalize this idea in higher dimensions as follows: De¯nition. Let f : D! Rm (D ⊆ Rn) be a vector ¯eld and a an interior point of D. Then f is di®erentiable at x = a if and only if there exists a linear map L : Rn ! Rm such that jjf(a + u) ¡ f(a) ¡ L(u)jj (¤) lim = 0: u!0 jjujj Note that the norm jj ¢ jj denotes the length of vectors in Rm in the numerator and in Rn in the denominator. This should not lead to any confusion, however. Lemma 1 Such an L, if it exists, is unique. Proof. Suppose we have L; M : Rn ! Rm satisfying (*) at x = a. Then jjL(u) ¡ M(u)jj jjL(u) + f(a) ¡ f(a + u) + (f(a + u) ¡ f(a) ¡ M(u))jj lim = lim u!0 jjujj u!0 jjujj jjL(u) + f(a) ¡ f(a + u)jj · lim u!0 jjujj jjf(a + u) ¡ f(a) ¡ M(u)jj + lim = 0: u!0 jjujj 2 Pick any non-zero v 2 Rn, and set u = tv, with t 2 R. Then, the linearity of L; M implies that L(tv) = tL(v) and M(tv) = tM(v). Consequently, we have jjL(tv) ¡ M(tv)jj lim = 0 t!0 jjtvjj jtj jjL(v) ¡ M(v)jj = lim t!0 jtj jjvjj 1 = jjL(v) ¡ M(v)jj: jjvjj Then L(v) ¡ M(v) must be zero. De¯nition. If the limit condition (¤) holds for a linear map L, we call L the total deriva- tive of f at a, and denote it by Taf. It is mind boggling at ¯rst to think of the derivative as a linear map. A natural question n which arises immediately is to know what the value of Taf is at any vector v in R . We will show in section 2.3 that this value is precisely f 0(a; v), thus linking the two generalizations of the one-dimensional derivative. Sometimes one can guess what the answer should be, and if (*) holds for this choice, then it must be the derivative by uniqueness. Here are two examples which illustrate this. (1) Let f be a constant vector ¯eld, i.e., there exists a vector w 2 Rm such that f(x) = w, for all x in the domain D. Then we claim that f is di®erentiable at any a 2 D0 with derivative zero. Indeed, if we put L(u) = 0, for any u 2 Rn, then (*) is satis¯ed, because f(a + u) ¡ f(a) = w ¡ w = 0. (2) Let f be a linear map. Then we claim that f is di®erentiable everywhere with Taf = f. Indeed, if we put L(u) = f(u), then by the linearity of f, f(a + u) ¡ f(a) = f(u), and so f(a + u) ¡ f(a) ¡ L(u) is zero for any u 2 Rn. Hence (*) holds trivially for this choice of L. Before we leave this section, it will be useful to take note of the following: Lemma 2 Let f1; : : : ; fm be the component (scalar) ¯elds of f. Then f is di®erentiable at a i® each fi is di®erentiable at a. Moreover, T f(v) = (T f1(v); T f2(v); : : : ; T fn(g)). An easy consequence of this lemma is that, when n = 1, f is di®erentiable at a i® the following familiar looking limit exists in Rm: f(a + h) ¡ f(a) lim ; h!0 h 3 0 0 allowing us to suggestively write f (a) instead of Taf. Clearly, f (a) is given by the vector 0 0 0 (f1(a); : : : ; fm(a)), so that (Taf)(h) = f (a)h, for any h 2 R. n Proof. Let f be di®erentiable at a. For each v 2 R , write Li(v) for the i-th component of (Taf)(v). Then Li is clearly linear. Since fi(a + u) ¡ fi(u) ¡ Li(u) is the i-th component of f(a + u) ¡ f(a) ¡ L(u), the norm of the former is less than or equal to that of the latter. This shows that (*) holds with f replaced by fi and L replaced by Li. So fi is di®erentiable for any i. Conversely, suppose each fi di®erentiable. Put L(v) = ((Taf1)(v);:::; (Tafm)(v)). Then L is a linear map, and by the triangle inequality, Xm jjf(a + u) ¡ f(a) ¡ L(u)jj · jfi(a + u) ¡ fi(a) ¡ (Tafi)(u)j: i=1 It follows easily that (*) exists and so f is di®erentiable at a. 2.2 Partial Derivatives n Let fe1; : : : ; eng denote the standard basis of R . The directional derivatives along the unit vectors ej are of special importance. 0 De¯nition. Let j · n. The jth partial derivative of f at x = a is f (a; ej), denoted by @f (a) or Djf(a). @xj @f @f Just as in the case of the total derivative, it can be shown that (a) exists i® i (a) @xj @xj exists for each coordinate ¯eld fi. Example: De¯ne f : R3 ! R2 by f(x; y; z) = (exsin(y); zcos(y)): @f All the partial derivatives exist at any a = (x ; y ; z ). We will show this for and leave 0 0 0 @y it to the reader to check the remaining cases. Note that 1 ex0sin(y0+h) ¡ ex0sin(y0) cos(y + h) ¡ cos(y ) (f(a + he ) ¡ f(a)) = ( ; z 0 0 ): h 2 h 0 h We have to understand the limit as h goes to 0. Then the methods of one variable calculus x0sin(y0) show that the right hand side tends to the ¯nite limit (x0cos(y0)e ; ¡z0sin(y0)), which 4 @f is (a). In e®ect, the partial derivative with respect to y is calculated like a one variable @y @f derivative, keeping x and z ¯xed. Let us note without proof that (a) is (sin(y )ex0sin(y0); 0) @x 0 @f and (a) is (0; cos(y )). @z 0 It is easy to see from the de¯nition that f 0(a; tv) equals tf 0(a; v), for any t 2 R. This 1 1 follows as h (f(a + h(tv)) ¡ f(a)) = t th (f(a + (ht)v) ¡ f(a)). In particular the Mean Value 0 Theorem for scalar ¯elds gives fi(a + hv) ¡ f(a) = hfi (a + t0hv) = hfi(a + ¿v) for some 0 · ¿ · h. We also have the following Lemma 3 Suppose the derivatives of f along any v 2 Rn exist near a and are continuous at a.
Recommended publications
  • The Directional Derivative the Derivative of a Real Valued Function (Scalar field) with Respect to a Vector
    Math 253 The Directional Derivative The derivative of a real valued function (scalar field) with respect to a vector. f(x + h) f(x) What is the vector space analog to the usual derivative in one variable? f 0(x) = lim − ? h 0 h ! x -2 0 2 5 Suppose f is a real valued function (a mapping f : Rn R). ! (e.g. f(x; y) = x2 y2). − 0 Unlike the case with plane figures, functions grow in a variety of z ways at each point on a surface (or n{dimensional structure). We'll be interested in examining the way (f) changes as we move -5 from a point X~ to a nearby point in a particular direction. -2 0 y 2 Figure 1: f(x; y) = x2 y2 − The Directional Derivative x -2 0 2 5 We'll be interested in examining the way (f) changes as we move ~ in a particular direction 0 from a point X to a nearby point . z -5 -2 0 y 2 Figure 2: f(x; y) = x2 y2 − y If we give the direction by using a second vector, say ~u, then for →u X→ + h any real number h, the vector X~ + h~u represents a change in X→ position from X~ along a line through X~ parallel to ~u. →u →u h x Figure 3: Change in position along line parallel to ~u The Directional Derivative x -2 0 2 5 We'll be interested in examining the way (f) changes as we move ~ in a particular direction 0 from a point X to a nearby point .
    [Show full text]
  • Lecture 11 Line Integrals of Vector-Valued Functions (Cont'd)
    Lecture 11 Line integrals of vector-valued functions (cont’d) In the previous lecture, we considered the following physical situation: A force, F(x), which is not necessarily constant in space, is acting on a mass m, as the mass moves along a curve C from point P to point Q as shown in the diagram below. z Q P m F(x(t)) x(t) C y x The goal is to compute the total amount of work W done by the force. Clearly the constant force/straight line displacement formula, W = F · d , (1) where F is force and d is the displacement vector, does not apply here. But the fundamental idea, in the “Spirit of Calculus,” is to break up the motion into tiny pieces over which we can use Eq. (1 as an approximation over small pieces of the curve. We then “sum up,” i.e., integrate, over all contributions to obtain W . The total work is the line integral of the vector field F over the curve C: W = F · dx . (2) ZC Here, we summarize the main steps involved in the computation of this line integral: 1. Step 1: Parametrize the curve C We assume that the curve C can be parametrized, i.e., x(t)= g(t) = (x(t),y(t), z(t)), t ∈ [a, b], (3) so that g(a) is point P and g(b) is point Q. From this parametrization we can compute the velocity vector, ′ ′ ′ ′ v(t)= g (t) = (x (t),y (t), z (t)) . (4) 1 2. Step 2: Compute field vector F(g(t)) over curve C F(g(t)) = F(x(t),y(t), z(t)) (5) = (F1(x(t),y(t), z(t), F2(x(t),y(t), z(t), F3(x(t),y(t), z(t))) , t ∈ [a, b] .
    [Show full text]
  • Cauchy Integral Theorem
    Cauchy’s Theorems I Ang M.S. Augustin-Louis Cauchy October 26, 2012 1789 – 1857 References Murray R. Spiegel Complex V ariables with introduction to conformal mapping and its applications Dennis G. Zill , P. D. Shanahan A F irst Course in Complex Analysis with Applications J. H. Mathews , R. W. Howell Complex Analysis F or Mathematics and Engineerng 1 Summary • Cauchy Integral Theorem Let f be analytic in a simply connected domain D. If C is a simple closed contour that lies in D , and there is no singular point inside the contour, then ˆ f (z) dz = 0 C • Cauchy Integral Formula (For simple pole) If there is a singular point z0 inside the contour, then ˛ f(z) dz = 2πj f(z0) z − z0 • Generalized Cauchy Integral Formula (For pole with any order) ˛ f(z) 2πj (n−1) n dz = f (z0) (z − z0) (n − 1)! • Cauchy Inequality M · n! f (n)(z ) ≤ 0 rn • Gauss Mean Value Theorem ˆ 2π ( ) 1 jθ f(z0) = f z0 + re dθ 2π 0 1 2 Related Mathematics Review 2.1 Stoke’s Theorem ¨ ˛ r × F¯ · dS = F¯ · dr Σ @Σ (The proof is skipped) ¯ Consider F = (FX ;FY ;FZ ) 0 1 x^ y^ z^ B @ @ @ C r × F¯ = det B C @ @x @y @z A FX FY FZ Let FZ = 0 0 1 x^ y^ z^ B @ @ @ C r × F¯ = det B C @ @x @y @z A FX FY 0 dS =ndS ^ , for dS = dxdy , n^ =z ^ . By z^ · z^ = 1 , consider z^ component only for r × F¯ 0 1 x^ y^ z^ ( ) B @ @ @ C @F @F r × F¯ = det B C = Y − X z^ @ @x @y @z A @x @y FX FY 0 i.e.
    [Show full text]
  • Notes on Partial Differential Equations John K. Hunter
    Notes on Partial Differential Equations John K. Hunter Department of Mathematics, University of California at Davis Contents Chapter 1. Preliminaries 1 1.1. Euclidean space 1 1.2. Spaces of continuous functions 1 1.3. H¨olderspaces 2 1.4. Lp spaces 3 1.5. Compactness 6 1.6. Averages 7 1.7. Convolutions 7 1.8. Derivatives and multi-index notation 8 1.9. Mollifiers 10 1.10. Boundaries of open sets 12 1.11. Change of variables 16 1.12. Divergence theorem 16 Chapter 2. Laplace's equation 19 2.1. Mean value theorem 20 2.2. Derivative estimates and analyticity 23 2.3. Maximum principle 26 2.4. Harnack's inequality 31 2.5. Green's identities 32 2.6. Fundamental solution 33 2.7. The Newtonian potential 34 2.8. Singular integral operators 43 Chapter 3. Sobolev spaces 47 3.1. Weak derivatives 47 3.2. Examples 47 3.3. Distributions 50 3.4. Properties of weak derivatives 53 3.5. Sobolev spaces 56 3.6. Approximation of Sobolev functions 57 3.7. Sobolev embedding: p < n 57 3.8. Sobolev embedding: p > n 66 3.9. Boundary values of Sobolev functions 69 3.10. Compactness results 71 3.11. Sobolev functions on Ω ⊂ Rn 73 3.A. Lipschitz functions 75 3.B. Absolutely continuous functions 76 3.C. Functions of bounded variation 78 3.D. Borel measures on R 80 v vi CONTENTS 3.E. Radon measures on R 82 3.F. Lebesgue-Stieltjes measures 83 3.G. Integration 84 3.H. Summary 86 Chapter 4.
    [Show full text]
  • MULTIVARIABLE CALCULUS (MATH 212 ) COMMON TOPIC LIST (Approved by the Occidental College Math Department on August 28, 2009)
    MULTIVARIABLE CALCULUS (MATH 212 ) COMMON TOPIC LIST (Approved by the Occidental College Math Department on August 28, 2009) Required Topics • Multivariable functions • 3-D space o Distance o Equations of Spheres o Graphing Planes Spheres Miscellaneous function graphs Sections and level curves (contour diagrams) Level surfaces • Planes o Equations of planes, in different forms o Equation of plane from three points o Linear approximation and tangent planes • Partial Derivatives o Compute using the definition of partial derivative o Compute using differentiation rules o Interpret o Approximate, given contour diagram, table of values, or sketch of sections o Signs from real-world description o Compute a gradient vector • Vectors o Arithmetic on vectors, pictorially and by components o Dot product compute using components v v v v v ⋅ w = v w cosθ v v v v u ⋅ v = 0⇔ u ⊥ v (for nonzero vectors) o Cross product Direction and length compute using components and determinant • Directional derivatives o estimate from contour diagram o compute using the lim definition t→0 v v o v compute using fu = ∇f ⋅ u ( u a unit vector) • Gradient v o v geometry (3 facts, and understand how they follow from fu = ∇f ⋅ u ) Gradient points in direction of fastest increase Length of gradient is the directional derivative in that direction Gradient is perpendicular to level set o given contour diagram, draw a gradient vector • Chain rules • Higher-order partials o compute o mixed partials are equal (under certain conditions) • Optimization o Locate and
    [Show full text]
  • Engineering Analysis 2 : Multivariate Functions
    Multivariate functions Partial Differentiation Higher Order Partial Derivatives Total Differentiation Line Integrals Surface Integrals Engineering Analysis 2 : Multivariate functions P. Rees, O. Kryvchenkova and P.D. Ledger, [email protected] College of Engineering, Swansea University, UK PDL (CoE) SS 2017 1/ 67 Multivariate functions Partial Differentiation Higher Order Partial Derivatives Total Differentiation Line Integrals Surface Integrals Outline 1 Multivariate functions 2 Partial Differentiation 3 Higher Order Partial Derivatives 4 Total Differentiation 5 Line Integrals 6 Surface Integrals PDL (CoE) SS 2017 2/ 67 Multivariate functions Partial Differentiation Higher Order Partial Derivatives Total Differentiation Line Integrals Surface Integrals Introduction Recall from EG189: analysis of a function of single variable: Differentiation d f (x) d x Expresses the rate of change of a function f with respect to x. Integration Z f (x) d x Reverses the operation of differentiation and can used to work out areas under curves, and as we have just seen to solve ODEs. We’ve seen in vectors we often need to deal with physical fields, that depend on more than one variable. We may be interested in rates of change to each coordinate (e.g. x; y; z), time t, but sometimes also other physical fields as well. PDL (CoE) SS 2017 3/ 67 Multivariate functions Partial Differentiation Higher Order Partial Derivatives Total Differentiation Line Integrals Surface Integrals Introduction (Continue ...) Example 1: the area A of a rectangular plate of width x and breadth y can be calculated A = xy The variables x and y are independent of each other. In this case, the dependent variable A is a function of the two independent variables x and y as A = f (x; y) or A ≡ A(x; y) PDL (CoE) SS 2017 4/ 67 Multivariate functions Partial Differentiation Higher Order Partial Derivatives Total Differentiation Line Integrals Surface Integrals Introduction (Continue ...) Example 2: the volume of a rectangular plate is given by V = xyz where the thickness of the plate is in z-direction.
    [Show full text]
  • The Matrix Calculus You Need for Deep Learning
    The Matrix Calculus You Need For Deep Learning Terence Parr and Jeremy Howard July 3, 2018 (We teach in University of San Francisco's MS in Data Science program and have other nefarious projects underway. You might know Terence as the creator of the ANTLR parser generator. For more material, see Jeremy's fast.ai courses and University of San Francisco's Data Institute in- person version of the deep learning course.) HTML version (The PDF and HTML were generated from markup using bookish) Abstract This paper is an attempt to explain all the matrix calculus you need in order to understand the training of deep neural networks. We assume no math knowledge beyond what you learned in calculus 1, and provide links to help you refresh the necessary math where needed. Note that you do not need to understand this material before you start learning to train and use deep learning in practice; rather, this material is for those who are already familiar with the basics of neural networks, and wish to deepen their understanding of the underlying math. Don't worry if you get stuck at some point along the way|just go back and reread the previous section, and try writing down and working through some examples. And if you're still stuck, we're happy to answer your questions in the Theory category at forums.fast.ai. Note: There is a reference section at the end of the paper summarizing all the key matrix calculus rules and terminology discussed here. arXiv:1802.01528v3 [cs.LG] 2 Jul 2018 1 Contents 1 Introduction 3 2 Review: Scalar derivative rules4 3 Introduction to vector calculus and partial derivatives5 4 Matrix calculus 6 4.1 Generalization of the Jacobian .
    [Show full text]
  • 1 Mean Value Theorem 1 1.1 Applications of the Mean Value Theorem
    Seunghee Ye Ma 8: Week 5 Oct 28 Week 5 Summary In Section 1, we go over the Mean Value Theorem and its applications. In Section 2, we will recap what we have covered so far this term. Topics Page 1 Mean Value Theorem 1 1.1 Applications of the Mean Value Theorem . .1 2 Midterm Review 5 2.1 Proof Techniques . .5 2.2 Sequences . .6 2.3 Series . .7 2.4 Continuity and Differentiability of Functions . .9 1 Mean Value Theorem The Mean Value Theorem is the following result: Theorem 1.1 (Mean Value Theorem). Let f be a continuous function on [a; b], which is differentiable on (a; b). Then, there exists some value c 2 (a; b) such that f(b) − f(a) f 0(c) = b − a Intuitively, the Mean Value Theorem is quite trivial. Say we want to drive to San Francisco, which is 380 miles from Caltech according to Google Map. If we start driving at 8am and arrive at 12pm, we know that we were driving over the speed limit at least once during the drive. This is exactly what the Mean Value Theorem tells us. Since the distance travelled is a continuous function of time, we know that there is a point in time when our speed was ≥ 380=4 >>> speed limit. As we can see from this example, the Mean Value Theorem is usually not a tough theorem to understand. The tricky thing is realizing when you should try to use it. Roughly speaking, we use the Mean Value Theorem when we want to turn the information about a function into information about its derivative, or vice-versa.
    [Show full text]
  • A Brief Tour of Vector Calculus
    A BRIEF TOUR OF VECTOR CALCULUS A. HAVENS Contents 0 Prelude ii 1 Directional Derivatives, the Gradient and the Del Operator 1 1.1 Conceptual Review: Directional Derivatives and the Gradient........... 1 1.2 The Gradient as a Vector Field............................ 5 1.3 The Gradient Flow and Critical Points ....................... 10 1.4 The Del Operator and the Gradient in Other Coordinates*............ 17 1.5 Problems........................................ 21 2 Vector Fields in Low Dimensions 26 2 3 2.1 General Vector Fields in Domains of R and R . 26 2.2 Flows and Integral Curves .............................. 31 2.3 Conservative Vector Fields and Potentials...................... 32 2.4 Vector Fields from Frames*.............................. 37 2.5 Divergence, Curl, Jacobians, and the Laplacian................... 41 2.6 Parametrized Surfaces and Coordinate Vector Fields*............... 48 2.7 Tangent Vectors, Normal Vectors, and Orientations*................ 52 2.8 Problems........................................ 58 3 Line Integrals 66 3.1 Defining Scalar Line Integrals............................. 66 3.2 Line Integrals in Vector Fields ............................ 75 3.3 Work in a Force Field................................. 78 3.4 The Fundamental Theorem of Line Integrals .................... 79 3.5 Motion in Conservative Force Fields Conserves Energy .............. 81 3.6 Path Independence and Corollaries of the Fundamental Theorem......... 82 3.7 Green's Theorem.................................... 84 3.8 Problems........................................ 89 4 Surface Integrals, Flux, and Fundamental Theorems 93 4.1 Surface Integrals of Scalar Fields........................... 93 4.2 Flux........................................... 96 4.3 The Gradient, Divergence, and Curl Operators Via Limits* . 103 4.4 The Stokes-Kelvin Theorem..............................108 4.5 The Divergence Theorem ...............................112 4.6 Problems........................................114 List of Figures 117 i 11/14/19 Multivariate Calculus: Vector Calculus Havens 0.
    [Show full text]
  • MATH 25B Professor: Bena Tshishiku Michele Tienni Lowell House
    MATH 25B Professor: Bena Tshishiku Michele Tienni Lowell House Cambridge, MA 02138 [email protected] Please note that these notes are not official and they are not to be considered as a sub- stitute for your notes. Specifically, you should not cite these notes in any of the work submitted for the class (problem sets, exams). The author does not guarantee the accu- racy of the content of this document. If you find a typo (which will probably happen) feel free to email me. Contents 1. January 225 1.1. Functions5 1.2. Limits6 2. January 247 2.1. Theorems about limits7 2.2. Continuity8 3. January 26 10 3.1. Continuity theorems 10 4. January 29 – Notes by Kim 12 4.1. Least Upper Bound property (the secret sauce of R) 12 4.2. Proof of the intermediate value theorem 13 4.3. Proof of the boundedness theorem 13 5. January 31 – Notes by Natalia 14 5.1. Generalizing the boundedness theorem 14 5.2. Subsets of Rn 14 5.3. Compactness 15 6. February 2 16 6.1. Onion ring theorem 16 6.2. Proof of Theorem 6.5 17 6.3. Compactness of closed rectangles 17 6.4. Further applications of compactness 17 7. February 5 19 7.1. Derivatives 19 7.2. Computing derivatives 20 8. February 7 22 8.1. Chain rule 22 Date: April 25, 2018. 1 8.2. Meaning of f 0 23 9. February 9 25 9.1. Polynomial approximation 25 9.2. Derivative magic wands 26 9.3. Taylor’s Theorem 27 9.4.
    [Show full text]
  • Notes for Math 136: Review of Calculus
    Notes for Math 136: Review of Calculus Gyu Eun Lee These notes were written for my personal use for teaching purposes for the course Math 136 running in the Spring of 2016 at UCLA. They are also intended for use as a general calculus reference for this course. In these notes we will briefly review the results of the calculus of several variables most frequently used in partial differential equations. The selection of topics is inspired by the list of topics given by Strauss at the end of section 1.1, but we will not cover everything. Certain topics are covered in the Appendix to the textbook, and we will not deal with them here. The results of single-variable calculus are assumed to be familiar. Practicality, not rigor, is the aim. This document will be updated throughout the quarter as we encounter material involving additional topics. We will focus on functions of two real variables, u = u(x;t). Most definitions and results listed will have generalizations to higher numbers of variables, and we hope the generalizations will be reasonably straightforward to the reader if they become necessary. Occasionally (notably for the chain rule in its many forms) it will be convenient to work with functions of an arbitrary number of variables. We will be rather loose about the issues of differentiability and integrability: unless otherwise stated, we will assume all derivatives and integrals exist, and if we require it we will assume derivatives are continuous up to the required order. (This is also generally the assumption in Strauss.) 1 Differential calculus of several variables 1.1 Partial derivatives Given a scalar function u = u(x;t) of two variables, we may hold one of the variables constant and regard it as a function of one variable only: vx(t) = u(x;t) = wt(x): Then the partial derivatives of u with respect of x and with respect to t are defined as the familiar derivatives from single variable calculus: ¶u dv ¶u dw = x ; = t : ¶t dt ¶x dx c 2016, Gyu Eun Lee.
    [Show full text]
  • Policy Gradient
    Lecture 7: Policy Gradient Lecture 7: Policy Gradient David Silver Lecture 7: Policy Gradient Outline 1 Introduction 2 Finite Difference Policy Gradient 3 Monte-Carlo Policy Gradient 4 Actor-Critic Policy Gradient Lecture 7: Policy Gradient Introduction Policy-Based Reinforcement Learning In the last lecture we approximated the value or action-value function using parameters θ, V (s) V π(s) θ ≈ Q (s; a) Qπ(s; a) θ ≈ A policy was generated directly from the value function e.g. using -greedy In this lecture we will directly parametrise the policy πθ(s; a) = P [a s; θ] j We will focus again on model-free reinforcement learning Lecture 7: Policy Gradient Introduction Value-Based and Policy-Based RL Value Based Learnt Value Function Implicit policy Value Function Policy (e.g. -greedy) Policy Based Value-Based Actor Policy-Based No Value Function Critic Learnt Policy Actor-Critic Learnt Value Function Learnt Policy Lecture 7: Policy Gradient Introduction Advantages of Policy-Based RL Advantages: Better convergence properties Effective in high-dimensional or continuous action spaces Can learn stochastic policies Disadvantages: Typically converge to a local rather than global optimum Evaluating a policy is typically inefficient and high variance Lecture 7: Policy Gradient Introduction Rock-Paper-Scissors Example Example: Rock-Paper-Scissors Two-player game of rock-paper-scissors Scissors beats paper Rock beats scissors Paper beats rock Consider policies for iterated rock-paper-scissors A deterministic policy is easily exploited A uniform random policy
    [Show full text]