Multivariable Analysis

Total Page:16

File Type:pdf, Size:1020Kb

Multivariable Analysis 1 - Multivariable analysis Roland van der Veen Groningen, 20-1-2020 2 Contents 1 Introduction 5 1.1 Basic notions and notation . 5 2 How to solve equations 7 2.1 Linear algebra . 8 2.2 Derivative . 10 2.3 Elementary Riemann integration . 12 2.4 Mean value theorem and Banach contraction . 15 2.5 Inverse and Implicit function theorems . 17 2.6 Picard's theorem on existence of solutions to ODE . 20 3 Multivariable fundamental theorem of calculus 23 3.1 Exterior algebra . 23 3.2 Differential forms . 26 3.3 Integration . 27 3.4 More on cubes and their boundary . 28 3.5 Exterior derivative . 30 3.6 The fundamental theorem of calculus (Stokes Theorem) . 31 3.7 Fundamental theorem of calculus: Poincar´elemma . 32 3 4 CONTENTS Chapter 1 Introduction The goal of these notes is to explore the notions of differentiation and integration in arbitrarily many variables. The material is focused on answering two basic questions: 1. How to solve an equation? How many solutions can one expect? 2. Is there a higher dimensional analogue for the fundamental theorem of calculus? Can one find a primitive? The equations we will address are systems of non-linear equations in finitely many variables and also ordinary differential equations. The approach will be mostly theoretical, schetching a framework in which one can predict how many solutions there will be without necessarily solving the equation. The key assumption is that everything we do can locally be approximated by linear functions. In other words, everything will be differentiable. One of the main results is that the linearization of the equation predicts the number of solutions and approximates them well locally. This is known as the implicit function theorem. For ordinary differential equations we will prove a similar result on the existence and uniqueness of solutions. To introduce the second question, recall what the fundamental theorem of calculus says. Z b f 0(x)dx = f(b) − f(a) a What if f is now a function depending on two or more variables? In two and three dimensions, vector calculus gives some partial answers involving div, grad, curl and the theorems of Gauss, Green and Stokes. How can one make sense of these and are there any more such theorems perhaps in higher dimensions? The key to understanding this question is to pass from functions to differential forms. In the example above this means passing from f(x) to the differential form f(x)dx. Taking the dx part of our integrands seriously clarifies all formulas and shows the way to a general fundamental theorem of calculus that works in any dimension, known as the (generalized) Stokes theorem: Z Z d! = ! Ω @Ω All the results mentioned in this paragraph are special cases of this powerful theorem. This is not calculus. We made an attempt to prove everything we say so that no black boxes have to be accepted on faith. This self-sufficiency is one of the great strengths of mathematics. The reader is asked to at least try the exercises. Doing exercises (and necessarily failing some!) is an integral part of mathematics. 1.1 Basic notions and notation Most of the material is standard and can be found in references such as Calculus on manifolds by M. Spivak. n n 1 n Pn i We will mostly work in R whose elements are vectors R 3 x = (x ; : : : ; x ) = i=1 x ei where ei is the i-th standard basis vector whose coordinates are all 0 except a 1 at the i-th place. Throughout the text I try to write f functions as A 3 a 7−! a2 + a + 1 2 B, instead of f : A ! B defined by f(a) = a2 + a + 1. An important function jj n p 2 2 2 is the (Euclidean) norm R 3 x 7−! R where j(x1; x2; : : : ; xn)j = x1 + x2 + : : : xn. As the name suggests it satisfies the triangle inequality jx + yj ≤ jxj + jyj. Another piece of notation is that of open and closed subsets of Rn. A set S ⊂ Rn is called open if it is the n n union of (possibly infinitely many) open balls Br(p) = fx 2 R : jx − pj < rg. The set S ⊂ R is closed if its n n complement R − S is open. For example the closed disk Br(p) = fx 2 R : jx − pj ≤ rg is closed. 5 6 CHAPTER 1. INTRODUCTION n n i i For a sequence of points (pk) 2 R and q 2 R we write limk!1 pk = q if for all i we have limk!1 pk = q . i n f m Here q is the i-th coordinate of q (not i-th power!). For a function R ⊃ A −! B ⊂ R we say limp!q f(p) = s if n for any sequence (pk) converging to q in R we have limk!1 f(pk) = s. The function f is said to be continuous at q if limp!q f(p) = f(q) and f is continuous if it is continuous at every point. f is uniformly continuous on set U if for all > 0 there is a δ > 0 such that for all u; v 2 U ju − vj < δ ) jf(u) − f(v)j < . A continuous function on a closed bounded U ⊂ Rn is uniformly continuous. Moreover, continuous functions Rm ! Rn send closed bounded subsets to closed bounded subsets. For more on open sets and continuity we refer to the course on metric and topological spaces next block. The above results and definitions should all be familliar at least in the case n = 1. When formulated in terms of the norm all proofs are identical in the case of Rn. Exercises Exercise 0. True or false: 1. B1(1) [ Bπ(5) is open. 2. [0; 1]10 is closed. 3. [0; 1) is open. n 4. f(x1; : : : xn) 2 R jx1 > 0g is open. 7 2 2 5. fx 2 R : x1 + ··· + x7 = 1g is open. Chapter 2 How to solve equations Under what conditions can a system of n real equations in k + n variables be solved? Naively one may hope that each equation can be used to determine a variable so that in the end k variables are left undetermined and all others are functions of those. For example consider the two systems of two equations on the left and on the right (k = 1; n = 2): x + y + z = 0 sin(x + y) − log(1 − z) = 0 (2.1) 1 −x + y + z = 0 ey − = 0 (2.2) 1 + x − z The system on the left is linear and easy to solve, we get x = 0 and y = −z. The system on the right is hard Figure 2.1: Solutions to the two systems. The yellow surface is the solution to the first equation, blue the second. The positive x; y; z axes are drawn in red, green, blue respectively. to solve explicitly but looks very similar near (0; 0; 0) since sin(x + y) ≈ x + y and log(1 − z) ≈ −z near zero. We will be able to show that just like in the linear situation, a curve of solutions passes through the origin. The key point is that the derivative of the complicated looking functions at the origin is precisely the linear function shown on the left. We will look at equations involving only differentiable functions. This means that locally they can be approximated well by linear functions. The goal of the chapter is to prove the implicit function theorem. Basically it says that the linear approximation decides whether or not a system of equations is solvable locally and if so how many solutions it has. This is illustrated in the figures above. Exercises Exercise 0. The non-linear equation sin(x − y) = 0 has a solution (x; y) = (0; 0). Find a linear equation that approximates the non-linear equation well near (0; 0). Exercise 1. (Linear case) Is it always true that a system of two linear equations in three unknowns has a line of solutions? Prove or 7 8 CHAPTER 2. HOW TO SOLVE EQUATIONS provide counter example. Exercise 2. (Non-linear equation) Solve z and y as a function of x subject to the equations. x2 + y2 + z2 = 1 x + y2 + z3 = 0 Exercise 3. (Linear equation) Write the system below as a matrix equation Ax = b for matrix A and vectors x; b. x + 3y + 2z = 1 x + y + z = 0 −x + y + 4z = 0 Exercise 4. (Three holes) Give a single equation in three unknowns such that the solution set is a bounded subset of R3, looks smooth and two-dimensional everywhere and has a hole. Harder: Can you increase the number of holes to three? 2.1 Linear algebra The basis for our investigation of equations is the linear case. Linear equations can neatly be summarized in terms of a single matrix equation Av = b. Here v is a vector in Rk+n, and b 2 Rn and A is an n × (k + n) matrix. In case b = 0 we call the equation homogeneous and the solution set is some linear subspace ker A = fv 2 Rk+njAv = 0g, the kernel of the map defined by A. In general, given a single solution p 2 Rk+n such that Ap = b the entire solution set fv 2 Rk+njAv = bg is the affine linear subspace (ker A) + p = fs + pjs 2 ker Ag. In discussing the qualitative properties of linear equations it is more convenient to think in terms of linear maps.
Recommended publications
  • A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- Sity, Pomona, CA 91768
    A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- sity, Pomona, CA 91768 In a recent calculus course, I introduced the technique of Integration by Parts as an integration rule corresponding to the Product Rule for differentiation. I showed my students the standard derivation of the Integration by Parts formula as presented in [1]: By the Product Rule, if f (x) and g(x) are differentiable functions, then d f (x)g(x) = f (x)g(x) + g(x) f (x). dx Integrating on both sides of this equation, f (x)g(x) + g(x) f (x) dx = f (x)g(x), which may be rearranged to obtain f (x)g(x) dx = f (x)g(x) − g(x) f (x) dx. Letting U = f (x) and V = g(x) and observing that dU = f (x) dx and dV = g(x) dx, we obtain the familiar Integration by Parts formula UdV= UV − VdU. (1) My student Victor asked if we could do a similar thing with the Quotient Rule. While the other students thought this was a crazy idea, I was intrigued. Below, I derive a Quotient Rule Integration by Parts formula, apply the resulting integration formula to an example, and discuss reasons why this formula does not appear in calculus texts. By the Quotient Rule, if f (x) and g(x) are differentiable functions, then ( ) ( ) ( ) − ( ) ( ) d f x = g x f x f x g x . dx g(x) [g(x)]2 Integrating both sides of this equation, we get f (x) g(x) f (x) − f (x)g(x) = dx.
    [Show full text]
  • Laplace Transform
    Chapter 7 Laplace Transform The Laplace transform can be used to solve differential equations. Be- sides being a different and efficient alternative to variation of parame- ters and undetermined coefficients, the Laplace method is particularly advantageous for input terms that are piecewise-defined, periodic or im- pulsive. The direct Laplace transform or the Laplace integral of a function f(t) defined for 0 ≤ t< 1 is the ordinary calculus integration problem 1 f(t)e−stdt; Z0 succinctly denoted L(f(t)) in science and engineering literature. The L{notation recognizes that integration always proceeds over t = 0 to t = 1 and that the integral involves an integrator e−stdt instead of the usual dt. These minor differences distinguish Laplace integrals from the ordinary integrals found on the inside covers of calculus texts. 7.1 Introduction to the Laplace Method The foundation of Laplace theory is Lerch's cancellation law 1 −st 1 −st 0 y(t)e dt = 0 f(t)e dt implies y(t)= f(t); (1) R R or L(y(t)= L(f(t)) implies y(t)= f(t): In differential equation applications, y(t) is the sought-after unknown while f(t) is an explicit expression taken from integral tables. Below, we illustrate Laplace's method by solving the initial value prob- lem y0 = −1; y(0) = 0: The method obtains a relation L(y(t)) = L(−t), whence Lerch's cancel- lation law implies the solution is y(t)= −t. The Laplace method is advertised as a table lookup method, in which the solution y(t) to a differential equation is found by looking up the answer in a special integral table.
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • CHAPTER 3: Derivatives
    CHAPTER 3: Derivatives 3.1: Derivatives, Tangent Lines, and Rates of Change 3.2: Derivative Functions and Differentiability 3.3: Techniques of Differentiation 3.4: Derivatives of Trigonometric Functions 3.5: Differentials and Linearization of Functions 3.6: Chain Rule 3.7: Implicit Differentiation 3.8: Related Rates • Derivatives represent slopes of tangent lines and rates of change (such as velocity). • In this chapter, we will define derivatives and derivative functions using limits. • We will develop short cut techniques for finding derivatives. • Tangent lines correspond to local linear approximations of functions. • Implicit differentiation is a technique used in applied related rates problems. (Section 3.1: Derivatives, Tangent Lines, and Rates of Change) 3.1.1 SECTION 3.1: DERIVATIVES, TANGENT LINES, AND RATES OF CHANGE LEARNING OBJECTIVES • Relate difference quotients to slopes of secant lines and average rates of change. • Know, understand, and apply the Limit Definition of the Derivative at a Point. • Relate derivatives to slopes of tangent lines and instantaneous rates of change. • Relate opposite reciprocals of derivatives to slopes of normal lines. PART A: SECANT LINES • For now, assume that f is a polynomial function of x. (We will relax this assumption in Part B.) Assume that a is a constant. • Temporarily fix an arbitrary real value of x. (By “arbitrary,” we mean that any real value will do). Later, instead of thinking of x as a fixed (or single) value, we will think of it as a “moving” or “varying” variable that can take on different values. The secant line to the graph of f on the interval []a, x , where a < x , is the line that passes through the points a, fa and x, fx.
    [Show full text]
  • MATH 162: Calculus II Framework for Thurs., Mar
    MATH 162: Calculus II Framework for Thurs., Mar. 29–Fri. Mar. 30 The Gradient Vector Today’s Goal: To learn about the gradient vector ∇~ f and its uses, where f is a function of two or three variables. The Gradient Vector Suppose f is a differentiable function of two variables x and y with domain R, an open region of the xy-plane. Suppose also that r(t) = x(t)i + y(t)j, t ∈ I, (where I is some interval) is a differentiable vector function (parametrized curve) with (x(t), y(t)) being a point in R for each t ∈ I. Then by the chain rule, d ∂f dx ∂f dy f(x(t), y(t)) = + dt ∂x dt ∂y dt 0 0 = [fxi + fyj] · [x (t)i + y (t)j] dr = [fxi + fyj] · . (1) dt Definition: For a differentiable function f(x1, . , xn) of n variables, we define the gradient vector of f to be ∂f ∂f ∂f ∇~ f := , ,..., . ∂x1 ∂x2 ∂xn Remarks: • Using this definition, the total derivative df/dt calculated in (1) above may be written as df = ∇~ f · r0. dt In particular, if r(t) = x(t)i + y(t)j + z(t)k, t ∈ (a, b) is a differentiable vector function, and if f is a function of 3 variables which is differentiable at the point (x0, y0, z0), where x0 = x(t0), y0 = y(t0), and z0 = z(t0) for some t0 ∈ (a, b), then df ~ 0 = ∇f(x0, y0, z0) · r (t0). dt t=t0 • If f is a function of 2 variables, then ∇~ f has 2 components.
    [Show full text]
  • Vector Calculus and Differential Forms with Applications To
    Vector Calculus and Differential Forms with Applications to Electromagnetism Sean Roberson May 7, 2015 PREFACE This paper is written as a final project for a course in vector analysis, taught at Texas A&M University - San Antonio in the spring of 2015 as an independent study course. Students in mathematics, physics, engineering, and the sciences usually go through a sequence of three calculus courses before go- ing on to differential equations, real analysis, and linear algebra. In the third course, traditionally reserved for multivariable calculus, stu- dents usually learn how to differentiate functions of several variable and integrate over general domains in space. Very rarely, as was my case, will professors have time to cover the important integral theo- rems using vector functions: Green’s Theorem, Stokes’ Theorem, etc. In some universities, such as UCSD and Cornell, honors students are able to take an accelerated calculus sequence using the text Vector Cal- culus, Linear Algebra, and Differential Forms by John Hamal Hubbard and Barbara Burke Hubbard. Here, students learn multivariable cal- culus using linear algebra and real analysis, and then they generalize familiar integral theorems using the language of differential forms. This paper was written over the course of one semester, where the majority of the book was covered. Some details, such as orientation of manifolds, topology, and the foundation of the integral were skipped to save length. The paper should still be readable by a student with at least three semesters of calculus, one course in linear algebra, and one course in real analysis - all at the undergraduate level.
    [Show full text]
  • Part IA — Vector Calculus
    Part IA | Vector Calculus Based on lectures by B. Allanach Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures. They are nowhere near accurate representations of what was actually lectured, and in particular, all errors are almost surely mine. 3 Curves in R 3 Parameterised curves and arc length, tangents and normals to curves in R , the radius of curvature. [1] 2 3 Integration in R and R Line integrals. Surface and volume integrals: definitions, examples using Cartesian, cylindrical and spherical coordinates; change of variables. [4] Vector operators Directional derivatives. The gradient of a real-valued function: definition; interpretation as normal to level surfaces; examples including the use of cylindrical, spherical *and general orthogonal curvilinear* coordinates. Divergence, curl and r2 in Cartesian coordinates, examples; formulae for these oper- ators (statement only) in cylindrical, spherical *and general orthogonal curvilinear* coordinates. Solenoidal fields, irrotational fields and conservative fields; scalar potentials. Vector derivative identities. [5] Integration theorems Divergence theorem, Green's theorem, Stokes's theorem, Green's second theorem: statements; informal proofs; examples; application to fluid dynamics, and to electro- magnetism including statement of Maxwell's equations. [5] Laplace's equation 2 3 Laplace's equation in R and R : uniqueness theorem and maximum principle. Solution of Poisson's equation by Gauss's method (for spherical and cylindrical symmetry) and as an integral. [4] 3 Cartesian tensors in R Tensor transformation laws, addition, multiplication, contraction, with emphasis on tensors of second rank. Isotropic second and third rank tensors.
    [Show full text]
  • MATH 162: Calculus II Differentiation
    MATH 162: Calculus II Framework for Mon., Jan. 29 Review of Differentiation and Integration Differentiation Definition of derivative f 0(x): f(x + h) − f(x) f(y) − f(x) lim or lim . h→0 h y→x y − x Differentiation rules: 1. Sum/Difference rule: If f, g are differentiable at x0, then 0 0 0 (f ± g) (x0) = f (x0) ± g (x0). 2. Product rule: If f, g are differentiable at x0, then 0 0 0 (fg) (x0) = f (x0)g(x0) + f(x0)g (x0). 3. Quotient rule: If f, g are differentiable at x0, and g(x0) 6= 0, then 0 0 0 f f (x0)g(x0) − f(x0)g (x0) (x0) = 2 . g [g(x0)] 4. Chain rule: If g is differentiable at x0, and f is differentiable at g(x0), then 0 0 0 (f ◦ g) (x0) = f (g(x0))g (x0). This rule may also be expressed as dy dy du = . dx x=x0 du u=u(x0) dx x=x0 Implicit differentiation is a consequence of the chain rule. For instance, if y is really dependent upon x (i.e., y = y(x)), and if u = y3, then d du du dy d (y3) = = = (y3)y0(x) = 3y2y0. dx dx dy dx dy Practice: Find d x d √ d , (x2 y), and [y cos(xy)]. dx y dx dx MATH 162—Framework for Mon., Jan. 29 Review of Differentiation and Integration Integration The definite integral • the area problem • Riemann sums • definition Fundamental Theorem of Calculus: R x I: Suppose f is continuous on [a, b].
    [Show full text]
  • Chapter 3. Linearization and Gradient Equation Fx(X, Y) = Fxx(X, Y)
    Oliver Knill, Harvard Summer School, 2010 An equation for an unknown function f(x, y) which involves partial derivatives with respect to at least two variables is called a partial differential equation. If only the derivative with respect to one variable appears, it is called an ordinary differential equation. Examples of partial differential equations are the wave equation fxx(x, y) = fyy(x, y) and the heat Chapter 3. Linearization and Gradient equation fx(x, y) = fxx(x, y). An other example is the Laplace equation fxx + fyy = 0 or the advection equation ft = fx. Paul Dirac once said: ”A great deal of my work is just playing with equations and see- Section 3.1: Partial derivatives and partial differential equations ing what they give. I don’t suppose that applies so much to other physicists; I think it’s a peculiarity of myself that I like to play about with equations, just looking for beautiful ∂ mathematical relations If f(x, y) is a function of two variables, then ∂x f(x, y) is defined as the derivative of the function which maybe don’t have any physical meaning at all. Sometimes g(x) = f(x, y), where y is considered a constant. It is called partial derivative of f with they do.” Dirac discovered a PDE describing the electron which is consistent both with quan- respect to x. The partial derivative with respect to y is defined similarly. tum theory and special relativity. This won him the Nobel Prize in 1933. Dirac’s equation could have two solutions, one for an electron with positive energy, and one for an electron with ∂ antiparticle One also uses the short hand notation fx(x, y)= ∂x f(x, y).
    [Show full text]
  • Proofs of the Product, Reciprocal, and Quotient Rules Math 120 Calculus I D Joyce, Fall 2013
    Proofs of the Product, Reciprocal, and Quotient Rules Math 120 Calculus I D Joyce, Fall 2013 So far, we've defined derivatives in terms of limits f(x+h) − f(x) f 0(x) = lim ; h!0 h found derivatives of several functions; used and proved several rules including the constant rule, sum rule, difference rule, and constant multiple rule; and used the product, reciprocal, and quotient rules. Next, we'll prove those last three rules. After that, we still have to prove the power rule in general, there's the chain rule, and derivatives of trig functions. But then we'll be able to differentiate just about any function we can write down. The product, reciprocal, and quotient rules. For the statement of these three rules, let f and g be two differentiable functions. Then Product rule: (fg)0 = f 0g + fg0 10 −g0 Reciprocal rule: = g g2 f 0 f 0g − fg0 Quotient rule: = g g2 A more complete statement of the product rule would assume that f and g are differ- entiable at x and conlcude that fg is differentiable at x with the derivative (fg)0(x) equal to f 0(x)g(x) + f(x)g0(x). Likewise, the reciprocal and quotient rules could be stated more completely. A proof of the product rule. This proof is not simple like the proofs of the sum and difference rules. By the definition of derivative, (fg)(x+h) − (fg)(x) (fg)0(x) = lim : h!0 h Now, the expression (fg)(x) means f(x)g(x), therefore, the expression (fg)(x+h) means f(x+h)g(x+h).
    [Show full text]
  • Tensor Calculus
    A Primer on Tensor Calculus David A. Clarke Saint Mary’s University, Halifax NS, Canada [email protected] June, 2011 Copyright c David A. Clarke, 2011 Contents Preface ii 1 Introduction 1 2 Definition of a tensor 3 3 The metric 9 3.1 Physical components and basis vectors ..................... 11 3.2 The scalar and inner products .......................... 14 3.3 Invariance of tensor expressions ......................... 17 3.4 The permutation tensors ............................. 18 4 Tensor derivatives 21 4.1 “Christ-awful symbols” .............................. 21 4.2 Covariant derivative ............................... 25 5 Connexion to vector calculus 30 5.1 Gradient of a scalar ................................ 30 5.2 Divergence of a vector .............................. 30 5.3 Divergence of a tensor .............................. 32 5.4 The Laplacian of a scalar ............................. 33 5.5 Curl of a vector .................................. 34 5.6 The Laplacian of a vector ............................ 35 5.7 Gradient of a vector ............................... 35 5.8 Summary ..................................... 36 5.9 A tensor-vector identity ............................. 37 6 Cartesian, cylindrical, spherical polar coordinates 39 6.1 Cartesian coordinates ............................... 40 6.2 Cylindrical coordinates .............................. 40 6.3 Spherical polar coordinates ............................ 41 7 An application to viscosity 42 i Preface These notes stem from my own need to refresh my memory on the fundamentals
    [Show full text]
  • Calculus: Intuitive Explanations
    Calculus: Intuitive Explanations Radon Rosborough https://intuitiveexplanations.com/math/calculus-intuitive-explanations/ Intuitive Explanations is a collection of clear, easy-to-understand, intuitive explanations of many of the more complex concepts in calculus. In this document, an “intuitive” explanation is not one without any mathematical symbols. Rather, it is one in which each definition, concept, and part of an equation can be understood in a way that is more than “a piece of mathematics”. For instance, some textbooks simply define mathematical terms, whereas Intuitive Explanations explains what idea each term is meant to represent and why the definition has to be the way it is in order for that idea to be captured. As another example, the proofs in Intuitive Explanations avoid steps with no clear motivation (like adding and subtracting the same quantity), so that they seem natural and not magical. See the table of contents for information on what topics are covered. Before each section, there is a note clarifying what prerequisite knowledge about calculus is required to understand the material presented. Section 1 contains a number of links to other resources useful for learning calculus around the Web. I originally wrote Intuitive Explanations in high school, during the summer between AP Calculus BC and Calculus 3, and the topics covered here are, generally speaking, the topics that are covered poorly in the textbook we used (Calculus by Ron Larson and Bruce Edwards, tenth edition). Contents 1 About Intuitive Explanations ................................. 4 2 The Formal Definition of Limit ................................ 5 2.1 First Attempts ...................................... 6 2.2 Arbitrary Closeness .................................... 7 2.3 Using Symbols .....................................
    [Show full text]