Implicit Differentiation

Total Page:16

File Type:pdf, Size:1020Kb

Implicit Differentiation opes. A Neurips 2020 Tut orial by Marc Deisenro A Swamps of integrationnumerical b Monte Carlo Heights C Fo r w w m a r o al r N izi ng Flo d -ba h ckw ime ar Of T d D B ea c h Hills There E Unrolling Backprop Bay f Expectations and Sl and fExpectations B ack Again: A Tale o Tale A Again: ack and th and opes. A Neurips 2020 Tut orial by Marc Deisenro A Swamps of integrationnumerical b Monte Carlo Heights C Fo r w Implicit of Vale w m a r o al r N izi ng Flo d -ba h G ckw ime ar Of T d D B Differentiation ea c h Hills There E Unrolling Backprop Bay f Expectations and Sl and fExpectations B ack Again: A Tale o Tale A Again: ack and th and Implicit differentiation Cheng Soon Ong Marc Peter Deisenroth December 2020 oon n S O g g ’s n Tu e t hC o d r n C i a a h l t M a o t a M r r n c e D e s i Motivation I In machine learning, we use gradients to train predictors I For functions f(x) we can directly obtain its gradient rf(x) I How to represent a constraint? G(x; y) = 0 E.g. conservation of mass 1 dy = 3x2 + 4x + 1 dx Observe that we can write the equation as x3 + 2x2 + x + 4 − y = 0 which is of the form G(x; y) = 0. High school gradients This is an equation y = x3 + 2x2 + x + 4 dy What is the gradient dx ? 2 Observe that we can write the equation as x3 + 2x2 + x + 4 − y = 0 which is of the form G(x; y) = 0. High school gradients This is an equation y = x3 + 2x2 + x + 4 dy What is the gradient dx ? dy = 3x2 + 4x + 1 dx 2 High school gradients This is an equation y = x3 + 2x2 + x + 4 dy What is the gradient dx ? dy = 3x2 + 4x + 1 dx Observe that we can write the equation as x3 + 2x2 + x + 4 − y = 0 which is of the form G(x; y) = 0. 2 Optimization with constraints Given a constrainted continuous optimization problem minx;yF (x; y) subject to G(x; y) = 0 We can solve the equality constraint to get y = g(x) and substitute into the objective F (x; g(x)), and calculate the gradient d F (x; g(x)) dx 3 Implicit function theorem I Solve for y = g(x) and use quotient rule I Directly differentiate, and use product rule Advanced high school gradients This is an equation y = x3 + 2x2 + xy + 4 dy What is the gradient dx ? 4 Advanced high school gradients This is an equation y = x3 + 2x2 + xy + 4 dy What is the gradient dx ? Implicit function theorem I Solve for y = g(x) and use quotient rule I Directly differentiate, and use product rule 4 Function of two variables Consider a function G(x; y) = 0 where x; y 2 R. Assume that near a particular point x0, we can write a closed form expression for y in terms of x, that is y = g(x) : 5 Solve for one variable Substituting y = g(x) into G(x; y), near x0, we get G(x; g(x)) = 0 : We calculate the derivative of G with respect to x using the chain rule, 0 Gx(x; g(x)) + Gy(x; g(x)) · g (x) = 0 : If Gy =6 0, then G g0(x) = − x : Gy 6 Implicit function theorem The implicit function theorem provides conditions under which we can write G(x; y) = 0 as y = g(x), and also conditions when g0(x) = − Gx . Gy Hand wavy intuition Given three variables (could be any topology) x; y; z and G(x; y) = z ; What are the conditions for the following to be well behaved? g(x) = y G(x; g(x)) = z : 7 Implicit function theorem as ansatz Ansatz 1. Explicitly solve one variable in terms of another 2. Chain the gradients I Ansatz = a way to look at problems I The implicit function theorem is a way to solve equations Generalizations depending on where equations live (Krantz and Parks, 2013): I Inverse function theorem I Constant rank theorem I Banach fixed point theorem I Nash-Moser theorem 8 Pointers to literature I From linear algebra to calculus (Hubbard and Hubbard, 2015; Spivak, 2008) I Implicit function theorem from variational analysis (Dontchev and Rockafellar, 2014) I Many versions of implicit function theorem (Krantz and Parks, 2013) I Deep Declarative Networks https://anucvml.github.io/ddn-cvprw2020/ Augustin-Louis Cauchy, 1916 9 Gradient of loss w.r.t. optimal value I Consider the problem of structured prediction (Nowozin et al., 2014) I Let a sample be (xn; yn) I Energy given the parameter of the learner w is E(yn; xn; w) I We assume that the best predictor is found by an optimization algorithm over y, ∗ y (xn; w) = optyE(y; xn; w) : I Measure the error with `(y; yn) I ∗ Loss per sample L(xn; yn; w) (assume predict with y ) I Want to take the gradient of loss L w.r.t. paramters w I but have an optimization problem inside (to find y∗) 10 Gradient of L w.r.t. w By implicit differentation (Do et al., 2007; Samuel and Tappen, 2009), the gradient of the dL loss L(xn; yn; w) with respect to the parameters dw has a closed form. Theorem ∗ ∗ Let y (w) = argminy E(y; w), and L(w) = `(y (w)). Then ( )− dL @2E @2E 1 d` = − : dw @w@y> @y@y> dy 11 Sketch: chain rule ( ) L(w) = ` optyE(y; w) By the chain rule, @L @` @y = : @w @y @w 12 Sketch: gradient with respect to energy E Denote by @E(y; w) g(y; w) = @y The gradient of g with respect to w is given by the chain rule again. Note that y is actually a function of w, i.e. g(y(w); w). @ @g @y @g g(y(w); w) = + : @w @y @w @w 13 Sketch: Stationarity conditions At optimality of E(y; w), its gradient is zero, i.e. g(y; w) = 0. Solving ( )− @y @g @g 1 = − @w @w @y @y @L Substituting @w into @w , ( )− @L @g @g 1 @` = − : @w @w @y @y 14 @L Sketch: Substitute back to obtain @w Recall the definition of g as gradient of energy E, @E(y; w) g(y; w) = @y And hence we get the Hessian with respect to the energy ( )− @L @2E @2E 1 d` = − : @w @w@y> @y@y> dy 15 Result: Gradient of L w.r.t. w The gradient of L(w) = `(argminy E(y(w); w)) with respect to w has a closed form. Theorem ∗ ∗ Let y (w) = argminy E(y; w), and L(w) = `(y (w)). Then ( )− dL @2E @2E 1 d` = − : dw @w@y> @y@y> dy Domke (2012) 16 Summary I We want to take the gradient with respect to an equality constraint I Implicit function theorem gives conditions where we can "invert" a derivative I Implicit function theorem as a way to solve equations I Useful to take a gradient over an optimum Backpropagation is just... the implicit function theorem 17 References Do, C. B., Foo, C.-S., and Ng, A. (2007). Efficient multiple hyperparameter learning for log-linear models. In Neural Information Processing Systems. Domke, J. (2012). Generic methods for optimization-based modeling. In AISTATS. Dontchev, A. L. and Rockafellar, R. T. (2014). Implicit Functions and Solution Mappings: A View from Variational Analysis. Springer, 2nd edition. Hubbard, J. H. and Hubbard, B. B. (2015). Vector Calculus, Linear Algebra, and Differential Forms: A Unified Approach. Matrix Editions, 5th edition. Krantz, S. G. and Parks, H. R. (2013). The Implict Function Theorem: History, Theory, and Applications. Springer. Nowozin, S., Gehler, P. V., Jancsary, J., and Lampert, C. H. (2014). Advanced Structured Prediction. MIT Press. 18 References (cont.) Samuel, K. and Tappen, M. (2009). Learning optimized map estimates in continuously-valued MRF models. In CVPR. Spivak, M. (2008). Calculus. Publish or Perish, Inc., 4th edition. 19.
Recommended publications
  • 33 .1 Implicit Differentiation 33.1.1Implicit Function
    Module 11 : Partial derivatives, Chain rules, Implicit differentiation, Gradient, Directional derivatives Lecture 33 : Implicit differentiation [Section 33.1] Objectives In this section you will learn the following : The concept of implicit differentiation of functions. 33 .1 Implicit differentiation \ As we had observed in section , many a times a function of an independent variable is not given explicitly, but implicitly by a relation. In section we had also mentioned about the implicit function theorem. We state it precisely now, without proof. 33.1.1Implicit Function Theorem (IFT): Let and be such that the following holds: (i) Both the partial derivatives and exist and are continuous in . (ii) and . Then there exists some and a function such that is differentiable, its derivative is continuous with and 33.1.2Remark: We have a corresponding version of the IFT for solving in terms of . Here, the hypothesis would be . 33.1.3Example: Let we want to know, when does the implicit expression defines explicitly as a function of . We note that and are both continuous. Since for the points and the implicit function theorem is not applicable. For , and , the equation defines the explicit function and for , the equation defines the explicit function Figure 1. y is a function of x. A result similar to that of theorem holds for function of three variables, as stated next. Theorem : 33.1.4 Let and be such that (i) exist and are continuous at . (ii) and . Then the equation determines a unique function in the neighborhood of such that for , and , Practice Exercises : Show that the following functions satisfy conditions of the implicit function theorem in the neighborhood of (1) the indicated point.
    [Show full text]
  • Integrated Calculus/Pre-Calculus
    Furman University Department of Mathematics Greenville, SC, 29613 Integrated Calculus/Pre-Calculus Abstract Mark R. Woodard Contents JJ II J I Home Page Go Back Close Quit Abstract This work presents the traditional material of calculus I with some of the material from a traditional precalculus course interwoven throughout the discussion. Pre- calculus topics are discussed at or soon before the time they are needed, in order to facilitate the learning of the calculus material. Miniature animated demonstra- tions and interactive quizzes will be available to help the reader deepen his or her understanding of the material under discussion. Integrated This project is funded in part by the Mellon Foundation through the Mellon Furman- Calculus/Pre-Calculus Wofford Program. Many of the illustrations were designed by Furman undergraduate Mark R. Woodard student Brian Wagner using the PSTricks package. Thanks go to my wife Suzan, and my two daughters, Hannah and Darby for patience while this project was being produced. Title Page Contents JJ II J I Go Back Close Quit Page 2 of 191 Contents 1 Functions and their Properties 11 1.1 Functions & Cartesian Coordinates .................. 13 1.1.1 Functions and Functional Notation ............ 13 1.2 Cartesian Coordinates and the Graphical Representa- tion of Functions .................................. 20 1.3 Circles, Distances, Completing the Square ............ 22 1.3.1 Distance Formula and Circles ................. 22 Integrated 1.3.2 Completing the Square ...................... 23 Calculus/Pre-Calculus 1.4 Lines ............................................ 24 Mark R. Woodard 1.4.1 General Equation and Slope-Intercept Form .... 24 1.4.2 More On Slope ............................. 24 1.4.3 Parallel and Perpendicular Lines .............
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • Notes on Calculus and Optimization
    Economics 101A Section Notes GSI: David Albouy Notes on Calculus and Optimization 1 Basic Calculus 1.1 Definition of a Derivative Let f (x) be some function of x, then the derivative of f, if it exists, is given by the following limit df (x) f (x + h) f (x) = lim − (Definition of Derivative) dx h 0 h → df although often this definition is hard to apply directly. It is common to write f 0 (x),or dx to be shorter, or dy if y = f (x) then dx for the derivative of y with respect to x. 1.2 Calculus Rules Here are some handy formulas which can be derived using the definitions, where f (x) and g (x) are functions of x and k is a constant which does not depend on x. d df (x) [kf (x)] = k (Constant Rule) dx dx d df (x) df (x) [f (x)+g (x)] = + (Sum Rule) dx dx dy d df (x) dg (x) [f (x) g (x)] = g (x)+f (x) (Product Rule) dx dx dx d f (x) df (x) g (x) dg(x) f (x) = dx − dx (Quotient Rule) dx g (x) [g (x)]2 · ¸ d f [g (x)] dg (x) f [g (x)] = (Chain Rule) dx dg dx For specific forms of f the following rules are useful in economics d xk = kxk 1 (Power Rule) dx − d ex = ex (Exponent Rule) dx d 1 ln x = (Logarithm Rule) dx x 1 Finally assuming that we can invert y = f (x) by solving for x in terms of y so that x = f − (y) then the following rule applies 1 df − (y) 1 = 1 (Inverse Rule) dy df (f − (y)) dx Example 1 Let y = f (x)=ex/2, then using the exponent rule and the chain rule, where g (x)=x/2,we df (x) d x/2 d x/2 d x x/2 1 1 x/2 get dx = dx e = d(x/2) e dx 2 = e 2 = 2 e .
    [Show full text]
  • Introduction to Manifolds
    Introduction to Manifolds Roland van der Veen 2018 2 Contents 1 Introduction 5 1.1 Overview . 5 2 How to solve equations? 7 2.1 Linear algebra . 9 2.2 Derivative . 10 2.3 Intermediate and mean value theorems . 13 2.4 Implicit function theorem . 15 3 Is there a fundamental theorem of calculus in higher dimensions? 23 3.1 Elementary Riemann integration . 24 3.2 k-vectors and k-covectors . 26 3.3 (co)-vector fields and integration . 33 3.3.1 Integration . 34 3.4 More on cubes and their boundary . 37 3.5 Exterior derivative . 38 3.6 The fundamental theorem of calculus (Stokes Theorem) . 40 3.7 Fundamental theorem of calculus: Poincar´elemma . 42 4 Geometry through the dot product 45 4.1 Vector spaces with a scalar product . 45 4.2 Riemannian geometry . 47 5 What if there is no good choice of coordinates? 51 5.1 Atlasses and manifolds . 51 5.2 Examples of manifolds . 55 5.3 Analytic continuation . 56 5.4 Bump functions and partitions of unity . 58 5.5 Vector bundles . 59 5.6 The fundamental theorem of calculus on manifolds . 62 5.7 Geometry on manifolds . 63 3 4 CONTENTS Chapter 1 Introduction 1.1 Overview The goal of these notes is to explore the notions of differentiation and integration in a setting where there are no preferred coordinates. Manifolds provide such a setting. This is not calculus. We made an attempt to prove everything we say so that no black boxes have to be accepted on faith. This self-sufficiency is one of the great strengths of mathematics.
    [Show full text]
  • Implicit Functions and Their Differentials in General Analysis*
    IMPLICIT FUNCTIONS AND THEIR DIFFERENTIALS IN GENERAL ANALYSIS* BY T. H. HILDEBRANDT AND LAWRENCEM. GRAVESf Introduction. Implicit function theorems occur in analysis in many different forms, and have a fundamental importance. Besides the classical theorems as given for example in Goursat's Cours d'Analyse and Bliss's Princeton Colloquium Lectures, and the classical theorems on linear integral equations, implicit function theorems in the domain of infinitely many variables have been developed by Volterra, Evans, Levy, W. L. Hart and others. Î The existence and imbedding theorems for solutions of differential equa- tions, as treated for example by Bliss, § have also received extensions to domains of infinitely many variables by Moulton, Hart, Barnett, Bliss and others. If Special properties in the case of linear differential equations in an infinitude of variables have been treated by Hart and Hildebrandt. || On the other hand, Hahn and Carathéodory** have made important generalizations of the notion of differential equation by removing conti- nuity restrictions on the derivatives and by writing the equations in the form of integral equations. *Presented to the Society March 25, 1921, and December 29, 1924; received by the editors January 28,1926. jNational Research Fellow in Mathematics. ÎSee Volterra, Fonctions de Lignes, Chapter 4; Evans, Cambridge Colloquium Lectures, pp. 52-72; Levy, Bulletin de la Société Mathématique de France, vol. 48 (1920), p. 13, Hart, these Transactions, vol. 18 (1917), pp. 125-150, where additional references may be found; also vol. 23 (1922), pp. 45-50, and Annals of Mathematics, (2), vol. 24 (1922)pp. 29 and 35. § Princeton Colloquium, and Bulletin of the American Mathematical Society, vol.
    [Show full text]
  • February 18, 2019 LECTURE 7: DIRECTIONAL DERIVATIVES
    February 18, 2019 LECTURE 7: DIRECTIONAL DERIVATIVES. 110.211 HONORS MULTIVARIABLE CALCULUS PROFESSOR RICHARD BROWN Synopsis. Today, we move into directional derivatives, a generalization of a partial deriva- tive where we look for how a function is changing at a point in any single direction in the domain. This gives a powerful tool, both conceptually as well as technically, to discuss the role the derivative of a function plays in exposing the properties of both functions on and sets within Euclidean space. We define the gradient of a real-valued function (finally) and its interpretations and usefulness, and move toward one of the most powerful theorems of multivariable calculus, the Implicit Function Theorem. The Directional Derivative. 7.0.1. Vector form of a partial derivative. Recall the definition of a partial derivative evalu- ated at a point: Let f : X ⊂ R2 ! R, x open, and (a; b) 2 X. Then the partial derivative of f with respect to the first coordinate x, evaluated at (a; b) is @f f(a + h; b) − f(a; b) (a; b) = lim : @x h!0 h Here, we vary only the first coordinate, leaving the y coordinate value b fixed, and write a (a + h; b) = (a; b) + (h; 0). In vector notation, this is like taking the vector a = , and b adding to it a small amount h, but only in the x-direction. Indeed, this means adding to a 1 the vector h = hi, where here we use the standard convention for unit vectors in 2 0 R 3 and R , namely i = e1, j = e2, etc.
    [Show full text]
  • Uniqueness, Continuity, and Existence of Implicit Functions in Constructive Analysis
    Uniqueness, Continuity, and Existence of Implicit Functions in Constructive Analysis Hannes Diener1 and Peter Schuster2 1 Universit¨atSiegen, Fachbereich 6: Mathematik Walter-Flex-Straße 3, 57072 Siegen, Germany [email protected] 2 Dipartimento di Filosofia, Universit`adegli Studi di Firenze Via Bolognese 52, 50139 Firenze, Italy [email protected] Abstract. We extract a quantitative variant of uniqueness from the usual hypotheses of the implicit functions theorem. This leads not only to an a priori proof of continuity, but also to an alternative, fully constructive existence proof. 1 Introduction To show the differentiability of an implicit function one often relies on its continu- ity. The latter is mostly seen as a by-product of the not uncommon construction of the implicit function as the limit of a uniformly convergent sequence of con- tinuous functions. We now show that the continuity of the implicit function is prior to its existence, and thus independent of any particular construction. More specifically, we deduce the continuity from a quantitative strengthening of the uniqueness, which in turn follows from the hypotheses one needs to impose on the equation the implicit function is expected to satisfy. The same quantitative strengthening of uniqueness enables us to ultimately give an alternative existence proof for implicit functions that is fully constructive in Bishop's sense. We use ideas from [6], which loc.cit. have only been spelled out in the case of implicit functions with values in R. The existence proof given in [6] therefore can rely on reasoning by monotonicity, whereas in the general case|treated in this paper|of implicit functions with values in Rm we need to employ an extreme value argument.
    [Show full text]
  • Chapter 6 Implicit Function Theorem
    Implicit function theorem 1 Chapter 6 Implicit function theorem Chapter 5 has introduced us to the concept of manifolds of dimension m contained in Rn. In the present chapter we are going to give the exact de¯nition of such manifolds and also discuss the crucial theorem of the beginnings of this subject. The name of this theorem is the title of this chapter. We de¯nitely want to maintain the following point of view. An m-dimensional manifold M ½ Rn is an object which exists and has various geometric and calculus properties which are inherent to the manifold, and which should not depend on the particular mathematical formulation we use in describing the manifold. Since our goal is to do lots of calculus on M, we need to have formulas we can use in order to do this sort of work. In the very discussion of these methods we shall gain a clear and precise understanding of what a manifold actually is. We have already done this sort of work in the preceding chapter in the case of hypermani- folds. There we discussed the intrinsic gradient and the fact that the tangent space at a point of such a manifold has dimension n ¡ 1 etc. We also discussed the version of the implicit function theorem that we needed for the discussion of hypermanifolds. We noticed at that time that we were really always working with only the local description of M, and that we didn't particularly care whether we were able to describe M with formulas that held for the entire manifold.
    [Show full text]
  • 13. Implicit Functions Consider the Curve Y2 = X in the Plane R2 , 2 2 C = { (X, Y) ∈ R | Y = X }
    13. Implicit functions Consider the curve y2 = x in the plane R2 , 2 2 C = f (x; y) 2 R j y = x g: This is not the graph of a function, and yet it is quite close to the graph of a function. Given any point on the graph, let's say the point (2; 4), we can always find open intervals U containing 2 and V containing 4 and a smooth function f : U −! V such that C \ (U × V ) is the graphp of f. Indeed, take U = (0; 1), V = (0; 1) and f(x) = x. In fact, we can do this for any point on the graph, apart from the origin. If it is above the x-axis, the function above works. If the point we pare interested in pis below the x-axis, replace V by (0; −∞) and f(x) = x, by g(x) = − x. How can we tell that the origin is a point where we cannot define an implicit function? Well away from the origin, the tangent line is not vertical but at the origin the tangent line is vertical. In other words, if we consider 2 F : R −! R; given by F (x; y) = y2 − x, so that C is the set of points where F is zero, then DF (x; y) = (−1; 2y): The locus where we run into trouble, is where 2y = 0. Somewhat amazingly this works in general: Theorem 13.1 (Implicit Function Theorem). Let A ⊂ Rn+m be an open subset and let F : A −! Rm be a C1-function.
    [Show full text]
  • Question 1. Calculate ∫ Sin -1 X Dx. Note. This Is an Example of A
    EXAMPLE: SECTION 7.1: INTEGRATION BY PARTS (THE ANTI-PRODUCT RULE) 110.109 CALCULUS II (PHYS SCI & ENG) PROFESSOR RICHARD BROWN Z Question 1. Calculate sin−1 x dx. Note. This is an example of a integral does not have the form of a product of functions. However, by thinking of one of the functions as a constant 1, the Integration by Parts procedure allows us to instead integrate the derivative of the inverse sine function. This will be easier. One other point is that this function sin−1 x is NOT \1 over the sine function", or 1 sin−1 x =6 = sec x: sin x This is a common mistake due to ambiguous notation. Rather, it is the function-inverse of the sine function, as in y = sin−1 x is the inverse function of sin y = x. To solve this, we will first compute the derivative of sin−1 x. Then we will use this fact. Proof. To compute the derivative of the function y = cos−1 x, we will instead use the inverse of this function sin y = x and use implicit differentiation. Here then thinking of y as an implicit function of x, we have d d (sin y) = (x) dx dx dy (cos y) = 1 dx dy 1 = : dx cos y This is not quite complete, as we have only expressed the derivative of y = sin−1 x in terms of the function y. We can fix this through an understanding of how triangles relate the trigonometric functions of their angles to their side lengths and such.
    [Show full text]
  • High-Order Quadrature Methods for Implicitly Defined Surfaces and Volumes in Hyperrectangles∗
    SIAM J. SCI. COMPUT. c 2015 Society for Industrial and Applied Mathematics Vol. 37, No. 2, pp. A993–A1019 HIGH-ORDER QUADRATURE METHODS FOR IMPLICITLY DEFINED SURFACES AND VOLUMES IN HYPERRECTANGLES∗ R. I. SAYE† Abstract. A high-order accurate numerical quadrature algorithm is presented for the evaluation of integrals over curved surfaces and volumes which are defined implicitly via a fixed isosurface of a given function restricted to a given hyperrectangle. By converting the implicitly defined geometry into the graph of an implicitly defined height function, the approach leads to a recursive algorithm on the number of spatial dimensions which requires only one-dimensional root finding and one-dimensional Gaussian quadrature. The computed quadrature scheme yields strictly positive quadrature weights and inherits the high-order accuracy of Gaussian quadrature: a range of different convergence tests demonstrate orders of accuracy up to 20th order. Also presented is an application of the quadrature algorithm to a high-order embedded boundary discontinuous Galerkin method for solving partial differential equations on curved domains. Key words. quadrature, integration, implicit surfaces, level set function, level set methods, high order AMS subject classifications. 65D30, 65N30 DOI. 10.1137/140966290 1. Introduction. In this paper, we develop high-order accurate numerical quadrature methods for the evaluation of integrals over curved surfaces and volumes whose geometry is defined implicitly via a fixed isosurface/level set of a smooth func- tion φ : Rd → R. In particular, assuming that the zero level set of φ defines the geome- try, let Γ = {x : φ(x)=0} denote the surface (“interface”) and let Ω = {x : φ(x) < 0} be the region on the negative side of the surface.
    [Show full text]