Solution of Ordinary Differential Equations in Gradient-Based Multidisciplinary Design Optimization
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
An Enhanced Parareal Algorithm Based on the Deferred Correction Methods for a Stiff System✩
Journal of Computational and Applied Mathematics 255 (2014) 297–305 Contents lists available at SciVerse ScienceDirect Journal of Computational and Applied Mathematics journal homepage: www.elsevier.com/locate/cam An enhanced parareal algorithm based on the deferred correction methods for a stiff systemI Sunyoung Bu a,∗, June-Yub Lee a,b a Institute of Mathematical Sciences, Ewha Womans University, Seoul 120-750, Republic of Korea b Department of Mathematics, Ewha Womans University, Seoul 120-750, Republic of Korea article info a b s t r a c t Article history: In this study, we consider a variant of the hybrid parareal algorithm based on deferred Received 31 March 2012 correction techniques in order to increase the convergence order even for the stiff system. Received in revised form 4 April 2013 A hybrid parareal scheme introduced by Minion (2011) [20] improves the efficiency of the original parareal by utilizing a Spectral Deferred Correction (SDC) strategy for a fine Keywords: propagator within the parareal iterations. In this paper, we use Krylov Deferred Correction Hybrid parareal algorithm (KDC) for a fine propagator to solve the stiff system and Differential Algebraic Equations Spectral deferred correction (DAEs) stably. Also we employ a deferred correction technique based on the backward Euler Krylov deferred correction Stiff system method for a coarse propagator in order to make the global order of accuracy reasonably Differential algebraic equation high while limiting the cost of sequential steps as small as possible. Numerical experiments on the efficiency of our method are promising. ' 2013 Elsevier B.V. All rights reserved. 1. Introduction Deferred correction methods can be used to obtain high order numerical solutions for a system of time dependent differential equations in the form of 0 f .y.t/; y .t/; t/ D 0; (1) y.0/ D y0: Deferred correction methods have mainly been used for boundary value problems. -
Chapter 16: Differential Equations
✐ ✐ ✐ ✐ Chapter 16 Differential Equations A vast number of mathematical models in various areas of science and engineering involve differ- ential equations. This chapter provides a starting point for a journey into the branch of scientific computing that is concerned with the simulation of differential problems. We shall concentrate mostly on developing methods and concepts for solving initial value problems for ordinary differential equations: this is the simplest class (although, as you will see, it can be far from being simple), yet a very important one. It is the only class of differential problems for which, in our opinion, up-to-date numerical methods can be learned in an orderly and reasonably complete fashion within a first course text. Section 16.1 prepares the setting for the numerical treatment that follows in Sections 16.2–16.6 and beyond. It also contains a synopsis of what follows in this chapter. Numerical methods for solving boundary value problems for ordinary differential equations receive a quick review in Section 16.7. More fundamental difficulties arise here, so this section is marked as advanced. Most mathematical models that give rise to differential equations in practice involve partial differential equations, where there is more than one independent variable. The numerical treatment of partial differential equations is a vast and complex subject that relies directly on many of the methods introduced in various parts of this text. An orderly development belongs in a more advanced text, though, and our own description in Section 16.8 is downright anecdotal, relying in part on examples introduced earlier. 16.1 Initial value ordinary differential equations Consider the problem of finding a function y(t) that satisfies the ordinary differential equation (ODE) dy = f (t, y), a ≤ t ≤ b. -
Recent Advances in Computational and Applied Mathematics
Recent Advances in Computational and Applied Mathematics Theodore E. Simos Editor Recent Advances in Computational and Applied Mathematics Editor Theodore E. Simos Department of Mathematics College of Sciences King Saud University P.O. Box 2455 Riyadh 11451 Saudi Arabia and Laboratory of Computational Sciences Department of Computer Science and Technology University of Peloponnese 22100 Tripolis Greece [email protected] ISBN 978-90-481-9980-8 e-ISBN 978-90-481-9981-5 DOI 10.1007/978-90-481-9981-5 Springer Dordrecht Heidelberg London New York Library of Congress Control Number: 2010938640 Mathematics Subject Classification (2000): 65, 65D20, 65Rxx, 65Lxx, 30, 30G35, 31B05, 31B35, 31, 33C05, 33C10, 33C15, 35A08, 35G15, 35Q40, 35R01, 41A60, 49Jxx, 49Nxx, 76-xx, 76A05, 81Q05, 81Qxx, 93Exx © Springer Science+Business Media B.V. 2011 Chapter 10 was created within the capacity of an US governmental employment and therefore is in the public domain. No part of this work may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, microfilming, recording or otherwise, without written permission from the Publisher, with the exception of any material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. Cover design: WMXDesign Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com) Preface This volume includes an exciting collection of papers on computational and applied mathematics presenting the recent advances in several areas of this field. All the papers have been peer reviewed by at least three reviewers. -
Numerical Methods (Wiki) GU Course Work
Numerical Methods (Wiki) GU course work PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sun, 21 Sep 2014 15:07:23 UTC Contents Articles Numerical methods for ordinary differential equations 1 Runge–Kutta methods 7 Euler method 18 Numerical integration 25 Trapezoidal rule 32 Simpson's rule 35 Bisection method 40 Newton's method 44 Root-finding algorithm 57 Monte Carlo method 62 Interpolation 73 Lagrange polynomial 78 References Article Sources and Contributors 84 Image Sources, Licenses and Contributors 86 Article Licenses License 87 Numerical methods for ordinary differential equations 1 Numerical methods for ordinary differential equations Numerical methods for ordinary differential equations are methods used to find numerical approximations to the solutions of ordinary differential equations (ODEs). Their use is also known as "numerical integration", although this term is sometimes taken to mean the computation of integrals. Many differential equations cannot be solved using symbolic computation ("analysis"). For practical purposes, however – such as in engineering – a numeric approximation to the solution is often sufficient. The algorithms studied here can be used to compute such an approximation. An alternative method is to use techniques from calculus to obtain a series expansion of the solution. Ordinary differential equations occur in many scientific disciplines, for instance in physics, chemistry, biology, and economics. In addition, some methods in numerical partial differential equations convert the partial differential equation into an ordinary differential equation, which must then be solved. Illustration of numerical integration for the differential equation The problem Blue: the Euler method, green: the midpoint method, red: the exact solution, The A first-order differential equation is an Initial value problem (IVP) of step size is the form, where f is a function that maps [t ,∞) × Rd to Rd, and the initial 0 condition y ∈ Rd is a given vector. -
An Analysis of the TR-BDF2 Integration Scheme Sohan
An Analysis of the TR-BDF2 integration scheme by Sohan Dharmaraja Submitted to the School of Engineering in Partial Fulfillment of the Requirements for the degree of Master of Science in Computation for Design and Optimization at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY September 2007 @ Massachusetts Institute of Technology 2007. All rights reserved. Author ............................ School of Engineering July 13, 2007 Certified by ...................... ...................... W. Gilbert Strang Professor of Mathematics Thesis Supervisor Accepted by I \\ Jaume Peraire Professor of Aeronautics and Astronautics MASSACHUSETTS INSTrTUTF Co-Director, Computation for Design and Optimization OF TECHNOLOGY SEP 2 7 2007 BARKER LIBRARIES 2 An Analysis of the TR-BDF2 integration scheme by Sohan Dharmaraja Submitted to the School of Engineering on July 13, 2007, in partial fulfillment of the requirements for the degree of Master of Science in Computation for Design and Optimization Abstract We intend to try to better our understanding of how the combined L-stable 'Trape- zoidal Rule with the second order Backward Difference Formula' (TR-BDF2) integra- tor and the standard A-stable Trapezoidal integrator perform on systems of coupled non-linear partial differential equations (PDEs). It was originally Professor Klaus- Jiirgen Bathe who suggested that further analysis was needed in this area. We draw attention to numerical instabilities that arise due to insufficient numerical damp- ing from the Crank-Nicolson method (which is based on the Trapezoidal rule) and demonstrate how these problems can be rectified with the TR-BDF2 scheme. Several examples are presented, including an advection-diffusion-reaction (ADR) problem and the (chaotic) damped driven pendulum. We also briefly introduce how the ideas of splitting methods can be coupled with the TR-BDF2 scheme and applied to the ADR equation to take advantage of the excellent modern day explicit techniques to solve hyperbolic equations. -
A Brief Note on Linear Multistep Methods
A brief note on linear multistep methods Alen Alexanderian* Abstract We review some key results regarding linear multistep methods for solving intial value prob- lems. Moreover, we discuss some common classes of linear multistep methods. 1 Basic theory Here we provide a summary of key results regarding stability and consistency of linear mul- tistep methods. We focus on the initial value problem, 0 y = f(x; y); x 2 [a; b]; y(a) = y0: (1.1) Here y(x) 2 R is the state variable, and all throughout, we assume that f satisfies a uniform Lipschitz condition with respect to y. Definition of a multistep method. The generic description of a linear multistep method is as follows: define nodes xn = a + nh, n = 0;:::;N, with h = (b − a)=N. The general form of a multistep method is p p X X yn+1 = αjyn−j + h βjf(xn−j; yn−j); n ≥ p: (1.2) j=0 j=−1 p p Here fαigi=0, and fβigi=−1 are constant coefficients, and p ≥ 0. If either αp 6= 0 or βp 6= 0, then the method is called a p + 1 step method. The initial values, y0; y1; : : : ; yp, must be obtained by other means (e.g., an appropriate explicit method). Note that if β−1 = 0, then the method is explicit, and if β−1 6= 0, the method is implicit. Also, here we use the convention that y(xn) is the exact value of y at xn and yn is the numerically computed approximation to y(xn); i.e., yn ≈ y(xn). -
Wiki: Linear Multistep Method "Adams Method" Redirects Here. for the Electoral Apportionment Method, See Method of Smallest Divisors
Wiki: Linear multistep method "Adams method" redirects here. For the electoral apportionment method, see Method of smallest divisors. Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The process continues with subsequent steps to map out the solution. Single-step methods (such as Euler's method) refer to only one previous point and its derivative to determine the current value. Methods such as Runge-Kutta take some intermediate steps (for example, a half-step) to obtain a higher order method, but then discard all previous information before taking a second step. Multistep methods attempt to gain efficiency by keeping and using the information from previous steps rather than discarding it. Consequently, multistep methods refer to several previous points and derivative values. In the case of linear multistep methods, a linear combination of the previous points and derivative values is used. Contents: 1. Definitions 2. Examples 3. Multistep Method Families 4. Analysis 5. First and second Dahlquist barriers 6. See also 7. References 8. External links 1. Definitions Numerical methods for ordinary differential equations approximate solutions to initial value problems of the form The result is approximations for the value of at discrete times : where h is the time step (sometimes referred to as ). A linear multistep method uses a linear combination of and to calculate the value of y for the desired current step. Multistep method will use the previous s steps to calculate the next value. -
Lecture 1 Introduction
Lecture 1 Introduction The great majority of di↵erential equations which describe real systems cannot be solved analytically, the equations can only be solved approximately using numerical algorithms. It is also the case that a great many problems are set in an evolutionary framework: given the state at one time, determine the state at some future time. In this course we try to set out the methodology whereby the behaviour of numerical algorithms to approximate such di↵erential systems can be studied systematically and the advantages, accuracy and pitfalls of such algorithms can be understood. Many systems, such as used in aircraft control, power station control, guidance systems are operated without human intervention according to computed solutions of di↵erential systems, so we have great interest in knowing that solutions are computed accurately and efficiently. Some examples of real systems are easy to find. 1. Motion under a central force field There are many examples in molecular or stellar dynamics where motion is determined by a central force field. The general title for this is an N-body problem. The example you are most likely to have already seen is that of just two bodies moving in a plane with a gravitational field. This simplifies to a one-body problem if one of the bodies has mass (M)sufficientlymassivethatitspositioncanbetakenasfixedatanorigin and the position of the second body, of mass m be given by (ˆx(tˆ), yˆ(tˆ)) with velocities (ˆu(tˆ), vˆ(tˆ)) as unknowns. Then, in a gravitational problem, applying simple mechanics shows that the motion will be described by d2xˆ GMm m = x,ˆ dtˆ2 − rˆ3 d2yˆ GMm m = y,ˆ dtˆ2 − rˆ3 2 2 2 11 3 1 2 wherer ˆ =ˆx +ˆy and G is a universal constant, G =6.67 10− m kg− s− .