Constrained Optimization

Total Page:16

File Type:pdf, Size:1020Kb

Constrained Optimization AA222: MDO 113 Thursday 26th April, 2012 at 16:05 Chapter 5 Constrained Optimization Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest in general nonlinearly constrained optimization theory and methods in this chapter. Recall the statement of a general optimization problem, minimize f(x) (5.1) n with respect to x 2 R (5.2) subject toc ^j(x) = 0; j = 1;:::; m^ (5.3) ck(x) ≥ 0; k = 1; : : : ; m (5.4) Example 5.21. Graphical Solution of a Constrained Optimization Problem Suppose we want to solve the following optimization problem, 2 minimize f(x) = 4x1 − x1 − x2 − 2:5 (5.5) with respect to x1; x2 (5.6) 2 2 subject to c1(x) = x2 − 1:5x1 + 2x1 − 1 ≥ 0; (5.7) 2 2 c2(x) = x2 + 2x1 − 2x1 − 4:25 ≤ 0 (5.8) How can we solve this? One intuitive way would be to make a contour plot of the objective function, overlay the constraints and determine the feasible region, as shown in Fig. 5.1. By inspection, it is easy to see where the constrained optimum is. At the optimum, only one of the constraints is active. However, we want to be able to find the minimum numerically, without having to plot the functions, since this is impractical in the general case. We can see that in this case, the feasible space comprises two disconnected regions. In addition, although all the functions involved are smooth, the boundaries of the feasible regions are nonmsooth. These characteristics complicate the numerical solution of constrained optimization problems. 5.1 Optimality Conditions for Constrained Problems The optimality conditions for nonlinearly constrained problems are important because they form the basis for algorithms for solving such problems. AA222: MDO 114 Thursday 26th April, 2012 at 16:05 Figure 5.1: Example contours and feasible regions for a simple constrained optimization problem. 5.1.1 Nonlinear Equality Constraints Suppose we have the following optimization problem with equality constraints, minimize f(x) n with respect to x 2 R subject toc ^j(x) = 0; j = 1;:::; m^ To solve this problem, we could solve form ^ components of x by using the equality constraints to express them in terms of the other components. The result would be an unconstrained problem with n − m^ variables. However, this procedure is only feasible for simple explicit functions. Joseph Louis Lagrange is credited with developing a more general method to solve this problem, which we now review. At a stationary point, the total differential of the objective function has to be equal to zero, i.e., @f @f @f T df = dx1 + dx2 + ··· + dxn = rf dx = 0: (5.9) @x1 @x2 @xn T Unlike unconstrained optimization, the infinitesimal vector dx = [ dx1; dx2;:::; dxn] is not ar- bitrary here; if x is the sought after minimum, then the perturbation x + dx must be feasible: c^j(x + dx) = 0. Consequently, the above equations does not imply that rf = 0. For a feasible point, the total differential of each of the constraints (^c1;::: c^m^ ) must also be zero: @c^j @c^j T d^cj = dx1 + ··· + dxn = rc^j dx = 0; j = 1;:::; m^ (5.10) @x1 @xn To interpret the above equation, recall that the gradient of a function is orthogonal to its contours. Thus, since the displacement dx satisfiesc ^j(x + dx) = 0 (the equation for a contour), it follow that dx is orthogonal to the gradient rc^j. AA222: MDO 115 Thursday 26th April, 2012 at 16:05 Lagrange suggested that one could multiply each constraint variation by a scalar λ^j and subtract it from the objective function, m^ n 0 m^ 1 X X @f X @c^j df − λ^j d^cj = 0 ) @ − λ^j A dxi = 0: (5.11) @x @x j=1 i=1 i j=1 i Notice what has happened: the components of the infinitesimal vector dx have become inde- pendent and arbitrary, because we have accounted for the constraints. Thus, for this equation to be satisfied, we need a vector λ^ such that the expression inside the parenthesis vanishes, i.e., m^ @f X @c^j − λ^j = 0; (i = 1; 2; : : : ; n) (5.12) @x @x i j=1 i which is a system of n equations and n + m unknowns. To close the system, we recognize that the m constraints must also be satisfied. Suppose we define a function as the objective function minus a weighted sum of the constraints, m^ X L(x; λ^) = f(x) − λ^jc^j(x) ) j=1 L(x; λ^) = f(x) − λ^T c^(x) (5.13) We call this function the Lagrangian of the constrained problem, and the weights the Lagrange multipliers. A stationary point of the Lagrangian with respect to both x and λ^ will satisfy m^ @L @f X @c^j = − λ^j = 0; (i = 1; : : : ; n) (5.14) @x @x @x i i j=1 i @L =c ^j = 0; (j = 1;:::; m^ ): (5.15) @λ^j Thus, at a stationary point of the Lagrangian encapsulates our required conditions: the constraints are satisfied and the gradient conditions (5.12) are satisfied. These first-order conditions are known as the Karush{Kuhn{Tucker (KKT) conditions. They are necessary conditions for the optimum of a constrained problem. As in the unconstrained case, the first-order conditions are not sufficient to guarrantee a local minimum. For this, we turn to the second-order sufficient conditions (which, as in the unconstrained case, are not necessary). For equality constrained problems we are concerned with the behaviour of 2 ^ the Hessian of the Lagrangian, denoted rxxL(x; λ), at locations where the KKT conditions hold. In particular, we look for positive-definiteness in a subspace defined by the linearized constraints. Geometrically, if we move away from a stationary point (x∗; λ^∗) along a direction w that satisfies the linearized constraints, the Lagrangian should look like a quadratric along this direction. To be precise, the second-order sufficient conditions are T 2 ∗ ^∗ w rxxL(x ; λ )w > 0; n for all w 2 R such that ∗ T rc^j(x ) w = 0; j = 1;:::; m:^ AA222: MDO 116 Thursday 26th April, 2012 at 16:05 2 1 0 -1 -2 -2 -1 0 1 2 Figure 5.2: Contour plot for constrained problem (5.18). Example 5.22. Problem with Single Equality Constraint Consider the following equality constrained problem: minimize f(x) = x1 + x2 (5.16) weight respect to x1; x2 (5.17) 2 2 subject toc ^1(x) = x1 + x2 − 2 = 0 (5.18) The objective function and constraint of the above problem are shown inp Fig. 5.2. By inspection we can see that the feasible region for this problem is a circle of radius 2. The solution x∗ is obviously (−1; −1)T . From any other point in the circle it is easy to find a way to move in the feasible region (the boundary of the circle) while decreasing f. In this example, the Lagrangian is ^ 2 2 L = x1 + x2 − λ1(x1 + x2 − 2) (5.19) And the optimality conditions are " 1 # − ^ 1 2λ1x1 x1 2λ^1 rxL = = 0 ) = 1 1 − 2λ^1x2 x2 2λ^1 2 2 1 r^ L = x + x − 2 = 0 ) λ^1 = ± λ1 1 2 2 To establish which are minima as opposed to other types of stationary points, we need to look T at the second-order conditions. Directions w = (w1; w2) that satisfy the linearized constraints are AA222: MDO 117 Thursday 26th April, 2012 at 16:05 given by ∗ T 1 rc^1(x ) w = (w1 + w2) = 0 λ^1 ) w2 = −w1 while the Hessian of the Lagrangian at the stationary points is ^ 2 −2λ1 0 rxL = : 0 −2λ^1 Consequently, the Hessian of the Lagrangian in the subspace defined by w is ^ T 2 ∗ −2λ1 0 w1 ^ 2 w rxxL(x )w = w1 −w1 = −4λ1w1 0 −2λ^1 −w1 ^∗ 1 In this case λ1 = − 2 corresponds to a positive-definite Hessian (in the space w) and, therefore, the T 1 1 T T solution to the problem is (x1; x2) = ( ; ) = (−1; −1) . 2λ^1 2λ^1 ∗ ∗ Also note that at the solution the constraint normal rc^1(x ) is parallel to rf(x ), i.e., there is ^∗ a scalar λ1 such that ∗ ^∗ ∗ rf(x ) = λ1rc^1(x ): We can derive this expression by examining the first-order Taylor series approximations to the objective and constraint functions. To retain feasibility with respect toc ^1(x) = 0 we require that c^1(x + d) = 0 ) T T c^1(x + d) =c ^1(x) +rc^1 (x)d + O(d d): | {z } =0 Linearizing this we get, T rc^1 (x)d = 0 : We also know that a direction of improvement must result in a decrease in f, i.e., f(x + d) − f(x) < 0: Thus to first order we require that f(x) + rf T (x)d − f(x) < 0 ) rf T (x)d < 0 : A necessary condition for optimality is that there be no direction satisfying both of these conditions. The only way that such a direction cannot exist is if rf(x) and rc^1(x) are parallel, that is, if rf(x) = λ^1rc^1(x) holds. By defining the Lagrangian function L(x; λ^1) = f(x) − λ^1c^1(x); (5.20) and noting that rxL(x; λ^1) = rf(x) − λ^1rc^1(x), we can state the necessary optimality condition ∗ ^∗ ∗ ^∗ as follows: At the solution x there is a scalar λ1 such that rxL(x ; λ1) = 0.
Recommended publications
  • Introduction to the Modern Calculus of Variations
    MA4G6 Lecture Notes Introduction to the Modern Calculus of Variations Filip Rindler Spring Term 2015 Filip Rindler Mathematics Institute University of Warwick Coventry CV4 7AL United Kingdom [email protected] http://www.warwick.ac.uk/filiprindler Copyright ©2015 Filip Rindler. Version 1.1. Preface These lecture notes, written for the MA4G6 Calculus of Variations course at the University of Warwick, intend to give a modern introduction to the Calculus of Variations. I have tried to cover different aspects of the field and to explain how they fit into the “big picture”. This is not an encyclopedic work; many important results are omitted and sometimes I only present a special case of a more general theorem. I have, however, tried to strike a balance between a pure introduction and a text that can be used for later revision of forgotten material. The presentation is based around a few principles: • The presentation is quite “modern” in that I use several techniques which are perhaps not usually found in an introductory text or that have only recently been developed. • For most results, I try to use “reasonable” assumptions, not necessarily minimal ones. • When presented with a choice of how to prove a result, I have usually preferred the (in my opinion) most conceptually clear approach over more “elementary” ones. For example, I use Young measures in many instances, even though this comes at the expense of a higher initial burden of abstract theory. • Wherever possible, I first present an abstract result for general functionals defined on Banach spaces to illustrate the general structure of a certain result.
    [Show full text]
  • Finite Elements for the Treatment of the Inextensibility Constraint
    Comput Mech DOI 10.1007/s00466-017-1437-9 ORIGINAL PAPER Fiber-reinforced materials: finite elements for the treatment of the inextensibility constraint Ferdinando Auricchio1 · Giulia Scalet1 · Peter Wriggers2 Received: 5 April 2017 / Accepted: 29 May 2017 © Springer-Verlag GmbH Germany 2017 Abstract The present paper proposes a numerical frame- 1 Introduction work for the analysis of problems involving fiber-reinforced anisotropic materials. Specifically, isotropic linear elastic The use of fiber-reinforcements for the development of novel solids, reinforced by a single family of inextensible fibers, and efficient materials can be traced back to the 1960s are considered. The kinematic constraint equation of inex- and, to date, it is an active research topic. Such materi- tensibility in the fiber direction leads to the presence of als are composed of a matrix, reinforced by one or more an undetermined fiber stress in the constitutive equations. families of fibers which are systematically arranged in the To avoid locking-phenomena in the numerical solution due matrix itself. The interest towards fiber-reinforced mate- to the presence of the constraint, mixed finite elements rials stems from the mechanical properties of the fibers, based on the Lagrange multiplier, perturbed Lagrangian, and which offer a significant increase in structural efficiency. penalty method are proposed. Several boundary-value prob- Particularly, these materials possess high resistance and lems under plane strain conditions are solved and numerical stiffness, despite the low specific weight, together with an results are compared to analytical solutions, whenever the anisotropic mechanical behavior determined by the fiber derivation is possible. The performed simulations allow to distribution.
    [Show full text]
  • A Variational Approach to Lagrange Multipliers
    JOTA manuscript No. (will be inserted by the editor) A Variational Approach to Lagrange Multipliers Jonathan M. Borwein · Qiji J. Zhu Received: date / Accepted: date Abstract We discuss Lagrange multiplier rules from a variational perspective. This allows us to highlight many of the issues involved and also to illustrate how broadly an abstract version can be applied. Keywords Lagrange multiplier Variational method Convex duality Constrained optimiza- tion Nonsmooth analysis · · · · Mathematics Subject Classification (2000) 90C25 90C46 49N15 · · 1 Introduction The Lagrange multiplier method is fundamental in dealing with constrained optimization prob- lems and is also related to many other important results. There are many different routes to reaching the fundamental result. The variational approach used in [1] provides a deep under- standing of the nature of the Lagrange multiplier rule and is the focus of this survey. David Gale's seminal paper [2] provides a penetrating explanation of the economic meaning of the Lagrange multiplier in the convex case. Consider maximizing the output of an economy with resource constraints. Then the optimal output is a function of the level of resources. It turns out the derivative of this function, if exists, is exactly the Lagrange multiplier for the constrained optimization problem. A Lagrange multiplier, then, reflects the marginal gain of the output function with respect to the vector of resource constraints. Following this observation, if we penalize the resource utilization with a (vector) Lagrange multiplier then the constrained optimization problem can be converted to an unconstrained one. One cannot emphasize enough the importance of this insight. In general, however, an optimal value function for a constrained optimization problem is nei- ther convex nor smooth.
    [Show full text]
  • Derivative, Gradient, and Lagrange Multipliers
    EE263 Prof. S. Boyd Derivative, Gradient, and Lagrange Multipliers Derivative Suppose f : Rn → Rm is differentiable. Its derivative or Jacobian at a point x ∈ Rn is × denoted Df(x) ∈ Rm n, defined as ∂fi (Df(x))ij = , i =1, . , m, j =1,...,n. ∂x j x The first order Taylor expansion of f at (or near) x is given by fˆ(y)= f(x)+ Df(x)(y − x). When y − x is small, f(y) − fˆ(y) is very small. This is called the linearization of f at (or near) x. As an example, consider n = 3, m = 2, with 2 x1 − x f(x)= 2 . x1x3 Its derivative at the point x is 1 −2x2 0 Df(x)= , x3 0 x1 and its first order Taylor expansion near x = (1, 0, −1) is given by 1 1 1 0 0 fˆ(y)= + y − 0 . −1 −1 0 1 −1 Gradient For f : Rn → R, the gradient at x ∈ Rn is denoted ∇f(x) ∈ Rn, and it is defined as ∇f(x)= Df(x)T , the transpose of the derivative. In terms of partial derivatives, we have ∂f ∇f(x)i = , i =1,...,n. ∂x i x The first order Taylor expansion of f at x is given by fˆ(x)= f(x)+ ∇f(x)T (y − x). 1 Gradient of affine and quadratic functions You can check the formulas below by working out the partial derivatives. For f affine, i.e., f(x)= aT x + b, we have ∇f(x)= a (independent of x). × For f a quadratic form, i.e., f(x)= xT Px with P ∈ Rn n, we have ∇f(x)=(P + P T )x.
    [Show full text]
  • Contactmethodsintegratingplasti
    Contact methodsintegrating plasticity B13/1 modelswith application to soil mechanics ChristianWeißenfels Leibniz Universität Hannover Contact methods integrating plasticity models with application to soil mechanics Von der Fakultät für Maschinenbau der Gottfried Wilhelm Leibniz Universität Hannover zur Erlangung des akademischen Grades Doktor-Ingenieur genehmigte Dissertation von Dipl.-Ing. Christian Weißenfels geboren am 30.01.1979 in Rosenheim 2013 Herausgeber: Prof. Dr.-Ing. Peter Wriggers Verwaltung: Institut für Kontinuumsmechanik Gottfried Wilhelm Leibniz Universität Hannover Appelstraße 11 30167 Hannover Tel: +49 511 762 3220 Fax: +49 511 762 5496 Web: www.ikm.uni-hannover.de © Dipl.-Ing. Christian Weißenfels Institut für Kontinuumsmechanik Gottfried Wilhelm Leibniz Universität Hannover Appelstraße 11 30167 Hannover Alle Rechte, insbesondere das der Übersetzung in fremde Sprachen, vorbehalten. Ohne Genehmigung des Autors ist es nicht gestattet, dieses Heft ganz oder teilweise auf photomechanischem, elektronischem oder sonstigem Wege zu vervielfältigen. ISBN 978-3-941302-06-8 1. Referent: Prof. Dr.-Ing. Peter Wriggers 2. Referent: Prof. Dr.-Ing. Karl Schweizerhof Tag der Promotion: 13.12.2012 i Zusammenfassung Der Schwerpunkt der vorliegenden Arbeit liegt in der Entwicklung von neuartigen Kon- zepten zur direkten Integration von drei-dimensionalen Plastizit¨atsmodellen in eine Kontaktformulierung. Die allgemeinen Konzepte wurden speziell auf Boden Bauwerk Interaktionen angewandt und im Rahmen der Mortar-Methode numerisch umgesetzt. Zwei unterschiedliche Strategien wurden dabei verfolgt. Die Erste integriert die Boden- modelle in eine Standard-Kontaktdiskretisierung, wobei die zweite Variante die Kon- taktformulierung in Richtung eines drei-dimensionalen Kontaktelements erweitert. Innerhalb der ersten Variante wurden zwei unterschiedliche Konzepte ausgearbeitet, wobei das erste Konzept das Bodenmodell mit dem Reibbeiwert koppelt. Dabei muss die H¨ohe der Lokalisierungszone entlang der Kontaktfl¨ache bestimmt werden.
    [Show full text]
  • Dual-Dual Formulations for Frictional Contact Problems in Mechanics
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Institutionelles Repositorium der Leibniz Universität... Dual-dual formulations for frictional contact problems in mechanics Von der Fakultät für Mathematik und Physik der Gottfried Wilhelm Leibniz Universität Hannover zur Erlangung des Grades Doktor der Naturwissenschaften Dr. rer. nat. genehmigte Dissertation von Dipl. Math. Michael Andres geboren am 28. 12. 1980 in Siemianowice (Polen) 2011 Referent: Prof. Dr. Ernst P. Stephan, Leibniz Universität Hannover Korreferent: Prof. Dr. Gabriel N. Gatica, Universidad de Concepción, Chile Korreferent: PD Dr. Matthias Maischak, Brunel University, Uxbridge, UK Tag der Promotion: 17. 12. 2010 ii To my Mum. Abstract This thesis deals with unilateral contact problems with Coulomb friction. The main focus of this work lies on the derivation of the dual-dual formulation for a frictional contact problem. First, we regard the complementary energy minimization problem and apply Fenchel’s duality theory. The result is a saddle point formulation of dual type involving Lagrange multipliers for the governing equation, the symmetry of the stress tensor as well as the boundary conditions on the Neumann boundary and the contact boundary, respectively. For the saddle point problem an equivalent variational inequality problem is presented. Both formulations include a nondiffer- entiable functional arising from the frictional boundary condition. Therefore, we introduce an additional dual Lagrange multiplier denoting the friction force. This procedure yields a dual-dual formulation of a two-fold saddle point structure. For the corresponding variational inequality problem we show the unique solvability. Two different inf-sup conditions are introduced that allow an a priori error analysis of the dual-dual variational inequality problem.
    [Show full text]
  • 18.02SC Notes: Lagrange Multipliers
    Lagrange Multipliers We will give the argument for why Lagrange multipliers work later. Here, we’ll look at where and how to use them. Lagrange multipliers are used to solve constrained optimization problems. That is, suppose you have a function, say f(x; y), for which you want to find the maximum or minimum value. But, you are not allowed to consider all (x; y) while you look for this value. Instead, the (x; y) you can consider are constrained to lie on some curve or surface. There are lots of examples of this in science, engineering and economics, for example, optimizing some utility function under budget constraints. Lagrange multipliers problem: Minimize (or maximize) w = f(x; y; z) constrained by g(x; y; z) = c. Lagrange multipliers solution: Local minima (or maxima) must occur at a critical point. This is a point where rf = λrg, and g(x; y; z) = c. Example: Making a box using a minimum amount of material. A box is made of cardboard with double thick sides, a triple thick bottom, single thick front and back and no top. It’s volume is fixed at 3. What dimensions use the least amount of cardboard? Answer: We did this problem once before by solving for z in terms of x and y and substi­ tuting for it. That led to an unconstrained optimization problem in x and y. Here we will do it as a constrained problem. It is important to be able to do this because eliminating one variable is not always easy. The box shown has dimensions x, y, and z.
    [Show full text]
  • ECON 301 Two Variable Optimization (With- and Without- Constraints)
    ECON 301 Two Variable Optimization (with- and without- constraints) Review of Some Derivative Rules 1. Partial Derivative Rules: U = xy ∂U/∂x = Ux = y ∂U/∂y = Uy = x a b a 1 b a b 1 U = x y ∂U/∂x = Ux = ax − y ∂U/∂y = Uy = bx y − a b xa a 1 b a b 1 U = x y− = b ∂U/∂x = Ux = ax − y− ∂U/∂y = Uy = bx y− − y − U = ax + by ∂U/∂x = Ux = a ∂U/∂y = Uy = b 1/2 1/2 1 1/2 1 1/2 U = ax + by ∂U/∂x = Ux = a 2 x− ∂U/∂y = Uy = b 2 y− 2. Logarithm (Natural log) ln x ¡ ¢ ¡ ¢ (a) Rules of natural log If Then y = AB ln y =ln(AB)=lnA +lnB y = A/B ln y =lnA ln B y = Ab ln y =ln(Ab−)=b ln A NOTE: ln(A + B) =lnA +lnB 6 (b) derivatives IF THEN dy 1 y =lnx dx = x dy 1 y =ln(f(x)) = f 0(x) dx f(x) · (c) Examples If Then 2 1 y =ln(x 2x) dy/dx = (x2 2x) (2x 2) 1/−2 1 1 − 1 −1 y =ln(x )= 2 ln xdy/dx= 2 x = 2x ¡ ¢¡ ¢ 3. The Number e dy if y = ex then = ex dx f(x) dy f(x) if y = e then = e f 0(x) dx · (a) Examples 3x dy 3x y = e dx = e (3) 7x3 dy 7x3 2 y = e dx = e (21x ) rt dy rt y = e dt = re 1 Using Calculus For Maximization Problems OneVariableCase If we have the following function y =10x x2 − we have an example of a dome shaped function.
    [Show full text]
  • Math 241: Multivariable Calculus, Lecture 15 Lagrange Multipliers
    Math 241: Multivariable calculus, Lecture 15 Lagrange Multipliers. go.illinois.edu/math241fa17 Wednesday, October 4th, 2017 go.illinois.edu/math241fa17. Math 241: Problem of the day Problem: Find the absolute maximum and minimum value of the function f (x; y) = x3 − x + y 2x + y + 2 on the unit disk D = f(x; y) j x2 + y 2 ≤ 1g. What guarantees that the max and min exist? go.illinois.edu/math241fa17. rf (x; y) = h3x2 − 1 + y 2; 2yx + 1i −1 If this is the zero vector we get 2xy + 1 = 0, or y = 2x . (Note that x nor y can be zero, so dividing by x or y is no problem.) Substitute in the first component to get 1 1 3x2 − 1 + = 0 =) 3x4 − x2 + = 0: 4x2 4 2 2 1 Put z = x , get 3z − z + 4 = 0. The discriminant b2 − 4ac = 1 − 3 = −2, so there are NO critical points. So we need to look at the boundary! Solution Part 1: Critical points If f (x; y) = x3 − x + y 2x + y + 2 the gradient vector is go.illinois.edu/math241fa17. −1 If this is the zero vector we get 2xy + 1 = 0, or y = 2x . (Note that x nor y can be zero, so dividing by x or y is no problem.) Substitute in the first component to get 1 1 3x2 − 1 + = 0 =) 3x4 − x2 + = 0: 4x2 4 2 2 1 Put z = x , get 3z − z + 4 = 0. The discriminant b2 − 4ac = 1 − 3 = −2, so there are NO critical points. So we need to look at the boundary! Solution Part 1: Critical points If f (x; y) = x3 − x + y 2x + y + 2 the gradient vector is rf (x; y) = h3x2 − 1 + y 2; 2yx + 1i go.illinois.edu/math241fa17.
    [Show full text]
  • Notes on Calculus and Optimization
    Economics 101A Section Notes GSI: David Albouy Notes on Calculus and Optimization 1 Basic Calculus 1.1 Definition of a Derivative Let f (x) be some function of x, then the derivative of f, if it exists, is given by the following limit df (x) f (x + h) f (x) = lim − (Definition of Derivative) dx h 0 h → df although often this definition is hard to apply directly. It is common to write f 0 (x),or dx to be shorter, or dy if y = f (x) then dx for the derivative of y with respect to x. 1.2 Calculus Rules Here are some handy formulas which can be derived using the definitions, where f (x) and g (x) are functions of x and k is a constant which does not depend on x. d df (x) [kf (x)] = k (Constant Rule) dx dx d df (x) df (x) [f (x)+g (x)] = + (Sum Rule) dx dx dy d df (x) dg (x) [f (x) g (x)] = g (x)+f (x) (Product Rule) dx dx dx d f (x) df (x) g (x) dg(x) f (x) = dx − dx (Quotient Rule) dx g (x) [g (x)]2 · ¸ d f [g (x)] dg (x) f [g (x)] = (Chain Rule) dx dg dx For specific forms of f the following rules are useful in economics d xk = kxk 1 (Power Rule) dx − d ex = ex (Exponent Rule) dx d 1 ln x = (Logarithm Rule) dx x 1 Finally assuming that we can invert y = f (x) by solving for x in terms of y so that x = f − (y) then the following rule applies 1 df − (y) 1 = 1 (Inverse Rule) dy df (f − (y)) dx Example 1 Let y = f (x)=ex/2, then using the exponent rule and the chain rule, where g (x)=x/2,we df (x) d x/2 d x/2 d x x/2 1 1 x/2 get dx = dx e = d(x/2) e dx 2 = e 2 = 2 e .
    [Show full text]
  • Constrained Optimization 5
    Constrained Optimization 5 Most problems in structural optimization must be formulated as constrained min- imization problems. In a typical structural design problem the objective function is a fairly simple function of the design variables (e.g., weight), but the design has to satisfy a host of stress, displacement, buckling, and frequency constraints. These constraints are usually complex functions of the design variables available only from an analysis of a finite element model of the structure. This chapter offers a review of methods that are commonly used to solve such constrained problems. The methods described in this chapter are for use when the computational cost of evaluating the objective function and constraints is small or moderate. In these meth- ods the objective function or constraints these are calculated exactly (e.g., by a finite element program) whenever they are required by the optimization algorithm. This approach can require hundreds of evaluations of objective function and constraints, and is not practical for problems where a single evaluation is computationally ex- pensive. For these more expensive problems we go through an intermediate stage of constructing approximations for the objective function and constraints, or at least for the more expensive functions. The optimization is then performed on the approx- imate problem. This approximation process is described in the next chapter. The basic problem that we consider in this chapter is the minimization of a function subject to equality and inequality constraints minimize f(x) such that hi(x) = 0, i = 1, . , ne , (5.1) gj(x) ≥ 0, j = 1, . , ng . The constraints divide the design space into two domains, the feasible domain where the constraints are satisfied, and the infeasible domain where at least one of the constraints is violated.
    [Show full text]
  • Tabu Search for Constraint Satisfaction
    Tabu Search for Constraint Solving and Its Applications Jin-Kao Hao LERIA University of Angers 2 Boulevard Lavoisier 49045 Angers Cedex 01 - France 1. Introduction The Constraint Satisfaction Problem (CSP) is a very general problem able to conveniently formulate a wide range of important applications in combinatorial optimization. These include well-known classical problems such as graph k-coloring and satisfiability, as well as many practical applications related to resource assignments, planning and timetabling. The feasibility problem of Integer Programming can also be easily represented by the CSP formalism. As elaborated below, we provide case studies for specific CSP applications, including the important realm of frequency assignment, and establish that tabu search is the best performing approach for these applications. The Constraint Satisfaction Problem is defined by a triplet (X, D, C) where: – X = {x1,· · ·,xn} is a finite set of n variables. – D = {Dx1 ,· · ·,Dxn} is a set of associated domains. Each domain Dxi specifies the finite set of possible values of the variable xi. – C = {C1,· · ·,Cp} is a finite set of p constraints. Each constraint is defined on a set of variables and specifies which combinations of values are compatible for these variables. Given such a triplet, the basic problem consists in finding a complete assignment of the values to the variables such that that all the constraints are satisfied. Such an assignment is called a solution to the given CSP instance. Since the set of all assignments is given by the Cartesian product Dx1 ×· · ·×Dxn, solving a CSP means to determine a particular assignment among a potentially huge search space.
    [Show full text]