Donald Wittman APPENDIX A1 Mathematical Background

Constrained : Example: Maximizing subject to a budget Greater Utility constraint. The triangle bounded by the y x and y axes and the budget constraint . is the feasible set. Choose the highest iso-utility curve (indifference curve) Budget Constraint - a line that is feasible. Note that the constraint holds with equality. x

Linear Programming:

The constraints and iso-curves are all linear. The constraints do not have to y Higher Utility hold with equality. That is, some of the Feasible constraints are not-binding (e.g., the set line going north-east). The solution will be at one of the points of intersection. x

Higher Utility Non-:

The solution may be anywhere, including the interior of the feasible set if the of highest utility is interior y Feasible Set to the feasible set. This would be the case in the drawing to the left if the feasible set were expanded to the right so that it covered the innermost circle. x

140

Closed Convex Sets

Almost all optimization problems make extensive use of convex sets. This statement holds true for maximization via calculus and linear programming as well as non-linear optimization problems. Today I will show that closed convex sets and hyperplanes are the bases behind the optimization techniques that we will cover. These concepts are meant to give insight into the more mechanical methods that we employ and to show that we are fundamentally using the identical method in all that we have covered in previous courses. The proofs regarding convex sets are thus not meant for memorization but to provide basic understanding.

We first start off with a number of definitions:

Convex Set:

Geometric definition--a set is convex if and only if a (straight) line connecting any two points in the set is also in the set.

Convex sets:

Non Convex sets:

Convex Set: Algebraic definition--Let x,y be two vectors in N space in the set S. E.g., x = (x1, x2) y = (y1, y2) . S is convex if and only if Px + [1−P] y ! S for all P such that 0 ≤ P ≤ 1.

141

2

x = x , x 1 2

y = y , y 1 2

Note that a convex set is not the same thing as a convex .

Boundary point: Intuitively boundary points are points, which are on the edge of a set. More formally, a point is a boundary point if every neighborhood (ball) around the point contains points not in the set and in the set. They may be members of the set; e.g., in the set 0 ≤ x ≤ 1, the numbers 0 and 1 are boundary points and members of the set. In the set 0 < x < 1, the numbers 0 and 1 are boundary points and not members of the set.

Closed set: a closed set is a set that includes all its boundary points. The following is a closed set: 0 ≤ x ≤ 1. In economics we almost always consider closed convex sets. The reason can be illustrated by considering the opposite. Maximize the amount of gold (G) you get if G < 1. There is no maximum!

Preference set: The set of all points, which are indifferent or preferred to a given point. Economics majors are acquainted with indifference curves. A preference set is the set of all points on the indifference curve plus the all of the points strictly preferred to the indifference curve. In the typical drawing of an indifference curve the preference set would include the indifference curve plus all of the points upwards and to the right of the indifference curve.

We can view the ordinary two-dimensional graph of indifference curves as a two-dimensional topographical map showing height in terms of contour lines. Here height is utility and the contour lines are indifference curves. Consider the following diagram. There is a mountain

142

(representing utility) sticking out from the paper, the higher the mountain the higher the utility. Assume that it is an ice cream cone cut in half lengthwise, filled with ice cream and put upside down on the paper. The cone part is the indifference curve formed by a horizontal slice parallel to the paper; the cone plus the ice cream represents all points such that utility is either equal to or greater. We assume the set of such points is convex.

y

Indifference curves

x

Theorem 1. Sufficient conditions for a local optimum to be a global optimum.

If the feasible set F is a closed and convex set and if the objective function is a on that set and the set of points (such that the function is indifferent or preferred) create a convex set, then the local optimum is a global optimum.

This is a very important result. When these conditions hold, we know that that myopic optimization will end up at a global optimum. If either the feasible set or the preference set is not convex, then the local optimum need not be a global optimum.

Feasible set is not convex. 2nd peak is a local optimal but not a global optimum

(See following page)

143

144

Preference set is not convex. Right-most is not a global optimal.

145

We next turn our attention toward hyperplanes.

Hyperplane: a line in N-space dividing the space in two.

In two-space a hyperplane is a line. E.g., y = a + bx or a = y - bx or in more general notation, A = c1x1 + c2x2. In three-space we have a plane: y = A + BX + CZ, or in more general notation, …a = c1x1 + c2x2 + c3x3. In four or more space we have a hyperplane. Note that y ≥ a

+ Bx or a ≥ c1x1 + c2x2 creates a convex halfspace.

HYPERPANES CREATE CONVEX HALFSPACES

y

x x 2

x 1

Supporting hyperplane has one or more points in common with a (closed) convex set but no interior points. The set lies to one side of the hyperplane.

SUPPORTING HYPERPLANES

146

Theorem 2. Given z a boundary point of a closed convex set, there is at least one supporting hyperplane at z.

Theorem 3: If 2 convex sets intersect but without interior points, there is a supporting hyperplane that separates them.

The following diagrams illustrate other cases of supporting hyperplanes.

147

In the following figure, the shaded area is the feasible set and the curved line is the indifference curve, which characterizes the preference set. The straight line is the hyperplane. We usually give it a less technical term – the budget set or price line. Given these prices, the point of intersection maximizes utility.

Linear program Y supporting hyperplane of objective function F X

Neoclassical optimization:

Neoclassical optimization (usually only consider edge) good Y Max U (x, y) + ¬ (g(x) - 5)

indifference curve Price line tangent

Feasible hyperplane separating the two convex sets good X

148

There are other results that are of use. The following two theorems state that the intersection of closed convex sets is closed and convex.

Theorem 4. The intersection of two convex sets is convex.

Proof: If x ! (S ∩ T) and y ! (S ∩ T), then

Z = Px + [1 - P] y ! S because S convex

Z = Px + [1 - P] y ! T because T convex

Therefore Z ! (S ∩ T)

Theorem 5. The intersection of two closed sets is closed.

149

Donald Wittman APPENDIX A2 Concave and Quasiconcave Functions

A. CONCAVE FUNCTIONS

f(x)

Concave functions

x

Three definitions of concavity:

(1) Concave Function: The line connecting any two points of the function lies on the function or below.

(2) Concave Function: Algebraic definition: Pf (x) +[1 ! P] f (y) " f (Px + [1 ! P]y). Note that x and y may be vectors in n-space

(3) Concave Function: Tangency definition: The tangent is always outside or on the function.

The sum of two concave functions is also concave.

150

In order to determine whether a function is concave, we generally would prefer a method that did not rely on drawing a picture of the function. Hence, we have the second test.

d 2 f If f !! (x) = " 0 , for all x, then the function is concave. This is a necessary and sufficient dx 2 condition for concavity.

Example A: f(x) = -x4 is concave since f''(x) = -12x2 ! 0 for all x.

Example B: f(x) = - x3 is not concave since f''(x) = - 6x and - 6x > 0 for x < 0.

Example C: x ≥ 0 and f(x) = − x3 is concave since f''(x) = − 6x ≤ 0 for all x ≥ 0.

Notice that the test for concavity has a great similarity to the test for a maximum. To test for a maximum, you find out if the is less than zero when the first derivative equals zero. To speak imprecisely, the test for a maximum discovers whether the function is locally concave. In contrast, the test for concavity requires the second derivative to be non-positive everywhere.

Maximization Problem from Calculus: Max f (x) = 12x ! 3x2 f ! (x) =12 " 6x = 0 x = 2 f !! (x) = "6 < 0 Therefore at x = 2, f(x) is a maximum and not a minimum.

tangent

tangent above function (locally) where f '(x) = 0. Going down hill.

The analogous test for concavity: f !! (x) = "6 < 0. Therefore f is concave.

151

An Important Property of Concavity:

If f(x) is concave, then the set of x such that f(x) ≥ k is convex for all k.

f(x) k

convex set convex set

152

B. MULTI-DIMENSIONAL CONCAVE FUNCTIONS

Next, we would like to expand our intuition by visualizing concave functions and convex sets in two dimensions. If f (x1, x2 ) = z , then f (x1, x2 )is concave if any vertical plane looks like the single dimension picture (see the slice in the right hand side of the following picture).

f(x ,x ) x 1 2 2 slice

x 1

What happens if we take a horizontal slice? We then we have an iso-function. If f is utility, iso- utility is an indifference curve. Looking at x1, x2 :

The concentric circles are indifference curves from the figure to the above and left. x 2

Notice that the set of x1, x2 such that x 1 f (x1, x2) ! k is a solid convex space.

If f is concave, then the set x1, x2 such that f (x1, x2) ! k is convex for all k.

The reverse relationship need not be true (a convex set does not imply a concave function).

Hessian of 2nd order test—sufficient conditions.

Let f i be the partial of f with respect to the partial of i.

f11 f12 H1 < 0 = H for all x1, x2 ! f concave where H1 = f11 and H2 = H f21 f 22 H2 > 0

Note well this is only a sufficient condition for concavity. The necessary and sufficient conditions have loose inequalities and consider permutations (this will be covered later).

153

C. CONVEX FUNCTIONS

Definitions of Convex functions

f(x) : The line connecting any two points on a convex function does not lie below the function. x

Convex Function, tangency definition: the function never lies below any tangent to the function.

Convex function (a third definition): A function f is convex if and only if -f is concave.

Hessian test: f !! (x) " 0 for all x is a necessary and sufficient condition for f being convex.

Note that the set of all x such that f (x) ! K is convex.

f(x) k

Students often get mixed up between concave and convex functions and convex sets. Think of a concave function as creating a cave. This will help distinguish a concave function from a convex function. While there is a convex set there is no such thing as a concave set. The set of x such that f(x) ≥ k is convex when f(x) is a concave function; the set of x such that f(x) ≤ k is convex when f(x) is a convex function.

154

D. QUASICONCAVITY

Intuitive concept for one dimension: the function has only one peak.

Quasiconcavity: Algebraic definition: f (px + [1! P]y) " f (x) and/or f (y)

f(x)

Notice that all concave functions are quasiconcave.

Alternative algebraic definition: a function is quasiconcave if the set of all x such that f (x) ! k is convex for all k.

f(x)

Any horizontal cut creates a convex set of indifferent or higher points.

Another lecture covers the hessian test for quasiconcavity.

155

Note well that the sum of two quasiconcave functions need not be quasi-concave. E.g., try x3 and -x2. For x ≥ 0 both are quasiconcave.

3 x

-x2

Adding them together: x 3 ! x 2 = 0 at x= 0 0 − 0 = 0

at x=1/2 1/8 −1/4 = −1/8

at x =1 1 − 1 = 0

A quasiconcave function has to have one peak no matter what way you slice it.

These two boxes below look like there is only one peak, but if you go from A to B you will drop down to C.

B A

C

DEFINTION: f is quasiconvex if −f is quasiconcave.

156

Donald Wittman APPENDIX A3 Semidefinite

Students are acquainted with second order conditions. In the single variable case, if the first order conditions are equal to zero (f'(x) = 0), then f'' < 0 is sufficient for a local maximum. These results can be generalized to the n-variable case by looking at the of second derivatives when the firsts order conditions are satisfied. Now this approach to maximization is very limited. It only tests for local maxima; there may be a much higher peak elsewhere. Furthermore, it provides only a sufficient condition. For example, -x4 is a maximum at x = 0; yet its second derivative is −12x2 = 0 at x = 0. Therefore the sufficiency test is not applicable.

In programming we are interested in global optima. We do not look at just a point (where f' = 0), but the function as a whole. We test whether f is a concave function. It is if f'' ≤ 0 for all relevant values of x (say for x ≥ 0). If there is more than one variable, we then deal with the Hessian of second derivatives (this will be explained later). Thus global optimization in non-linear programming uses related ideas to the local optimization analysis (which in crude but incorrect terms finds local concavity) typically found in courses in intermediate economics, but there are three important differences: (1) Conditions on the second derivative of f apply to all values of x not just where the first order conditions are satisfied. (2) Inequalities need not be strict (zero values are allowed). (3) Permutations of the Hessian of second order partials are considered.

We will start with strict concavity and strict convexity, which are easier to handle. We know from the single variable case that if f'' < 0, then f(x) is a strictly concave function. The two- variable test is a bit more complicated as one has to make sure that the function is concave not only in one direction but in any direction. To start with, suppose that f is a function of three variables, f(x1, x2, x3). Letting fi stand for the partial of f with respect to xi, we have the following matrix of second derivatives:

157

! f11 f12 f13 $ # & f f f . # 21 22 23& "# f 31 f 32 f33%&

Note that the matrix is symmetric as fij = fji by Young’s theorem. We will now look at the determinants (denoted by ||) formed from this matrix (denoted by []).

If the following holds, then f is strictly concave:

f11 f12 f13 f11 f12 f11 < 0, = f11 f22 ! f 21 f12 > 0, f 21 f22 f 23 < 0. f 21 f22 f 31 f32 f 33

f f f 11 12 13 Note that there is a simple way of calculating f 21 f 22 f23

f 31 f 32 f33

f f f f f 11 12 13 11 12 f 21 f 22 f23 f 21 f22 = f11 f 22 f 33 + f12 f 23 f 31 + f13 f21 f 32 ! f 31 f 22 f13 ! f 32 f 23 f11 ! f33 f 21 f12 f 31 f 32 f33 f 31 f32

If the determinants derived from the matrix of second derivatives of f are negative definite for all relevant values of x, then f is strictly concave. This holds for any number of variables, not just 3.

Why not only if? f(x) = –x4 is strictly concave, but its second derivative = 0 at x = 0.

The matrix is known as being positive definite if the determinants of the successive principal minors are all positive for all relevant values of x.

158

If the determinants derived from the matrix of second derivatives of f are positive definite for all relevant values of x, then f is strictly convex.

Things get a bit more complicated when the above inequalities are no longer strict.

Let H k be a k-by-k formed by any permutation from the set of N variables along the diagonal of H, an N-by-N matrix. Note that the bar is to remind us that we are permuting.

! f11 f12 ... f1n $ # & f f ... f H =# 21 22 2n & # ...... & # & " f n1 f n2 ... f nn %

! f11 f12$ Thus if H =# & , then H1 = f11 and f22 " f 21 f 22%

f11 f12 f 22 f 21 H 2 = = f11 f 22 - f 21 f12 and = f11 f 22 - f21 f12 f21 f 22 f12 f11

As can be seen, the second expression is redundant. This is for two reasons: 1) the second determinant is equivalent to the first because we have interchanged a row and then a column, which does not alter the value of the determinant. 2) In the second place, we will only be considering second partials so that the matrices are symmetric. Therefore this alteration will have no effect.

159

! f11 f12 f13 $ # & If H = f f f # 21 22 23& "# f 31 f 32 f33%&

then H1 = f11 and f 22 and f 33

f11 f12 f22 f 21 f11 f13 f33 f 31 f22 f 23 f33 f 32 H 2 = and and and and and f21 f 22 f12 f11 f31 f 33 f13 f11 f32 f 33 f23 f 22

Note that the second determinant is redundant of the first, the fourth redundant of the third and the sixth redundant of the fifth.

Finally, H 3 = |H|.

Note that the diagonals in the sub-determinants are always fii from the original diagonal and that the off diagonal elements are matched to the diagonal. The same holds true for a bordered hessian, which we will get to later (the borders remain borders, diagonals remain diagonals and other elements match).

Thus f11 f 23 and f11 f 22 are inappropriate.

f 23 f 22 f 21 f12

The matrix is negative semidefinite if for all relevant values of x, all of the permutations of the determinants of the successive principal minors alternate in starting with a minus. That is, if for all the permutations, H1 ! 0;H 2 " 0;H 3 ! 0, then H is a negative semidefinite matrix.

A matrix of second derivatives of f is negative semidefinite if and only if f is concave.

If for all the permutations H1 ! 0;H 2 ! 0;H 3 ! 0, then H is a positive semidefinite matrix.

160

A matrix of second derivatives of f is positive semidefinite if and only if f is convex.

Just for fun (if you call this fun), I have written this out for the 4-by-4 case, where aij = fij.

a a a a 11 12 13 14 a21 a22 a23 a24 If H = a a a a then H1 = a11 and a22 and a33 and a44 31 32 33 34 a41 a42 a43 a44

a11 a12 a11 a13 a11 a14 a22 a23 a22 a24 a33 a34 H 2 = and and and and and a21 a22 a31 a33 a41 a44 a32 a33 a42 a44 a43 a44

a11 a12 a13 a11 a12 a14 a11 a13 a14 H3 = a21 a22 a23 and a21 a22 a24 and a31 a33 a34 a31 a32 a33 a41 a42 a44 a41 a43 a44

a22 a23 a24 and a32 a33 a34 a42 a43 a44

H 4 =| H |

A matrix is positive semidefinite if for all values the derived determinants

H1 ! 0, H2 ! 0, H3 ! 0

A matrix is negative semidefinite if for all values the derived determinants H1 ! 0, H2 " 0, H3 ! 0

161

We next consider the tests for quasiconcavity and quasiconvexity. Again, we will use the concepts of negative (or positive) definite of determinants and negative (or positive) semidefinite series of determinants, but this time the determinants will be bordered with the first derivatives and therefore the signs get reversed.

It is easy to get confused about signs. Here is the way that I remember. For concavity, we have a negative semidefinite expansion 1!1 " 0;2 ! 2 # 0;3! 3 " 0; etc. In testing for quasiconcavity, we reverse signs because we have a border: hence, 1!1 " 0;2 ! 2 # 0;3! 3 " 0; etc. As we will see, with a border, the following will always be true1!1 = 0;2 ! 2 " 0. So, we will only have to pay attention starting with the 3! 3determinant. For convexity, we have a positive semidefinite expansion 1!1 " 0;2 ! 2 " 0;3! 3 " 0; etc. In testing for quasiconvexity, we again reverse signs because we have a border: 1!1 " 0;2 ! 2 " 0;3! 3 " 0; etc. Again it is true that with a border, the following will always be true1!1 = 0;2 ! 2 " 0. So, we will only have to pay attention starting with the 3! 3 determinant.

We start with the two-variable case. Notice right away that the border is composed of first derivatives. Notice also that if we consider successive determinants, the first determinant will always equal 0 and the second will always be negative or 0. So we start with the third. If the 3x3 determinant is strictly positive (and the 4x4 determinant is strictly negative when we have 3 variables, and so forth), we have a negative definite symmetric matrix.

! 0 f1 f2 $ # & f f f # 1 11 12 & "# f 2 f 21 f22%&

If the bordered hessian of f is negative definite (starting with a positive for the 1x1), then f is strictly quasiconcave. If the bordered hessian of −f is negative definite (starting with a positive for the 1x1), then f is strictly quasiconvex.

162

Note that employing the minus sign in front of f, means the value of the determinant is opposite of the value of the determinant for f when there are an odd number of columns and the same value of the determinant as for f when there are an even number of columns. Hence if the expansion is always negative, then f is strictly quasiconvex.

When the value of one or more determinants is 0, we have to go back to permutations and semi- definiteness, with loose inequalities.

The bordered hessian of f is negative semidefinite (starting with a positive for the 1x1), if and only if f is quasiconcave. The bordered hessian of −f is negative semidefinite (starting with a positive for the 1x1), if and only if f is quasiconvex.

Example 1. f(x,y) = x + y2 where x, y ≥ 0.

fx = 1; fy = 2y

fxx = 0; fyy = 2; fxy = 0

0 0 H = 0 2

H1 = 0, 2 ! 0; H2 = 0

Hence, we have a positive semidefinite determinant and f is convex. This example illustrates why we need to do permutations. If we did not do the second permutation, we could have said that f is negative semidefinite, but this would be incorrect.

163

Example 2. f(x,y) = x2 + y2 where x, y ≥ 0.

fx = 2x; fy = 2y

fxx = 2; fyy = 2; fxy = 0

2 0 H = 0 2

H1 = 2 > 0; H2 = 4 > 0

Therefore we have a positive and a strictly convex function.

We know that a convex function is quasiconvex, as well. But it is useful to demonstrate this by looking at the bordered hessian (this is also a good way to check about the appropriate signing of a bordered hessian if you forget). So let us look at the bordered hessian. We start with the 3x3 border as the 2x2 determinants are never positive when we have a bordered hessian and the 1x1 is always 0.

0 f1 f2 0 2x 2y 0 2x

H = f1 f11 f12 = 2x 2 0 2x 2 2y f2 f21 f22 2y 0 2 0

= !8y2 ! 8x2 " 0 . So it is quasiconvex.

I have actually given you a slightly different algorithm. I said that if you multiply f by minus 1, and the –f function is quasiconcave, then f is convex. So let us multiply by –1.

0 !2x !2y 0 !2x H = !2x !2 0 !2x !2 !2y 0 !2 !2y 0

= 8y2 + 8x2 ! 0. So –f is quasiconcave and therefore f is quasiconvex.

164

Example 3. f(x,y) = x +2y ! 3xy where x, y ≥ 1

fx = 1 ! 3y; fy = 2 ! 3x

fxx = 0; fyy = 0; fxy = !3

0 !3 H = H = 0, H = 0; H = !9 < 0 !3 0 1 1 2

Therefore f is neither concave nor convex. Note that we have two values for H1 as we have to deal with each permutations, and that we have to deal with permutations because we have zeros and semidefiniteness. This time we need to go to the bordered hessian because we could not establish concavity or convexity.

We next test for quasiconcavity and quasiconvexity:

0 1 ! 3y 2 ! 3x 0 1 ! 3y H = 1 ! 3y 0 !3 1 ! 3y 0 2 ! 3x !3 0 2 ! 3x !3

= !6(1 ! 3y)(2 ! 3x) < 0 because x, y ≥ 1. So, quasiconconvex. Note well that if x, y ≥ 0 , f is neither quasiconcave nor quasiconvex because y < 1/3 and x > 2/3 would generate a positive number while y > 1/3 and x > 2/3 would generate a negative number. Concavity and convexity are global conditions not local.

Example 4: f(x,y) = x3y x ≥ 0, y ≥ 0

2 3 2 fx = 3x y fxx = 6xy fy = x fxy = 3x fyy = 0

2 6xy 3x H1 = 6xy and 0

2 4 3x 0 H2 = -9x

165

Therefore f is neither concave or convex.

Let us see if the function is quasiconcave or quasiconvex by looking at the bordered hessian.

0 3x2y x3 0 3x2y 3x2y 6xy 3x2 3x2y 6xy 2 2 x3 3x 0 x3 3x

= 9x7y + 9x7y - 6x7y = 12x7y ≥ 0 for x, y ≥ 0

Therefore negative semidefinite and therefore quasiconcave.

Note that when we have a bordered hessian, we do not pay attention to H1 as it is always 0 (the border remains the same) and H 2 as all 2x2 are negative (here too, remember that the border remains with the 0 in the upper left-hand corner). Note that in the case we were saved from doing permutations as we only had the 3x3 matrix to worry about.

Note that a function can be both concave and quasiconvex or both convex and quasiconcave. A straight line is both concave and convex (as well as quasiconcave and quasiconvex).

Example 5

Suppose that f = −x2 − y2 − z2 and that x, y, and z are non-negative. We know that −x2 is concave and that the same holds true for the other two variables. We also know that the sum of concave functions is concave and that a concave function is quasiconcave. Nevertheless, let us test directly whether f is quasiconcave as this exercise will help to cement our understanding.

166

0 f f f x y z 0 !2x !2y !2z fx fxx fxy fxz !2x !2 0 0 H = = fy fyx fyy fyz !2y 0 !2 0 !2z 0 0 !2 fz fzx fzy fzz

0 fx 0 fy 0 fz H 2 = and and It is always the case that these are ! 0. fx fxx fy fyy fz fzz

0 f f x y 0 !2x !2y 0 !2x !2z 0 !2z !2y

H 3 = fx fxx fxy = !2x !2 0 , !2x !2 0 , !2z !2 0 2y 0 2 2z 0 2 2y 0 2 fy fyx fyy ! ! ! ! ! !

These equal 8x2 +8y2 ! 0; 8x2 +8z2 ! 0; and 8z2 +8y2 ! 0, respectively.

Notice once again that if I switch 1 row and 1 column, the value of the determinant remains the same. In the last case, the z and y rows and z and y columns were switched.

0 f f f x y z 0 !2x !2y !2z fx fxx fxy fxz !2x !2 0 0 H 4 = H = = fy fyx fyy fyz !2y 0 !2 0 !2z 0 0 !2 fz fzx fzy fzz

Expanding by co-factors (by going down the last column) and repeating the first two rows for easy multiplication, we get,

!2x !2 0 !2x !2 0 !2x !2y 0 !2x H = 2z !2y 0 !2 !2y 0 ! 2 !2x !2 0 !2x !2 = !16z2 !16y2 !16x2 " 0 !2z 0 0 !2z 0 !2y 0 !2 !2y 0

167

So, the bordered hessian is negative semidefinite and therefore the function is quasiconcave. Note that because of the border, we have positive, negative, positive, negative starting with the first 1 by 1 in contrast to the test for concavity where negative semidefinite means having negative, positive, negative, positive without any border.

Example 6

Suppose that f = x2 + y2 + z2 and that x, y, and z are non-negative. We know that x2 is convex and that the same holds true for the other two variables. We also know that the sum of convex functions is convex and that a convex function is quasiconvex. Nevertheless let us test directly whether f is quasiconvex as this exercise will further help to cement our understanding.

0 f f f x y z 0 2x 2y 2z fx fxx fxy fxz 2x 2 0 0 H = = fy fyx fyy fyz 2y 0 2 0 2z 0 0 2 fz fzx fzy fzz

0 fx 0 fy 0 fz H 2 = and and It is always the case that these are ! 0. fx fxx fy fyy fz fzz

0 f f x y 0 2x 2y 0 2x 2z 0 2z 2y

H 3 = fx fxx fxy = 2x 2 0 , 2x 2 0 , 2z 2 0 2y 0 2 2z 0 2 2y 0 2 fy fyx fyy

These equal −8x2 −8y2 ! 0, −8x2 −8z2 ! 0, and −8z2 −8y2 ! 0, respectively.

168

0 f f f x y z 0 2x 2y 2z fx fxx fxy fxz 2x 2 0 0 H 4 = H = = fy fyx fyy fyz 2y 0 2 0 2z 0 0 2 fz fzx fzy fzz

Expanding by co-factors (by going down the last column) and repeating the first two rows for easy multiplication, we get,

2x 2 0 2x 2 0 2x 2y 0 2x H = !2z 2y 0 2 2y 0 + 2 2x 2 0 2x 2 = !16z2 !16y2 !16x2 " 0 2z 0 0 2z 0 2y 0 2 2y 0

Notice that that each expansion is now negative. We do not have a commonly used term for an all-negative expansion. We use positive semidefinite for an all-positive expansion, and negative semidefinite for an alternating negative and positive expansion (when there is no border) and an alternating positive and negative expansion (when there is a border). That is why the easiest way to remember all the rules is to test for convexity and quasiconvexity by multiplying the function by minus one and testing for concavity or quasiconcavity.

HOMEWORK:

Find out if concave, convex, quasiconcave, quasiconvex or none of the above for x, y ≥ 0. Your homework will be completed faster if you remember that a sum of concave functions is concave (but remember that a sum of quasiconcave functions need not be quasiconcave).

2 1) (x + y) + x

2) x3 + y1/2

3) x1/3 ! 3xy

169

4) x1/2 + y1/3 + 3(xy)1/2

5) 2 log(x+1) + (1/3)log(y+1), where log means natural log.

170

Donald Wittman APPENDIX A4 Necessary Conditions

Nonlinear programming is a generalization of linear programming, unconstrained maximization and constrained maximization.

1) The maximization problem has the following general format:

Maximize ! = f (x1, x2 ,x3 . . .xN )

subject to 1 g (x1, x2, . . . xN ) ! R1 2 g (x1, x2, . . .xN ) ! R2 . .

. . M g (x1, x2, . . .xN ) ! RM

xi " 0

Max ! = x 100 " x 2 " x 300 " x 3 + x x = f x ,x 1[ 1 ] 2[ 2 ] 1 2 ( 1 2) S.T. g1 x , x x x2 7 ( 1 2 ) = 1 2 ! 2 2 g (x1, x2 ) = x1 + 3x2 ! 5

x1, x2 ! 0

j th Remember that g (x1, x2 . . .) is the j function of the variables x1, x2 . . . we could have g(x1 . . .), h(x . . .), k(x . . .) but this would be notationally difficult.

If f and g j are linear functions then we have a linear program. If we have no constraints but f is nonlinear then we have optimization using calculus. If we have the constraints holding with strict equality, we have optimization with equality constraints.

171

2) Graphical Representation of .

Nonlinear programming does not have the graphical restrictions found in standard economic undergraduate texts. This can be illustrated via several examples.

2 2 Minimize C = (x1 ! 4) + (x2 ! 4) subject to

2x1 + 3x2 ! 6

!3x1 ! 2x2 " !12

x1, x2 ! 0

6

5 4

2 10 3 13, 13 2!!!!!!!!!!!!!!2 2

1

0

0 1 2 3 4

The feasible set is shaded in. A) Optimal solution as seen here need not be located at an extreme point. B) Only one constraint exactly fulfilled even though two constraints (including non- negativity there are four constraints) and two unknowns.

172

Another example:

2 2 Minimize C = (x1 ! 4) + (x2 ! 4) subject to

x1 + x2 ! 5

!x1 " !6

!2x2 " !11

x1, x2 ! 0 6 5 4 3 2 1

1 2 3 4 5 6

Again the feasible set is shaded in. In this example, the optimal solution does not lie on the boundary and thus none of the constraints are exactly fulfilled. So can't narrow down to boundary points. Here the constrained and unconstrained optimums are identical.

173

Our third example displays a problem—non convexity – that we will try to avoid. In this case, we have a non-convex constraint space (we may also have a non-convex preference set which would also cause problems) We have illustrated this before, but now we do so in a different way..

Maximize ! = 2 x1 + x2

2 !x1 + 4x1 ! x2 " 0

2x1 + 3x2 ! 12

x1, x2 ! 0 6 5 4 3 2 1

1 2 3 4 5 6

The straight lines are the indifference curves. The further to the right, the higher the indifference curve. We have a local maximum at 2, 3 but not a global maximum.

174

3) Kuhn Tucker Necessary Conditions

In calculus, the FOC or necessary conditions for an interior maximum were f i (x1, x2 ) = 0 for all i. We now consider the necessary conditions for a maximum in nonlinear programming. These are known as the Kuhn-Tucker necessary conditions. We start off with a very simple problem—a maximization problem with non-negativity constraints. This will enable us to see the logic behind the Kuhn-Tucker conditions.

Max f (x) x ≥ 0

0 (A) 0 0 (C) (B) where f (x) is a one variable .

df Let f ! " . If a local maximum occurs in an interior point (Figure A), the first order condition dx is f ! (x) = 0 . This is the same necessary condition as the free extremum. Even in the second example (Figure B), where we have a local extremum at a boundary point, f ! (x) = 0 . As a third possibility, we may have a local maximum at a boundary even though f ! (0) < 0 .

Thus the necessary conditions for a maximum are: A. f ! (x) = 0 and x > 0 B. f ! (x) = 0 and x = 0 C. f ! (x) < 0 and x = 0

For (C) we are at the lower boundary and we cannot decrease x more and make f (x) larger.

These conditions can be stated more elegantly as follows: f ! (x) ≤ 0 x ≥ 0 and x f ! (x) = 0 non-negativity says at least one constraint must equal zero

175

For a minimum the necessary conditions are as follows: f ! (x) ≥ 0 x ≥ 0 and x f ! (x) = 0

More generally, when ! = f (x1, x2 . . . xN ) = f (x), x is a vector, and xi ! 0, then the K-T !f necessary conditions for a maximum are: fi (x) = " 0 xi # 0 xi f i = 0 !xi Finally, we consider the most general case with inequality constraints: Max ! = f x = f x , x , . . . x ( ) ( 1 2 N ) subject to 1 1 g (x) = g (x1, x2 , . . .xN ) ! R1

g 2 (x ) ! R2 g M (x) ! RM

xi ! 0 This can be rewritten as a

1 2 M "Max" W(x1, x2 , . . . xN ,! ,! , . . .! )

1 1 1 2 2 2 M M M = f (x1, x2, . . . xN ) + ! (R " g ) + ! (R " g ) . . . ! (R " g )

Max is in quotes. Since we are finding a saddle-point not a maximum.

Necessary condition—K-Tucker necessary conditions

!W 1 1 2 2 M M "W = f i " # gi " # gi . . . # gi $ 0 xi ! 0 and • xi = 0. !xi "xi

!W j j i !W i i = R # g $ 0 ! " 0 and • " = 0.. !" !"i

176

Note that the Lagrangian method has converted the constraints into first order conditions. Also !W note that the KT complementary slack requirements # " i = 0 means that we have only !" i added zero's to the objective function.

A natural question is to ask how the non non-negativity constraints work. A simple example in the homework should provide insight.

177

4. Objective Function Requirements

A. f (x1, x2, . . . xn ) is either

(i) concave or

* * * * (ii) quasiconcave and there exists no point x = (x1 , . . . , xn ) such that f i (x ) = 0 for all i.

The rationale behind these assumptions: Say you have a concave objective function like the following diagram: f(x)

x

Then everything is OK. When you max you end up at a global max. The same holds true if function looked like this.

But, if the objective function looks like the following,

then one could find what looks like a "max" at the using the 2nd . So, to avoid this problem we must rule out complete local satiation.

178

4) Arrow Enthoven Constraint Qualifications

In order to avoid certain pathologies the following assumptions concerning the constraints must be true. These are known as the Arrow Enthoven Constraint Qualifications:

a) Every constraint function g j (x) in the maximization problem is twice differentiable and quasiconvex in (an open set containing) the non negative orthant. b) There exists a point x0 in the non negative orthant such that all the constraints are non- binding (although satisfied). That is, the feasible set has an interior point. c) One of the following is true: i. g j (x) is convex, or * i * ii. There exists no point x in the feasible set where g j (x ) = 0 for all i and j.

Discussion of points (a), (b) and (c):

(a) We know from earlier lectures that, for a maximization problem, quasiconvex constraints create convex feasible sets (that is, if f(x) is quasiconvex, the set of all x such that f(x) ! k is convex). From the opposite perspective, we have seen diagrams illustrating how non-convex feasible sets may yield non-optimal solutions even when the indifference curve is tangent to the feasible set. So requirement (a) is ensuring convex feasible sets. In a homework exercise, we will engage in a mathematical example showing that the KT conditions are not sufficient for a maximum when quasiconvexity of the constraints is violated. Fortunately, we already know how to test for quasiconvexity.

(b) In a homework exercise we will see via a simple example the problems that arise when (b) is violated.

179

(c) This is getting rid of constraint inflection points where the derivative is 0 (and is thus parallel to getting rid of objective function inflection points). If i holds, that is, the function is convex, then we do not have to worry about inflection points. If ii holds, then we do not have to worry either even though we only have quasi-convexity because the condition rules out the possibility of inflection points.

Example: Testing to see whether c is satisfied.

There is one constraint, LOG(x1 +1) ! LOG(x2 +1) " 7 . Then g1 = 1/ (x1 +1) and g2 =

2 2 −1/(x1 +1); g11 = !1/ (x1 +1) ;g12 = 0;g21 = 0;g22 = 1/ (x2 +1) . Looking at the hessian of second !1/ (x +1)2 0 derivatives, we can immediately see that the hessian 1 is not positive 2 0 1/ (x2+1) definite. So, the constraint is not convex. Hence, c.i is not satisfied.

We next test for quasiconvexity.

0 1/ (x +1) !1/ (x +1) 1 2 0 1/ (x1+1) 2 2 1/ (x +1) !1/ (x +1) 0 1/ (x1+1) 1 1 !1/ (x1+1) 2 !1/ (x +1) 0 1/ (x +1) !1/ (x2+1) 0 2 2 2 2 2 2 = 1/ (x2+1) (x1+1) !1/ (x2+1) (x1+1) = 0

So the constraint is quasiconvex (as well as being quasiconcave).

(c.ii) is satisfied as there are no set of values where both g1 and g2 = 0. (Indeed there are no non- negative values of x1 and x2 where either g1 or g2 is equal to 0.) Hence the AECQ conditions are satisfied.

180

6. Sufficient Conditions for a Global Maximum

Max ! = f (x ) x = x1, x2 ,x3, . . . xn

ST g j (x) ! Rj j =1, . . ., M

xi ! 0 i =1, . . . , N

If the following is true then x* gives a global maximum.

(1) The objective function, f, is differentiable and quasiconcave in an open set containing the non-negative orthant.

(2) Each constraint, g j , is differentiable and quasiconvex in an open set containing the non- negative orthant.

(3) x* satisfies the Kuhn-Tucker necessary conditions.

(4) The set g j satisfies the AECQ conditions (this is a redundant because an element of (3) is that the AECQ or another set of constraint qualification hold): a) g j is either convex

i * or quasiconvex but there exists no point x* such that g j (x ) = 0 for all i, j, b) There is an interior point to the constraint set.

(5) The second derivative of f exists and a) f (x) is concave

* * b) or quasiconcave and there exists no point x such that f i (x ) = 0 for all i (rules out the local satiation).

181

A word about necessary and sufficient conditions: Sufficient means that if the conditions hold you will have a maximum; necessary means that if the conditions do not hold, you will not have a maximum (equivalently, only if the conditions hold is a maximum possible).

HOMEWORK

A. In the following examples, first establish (1) whether the constraint qualifications are met and (2) whether the conditions on the objective function are met. Then find the KT necessary conditions and solve for all values. If (1) or (2) are not satisfied show that the KT necessary conditions can be met, yet the objective is not achieved or that it is impossible for the K-T conditions to be met. In all cases, x ≥ 0. Also solve graphically.

Note that a of one variable is quasiconcave and quasiconvex. In some of these problems, certain conditions are violated. Therefore the K-T conditions are no longer necessary and sufficient. Therefore it is possible for the KT conditions to be satisfied and yet the point not be a maximum.

1. Max f (x) = x subject to x ≤ 10 (Note that λ is the value of relaxing the constraint and that it is at least as large as the marginal value of x.)

2. Max f (x) = x2 subject to x3 ≤ 8

3. Max f (x) = x.2 subject to x3 ≤ 8 and x ≤ 10

2 2 4. Max f (x) = x + y subject to (x ! 2) + y " 4 and (x ! 2) ! y " !4 . In this case, the AECQ conditions cannot be satisfied.

182

HOMEWORK 2 NOTE: other homework assignments may be substituted for this and the following homework assignments. In each case do the following: 1. Check if objective function satisfies either: A. Concavity or B. Quasiconcavity and non-satiation 2. Check if constraints satisfy Arrow-Enthoven conditions 3. Write out Lagriangian 4. Write out Kuhn-Tucker conditions 5. Provide an economic interpretation of Kuhn-Tucker conditions. E.g. marginal something equals marginal something else. 6. If possible determine which KT equations hold with equality 7. Perform requested comparative statics based on equalities in 6.

A. A Firm maximizes Profit = [A - BQ(K, L)]Q(K, L) -wL - Ki

Let Q(K, L) be the production function. Q has a hessian of second order derivitives that is negative definite. A > 2BQ. A, B > 0. QL, QK > 0. QKL = 0.

Answer questions 1-5. Also show that Profits are homogeneous of degree 0 in A, B i and w. Show that the Profit function is convex.

Find the effect of an increase in A on K; the effect of an increase of w on K, the effect of an increase in A on Profits; and the effect of an increase in i on profits.

183

B. Peak load Pricing C.E.G. Competitive Firm P1, P2 > 0 Prices in the two periods given. Q1 — supply of good in period 1 Q1 > 0 Q2 — supply of good in period 2 Q2 > 0 Assume that Q2 > Q1 variable cost = C = AQ12 + BQ1Q2 + DQ22A, B, D > 0 4AD > B2 i > 0 interest rate which is the cost of capital K ≥ 0 Max ∏ = P1Q1 + P2Q2 – AQ12 – BQ1Q2 – DQ22 – Ki Present value of price

Subject to .01 Q1 ≤ K supply of good uses 1% of capacity .01 Q2 ≤ K

Find the effect of a change in P1 on Q2

Note that Q1 and Q2 are not functions of K and L, but primitives (variables). The firm is choosing Q1, Q2 and K subject to constraints.

C. A Shipping Problem A monopolist ships from point A to B and back. We assume that the demand curve is higher for going out than for coming back.

X0 = Number of paid trips out

XB = Number of paid trips back

X0 ≥ XB Treat this as a constraint (that is, a ship cannot return if it does not go out).

We assume that X0 > XB and that AB > 2BXB

P0 = A0 – BX0 A0, AB , B ≥ 0 Pb= AB – BXB

184

HOMEWORK 3

(1) In order for WAIBPC to be a profit function, what restrictions are there on A and B and C. W = Wage, I = Interest Rate, P = Price of Good.

1 1 (2) Suppose Q(K, L) = K 3 L3 and the firm is competitive in input and ouput markets. A) Derive Profit maximizing relations. B) What are second order conditions? C) Find the of the isoquant. D) Show that ∏ is homogeneous of degree 1.

(3) A monopolist has the following demand curve: P = A – BQ A, B > 0 CQ2 is the cost curve C > 0 A) Find FOC, SOC. B) Find the effect of an increase in C on Q. C) Find the effect of an increase in C on ∏. D) Write out the profit function, ∏*, explicitly. Remember that a profit function is a function of the exogenous variables only.

The cost of a round trip is C(Xo)2 where C is a constant.

Find the effect of an increase of AB on X0.

185

HOMEWORK 4

1 1 Starting with a Cobb Douglas production function Q = K 3 L3 , find L*, K* and ∏*. Show that

∏* satisfies the properties of a profit function then derive the cost function.

This is a lot of mathematical manipulation, but you need to do it to see how everything fits together.

186

HOMEWORK 5

Q B, W B i 2B (wi)1/2 1. Show that [ + i + 3 ] is a cost function.

2. Starting with a Cobb-Douglas technology (with decreasing returns to scale).

a. Derive the first order conditions for profit maximization. (3) b. Show that the second order conditions are satisfied. (4) c. Determine the elasticity of substitution. (5)

3. Suppose that Q(K,L) is concave with a negative definite Hessian of second order partials.

a. Find the first order conditions for the minimizing cost subject to an output constraint. (3) b. Find the effect of a change in W on cost. (3)

4. Again assume Q(K,L) has a negative defnite Hessian of second partials.

a. Find the first order conditions of a perfectly competitive profit maximizing firm (3) and show the relationship between marginal products and factor prices when L,K>0. (1) b. Show that profits are homogeneous of degree O in input and output prices. (3) c. Show that if there are constant returns to scale, there are zero profits. (4) d. Find the effect of a change in i on profits. (3)

5. Suppose that Q(K,L) = K+L.

a. Derive the isoquant and draw it. (3) b. Derive the minimal cost function (4) and derive, explain, and draw the isocost curve. (4) c. Derive and draw the factor demand curve for labor. (3)

187