Mixed Integer Strategies for Parameter Optimization

Rui Li [email protected] Group, Leiden Institute of Advanced Computer Science, Leiden University, Leiden, 2333 CA, The Netherlands Michael T.M. Emmerich [email protected] Natural Computing Group, Leiden Institute of Advanced Computer Science, Leiden University, Leiden, 2333 CA, The Netherlands Jeroen Eggermont [email protected] Division of Image Processing, Department of Radiology C2S, Leiden University Medical Center, Leiden, 2300 RC, The Netherlands Thomas Back¨ [email protected] Natural Computing Group, Leiden Institute of Advanced Computer Science, Leiden University, Leiden, 2333 CA, The Netherlands M. Schutz¨ [email protected] Blumenthalstraße 14, 69120 Heidelberg, Germany J. Dijkstra [email protected] Division of Image Processing, Department of Radiology C2S, Leiden University Medical Center, Leiden, 2300 RC, The Netherlands J.H.C. Reiber [email protected] Division of Image Processing, Department of Radiology C2S, Leiden University Medical Center, Leiden, 2300 RC, The Netherlands

Abstract Evolution strategies (ESs) are powerful probabilistic search and optimization algo- rithms gleaned from biological evolution theory. They have been successfully applied to a wide range of real world applications. The modern ESs are mainly designed for solving continuous parameter optimization problems. Their ability to adapt the param- eters of the multivariate normal distribution used for mutation during the optimization run makes them well suited for this domain. In this article we describe and study mixed integer evolution strategies (MIES), which are natural extensions of ES for mixed inte- ger optimization problems. MIES can deal with parameter vectors consisting not only of continuous variables but also with nominal discrete and integer variables. Following the design principles of the canonical evolution strategies, they use specialized muta- tion operators tailored for the aforementioned mixed parameter classes. For each type of variable, the choice of mutation operators is governed by a natural metric for this variable type, maximal entropy, and symmetry considerations. All distributions used for mutation can be controlled in their shape by means of scaling parameters, allowing self-adaptation to be implemented. After introducing and motivating the conceptual design of the MIES, we study the optimality of the self-adaptation of step sizes and mutation rates on a generalized (weighted) sphere model. Moreover, we prove global

C 2013 by the Massachusetts Institute of Technology 21(1): 29–64

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

convergence of the MIES on a very general class of problems. The remainder of the article is devoted to performance studies on artificial landscapes (barrier functions and mixed integer NK landscapes), and a case study in the optimization of medical image analysis systems. In addition, we show that with proper handling techniques, MIES can also be applied to classical mixed integer problems.

Keywords Evolution strategies, mixed integer evolution strategies, NK landscapes.

1 Introduction The classical problem formulation of continuous parameter optimization reads: min{f (x)|x ∈ M ⊆ Rn}. Aside from this, in and combinatorial optimization discrete search spaces are considered, such as Zn or Bn.Mostofthere- search in optimization theory is either focused on discrete search spaces or focused on continuous search spaces. However, there are many practical optimization problems from industry in which the set of decision variables involves continuous as well as dis- crete variables. These problems are classically referred to as mixed integer optimization problems (Dakin, 1965; Wolsey and Nemhauser, 1999). Classical techniques from mathematical programming have focused mainly on problems of which the structure is clearly defined by a set of mathematical expressions. For instance, mixed integer nonlinear programming (MINLP; Bussieck and Pruessner, 2003) is a natural approach of formulating problems where continuous and discrete parameters need to be optimized simultaneously. It is based on the assumption that the search space can efficiently be explored using a divide and conquer scheme. As opposed to these white box optimization problems, less attention has been paid to black box optimization problems where the structure of the objective function is not fully known. This type of problem became very apparent in applications where objective functions are based on large-scale simulation models. In this article, we discuss a promising approach for solving such problems, the so-called mixed integer evolution strategies (MIES). Evolution strategies (ESs; Back,¨ 1996; Rechenberg, 1994; Schwefel, 1995) are robust strategies that are frequently used for continuous optimization. They belong to the class of randomized search and are based on principles of biological evolution theory, such as selection, recombination, and mutation. As flexible and robust search techniques, they have been successfully applied to various real- world problems. The dynamic behavior of ESs was subject to thorough theoretical and empirical studies. For instance, Schwefel compared it to traditional direct optimization and Back¨ to other evolutionary algorithms. Theoretical studies of the conver- gence behavior of the ES were carried out for instance by Beyer (2001), Oyman (Oyman et al., 2000), and Rudolph (1997). The results indicate that the ES is a robust optimization tool that can deal with a large number of practically relevant function classes, including discontinuous and multimodal functions with many variables. This article discusses the MIES, an extension of the ES for the simultaneous opti- mization of continuous, integer, and nominal discrete parameters. It combines muta- tion operators of ESs in the continuous domain (Schwefel, 1995), for integer program- ming (Rudolph, 1994), and for binary search spaces (Back,¨ 1996). These operators have in common that they have certain desirable properties, such as symmetry, scalability, and maximal entropy, the details of which will be discussed later. The MIES was originally developed for optical filter optimization (Back;¨ Schutz¨ and Sprave, 1996), and chem- ical engineering plant optimization (Emmerich et al., 2000; Groß, 1999). Recently, as

30 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

discussed in this contribution, it has been used in the context of medical image analysis (Li, Emmerich, Bovenkamp et al., 2006; Bovenkamp et al., 2006). In the latter work, the MIES convergence behavior on various artificial landscapes was studied empirically, including a collection of single-peak landscapes (Emmerich et al., 2000) and landscapes with multiple peaks (Li, Emmerich, Eggermont, and Bovenkamp, 2006; Li, Emmerich, Eggermont, Bovenkamp, et al., 2006). This article summarizes and extends recent studies of MIES and their applications in various fields of science and technology. It extends the existing work by providing a self-contained detailed description of the and its motivation. Moreover, we provide theoretical studies on the optimality of the step size adaptation. A theorem on the convergence with probability one to the global optimum is derived for a very general class of functions. In addition, the article surveys existing applications of the MIES in systems design and extends empirical results by assessing the performance of the MIES on a set of classical MINLP benchmark problems. For the latter study, constraint handling methods are introduced into the framework. In Section 2, we provide a mathematical definition of mixed integer parameter op- timization and discuss methodologies applied for solving such problems. In Section 3, the MIES is introduced and we discuss its conceptual design. We study the behav- ior of the algorithm in artificial landscapes in Section 4. In addition, we introduce extensions of landscape models that are used typically either in binary optimization (NK-landscapes) or continuous optimization (barrier model). Furthermore, we apply MIES to the selected MINLP test problems by using constraint handling techniques, more specifically, penalty function methods. In Section 5, we report on the application of MIES to real-world problems. The article concludes with a summary discussion of the obtained results and discusses promising paths for future work.

2 Problem Definition and Related Work Many application problems from industry involve the simultaneous use of continuous, integer, and nominal discrete objective variables. The problem of mixed integer param-

eter optimization can be formalized as follows: let r1,...,rnr denote a set of real-valued

decision variables, z1,...,znz denote a set of integer decision variables, and d1,...,dnd denote a set of nominal discrete decision variables, each of which is taken from a finite domain. The finite domains for the nominal discrete variables will be denoted with D(1), ...,D(nd ). We do not encode nominal discrete variables as integers, in order to exploit the fact that there is no meaningful a priori ordering given for their domain. Fur- thermore, let f : Rnr × Znz × D(1) ×···×D(nd ) → R denote an objective function to be nr nz (1) (n ) minimized, and gi : R × Z × D ×···×D d → R,i= 1,...,ng denote constraint functions. Then the mixed integer optimization problem is defined as:

f (r ◦ z ◦ d) → min (1)

gi (r ◦ z ◦ d) ≤ 0,i= 1,...,m

hj (r ◦ z ◦ d) = 0,j= 1,...,n ∈ min max = ri [ri ,ri ],i 1,...,nr ∈ min max = zi [zi ,zi ],i 1,...,nz (i) di ∈ D ,i= 1,...,nd

Evolutionary Computation Volume 21, Number 1 31

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

min max Here, the constants ri and ri define lower and upper bounds for the real variables min max and the constants zi and zi define lower and upper bounds for the integer variables. The symbol ◦ denotes tuple concatenation. As we are interested in the black box scenario, we assume that the structure of f, gi ,i = 1,...,ng and hj ,j = 1,...,nh is unknown, or that we can only make some very general statements about it, such as continuity assumptions based on a similarity measure defined on the search space. As a result of this, it becomes harder to apply stan- dard techniques from mathematical programming, so-called mixed integer nonlinear programming methods (Floudas, 1995a) to solve them deterministically, such as outer approximation (OA; Duran and Grossmann, 1986), (BB; Borchers and Mitchell, 1994), and generalized Benders decomposition (Geoffrion, 1972). In cases where mathematical programming techniques fail, for mixed integer optimization can be an interesting method to heuristically search for solutions that improve the objective function value. In order to solve mixed integer optimization problems with metaheuristics, two general approaches can be considered.

1. Hierarchical Approach. Separate the discrete problem from the continuous problem by optimizing the discrete variables in a higher level optimization prob- lem and treating the optimization of the continuous parameters as a subprob- lem (Lohmann, 1992; Rechenberg, 1994; Utecht and Trint, 1994)

2. Simultaneous Approach. Optimize discrete and continuous parameters simul- taneously. In this approach, we consider that a similarity of parameter vectors due to an appropriate metric is equivalent to being positively correlated to the similarity in function values (Groß, 1999; Schutz¨ and Sprave, 1996).

In the following, we will study the simultaneous optimization of parameters. There are two reasons why we favor this approach over the hierarchical approach. First, the hierarchical approach requires a suboptimization of continuous parameters for each set of discrete parameters chosen in the outer level. This can be very time-consuming. Second, in the hierarchical approach, it is difficult to consider correlations between discrete and continuous variables, as they are strictly separated from each other.

3 Mixed Integer Evolution Strategies This section will provide a detailed description of the MIES, and then discuss and study its convergence behavior thoroughly.

3.1 Algorithm Description The problem of designing an ES for a new type of search space breaks down into three subtasks: (1) definition of the generational cycle, (2) definition of the individual representation, and (3) definition of variation operators for the representation of choice. These subtasks will be discussed next. + The chosen algorithm will be an instantiation of a (μ , λ)-ES for mixed integer + spaces. It generalizes the more common (μ , λ)-ES for continuous spaces. 3.1.1 Generational Cycle The main procedure of the ES is described in Algorithm 1. After a uniform random initialization and evaluation of the first population P (0) of μ individuals (parame- tervectorstakenfromanindividualspaceI) and setting the generation counter t to zero, the main loop of the algorithm starts. In a first step of the iteration, the algorithm

32 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Algorithm 1 (μ+,λ)-Evolution strategy t ← 0 initialize population P (t) ∈ Iμ evaluate the μ initial individuals with objective function f while termination criteria not fulfilled do for all i ∈{1,...,λ} do

choose uniform randomly parents ci1 and ci2 from P (t) (repetition is possible) ← xi mutate(recombine(ci1 ,ci2 )) Q(t) ← Q(t) ∪{xi } end for P (t + 1) ← μ individuals with best objective function value from P (t) ∪ Q(t)(plus), or Q(t) (comma) t ← t + 1 end while

generates the set Q(t)ofλ new offspring individuals, each of them obtained by the following procedure. Two individuals are randomly selected from P (t) and an offspring is generated by recombining these parents and then mutating (random perturbation) the individual resulting from the recombination. In the next step of the iteration, the λ offspring individuals are evaluated using the objective function to rank the individuals (the lower the objective function value, the better the rank). In case of a (μ + λ) selection, the μ best individuals out of the union of the λ offspring individuals and the μ parental individuals are selected. In case of a (μ, λ) selection, the μ best individuals out of the λ offspring individuals are selected. The selected individuals form the new parent population P (t + 1). After this, the generation counter is incremented. The generational loop is repeated until the termination criterion1 is fulfilled. 3.1.2 Representation An individual in an ES contains the information about one solution candidate. The contents of parent individuals is inherited by offspring individuals and is subject to variation. The standard representation of a solution in an ES individual is a continuous vector. In addition, parameters of the probability distribution used in the mutation (such as standard deviations or step sizes) are stored in the individual. The latter parameters are referred to as strategy parameters. To solve mixed integer problems with an ES we extend the real vector representation of individuals by introducing integer and nominal discrete variables as well as strategy parameters related to them. The domain of an individual then reads: I = ×···× × ×···× × ×···× × R1 Rnr Z1 Znz D1 Dnd As

Here, As denotes the domain of strategy parameters and is defined as: + nσ nς np As = R+ × [0, 1] ,nσ ≤ nr ,nς ≤ nz,np ≤ nd

1In most cases a maximal number of generations is taken as the termination criterion.

Evolutionary Computation Volume 21, Number 1 33

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

An individual of a population P (t) in generation t is denoted as:

= a (r1,...,rnr ,z1,...,znz ,d1,...,dnd ,σ1,...,σnσ ,ς1,...,ςnς ,p1,...,pnp )

The so-called object variables r1,...,rnr ,z1,...,znz ,d1,...,dnd determine the objective

function value and thus the fitness of the individual (cf. Equation (1)). Here, r1,...,rnr

denote real valued, z1,...,znz integer valued, and d1,...,dnd nominal discrete variables.

The so-called strategy variables σ1,...,σnσ are standard deviations used in the mutation

of the real valued variables, and ς1,...,ςnς denote mean step sizes in the mutation of

the integer parameters. Finally, p1,...,pnp denote mutation probabilities (or rates) for the nominal discrete object parameters. All these parameters are subject to inheritance, recombination, and mutation within Algorithm 1. Object variables are initialized uni- formly within their domain. The number of mutation parameters is variable. In this work, we consider the case of one single mutation parameter for each type of variable, and the case of one mutation parameter for each single variable.

3.1.3 Recombination The recombination operator can be subdivided into two steps, selection of the parents and recombining the selected parents. In this article, we apply local recombination which works with two recombination partners. The two recombination partners c1 ∈ I and c2 ∈ I are chosen randomly, uniformly distributed, from the parental generation for each of the offspring individuals. The information contained in these individuals is combined in order to generate an offspring individual. In ESs, two recombination types are commonly used: dominant and intermediate recombination (Schwefel, 1995). In a dominant (or) discrete recombination the operator chooses randomly one of the corresponding parental parameters for each offspring vec- tor position. Intermediate recombination computes the arithmetic mean of both parental vectors and thus, in general, can only be applied for continuous object variables and strategy variables. In mixed integer ES, dominant recombination is used for the solution parameters while intermediate recombination is used for the strategy parameters.

3.1.4 Mutation For the parameter mutation, standard mutations with maximal entropy for real, inte- ger and discrete parameter types are combined, as described previously (Back,¨ 1996; Rudolph, 1994; Schutz¨ and Sprave, 1996; Schwefel, 1995). The choice of mutation oper- ators was guided by the following requirements for a mutation in general search spaces (cf. Droste and Wiesmann, 2000; Rudolph, 1994; Beyer, 2007).

• Accessibility. Every point of the individual search space should be accessible from every other point by means of a finite number of applications of the mutation operator.

• Feasibility. The mutation should produce feasible individuals. This guideline can be crucial in search spaces with a high number of infeasible solutions.

• Symmetry. No additional bias should be introduced by the mutation operator.

• Similarity. Evolution strategies are based on the assumption that a solution can be gradually improved. This means it must be possible to generate similar solutions by means of mutation.

34 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Algorithm 2 Mutation of real valued parameters

input: r1,...,rnr ,σ1,...,σnr r ,...,r ,σ ,...,σ output: 1 nr 1 nσ control parameters: nσ ∈{1,nr } Nc ← N(0, 1) {generate and store a normally distributed random number}

τ ← √1 ; τ ← √ 1√ {initialize global and local learning rate} 2nr 2 nr

if nσ = 1 {single step size mode} then = σ1 σ1 exp(τNc) for all i ∈{1,...,nr } do ← + ri ri σ1N(0, 1) end for else {multiple step size mode}

for all i ∈{1,...,nr } do ← + σi σi exp(τNc τ N(0, 1)) ← + ri ri σi N(0, 1) end for end if {interval boundary treatment}

for all i ∈{1,...,nr } do

r ← T min max r i ri ,ri ( i ) end for

• Scalability. There should be an efficient procedure by which the strength of the impact of the mutation operator on the fitness values can be controlled.

• Maximal Entropy. If there is no additional knowledge about the objective func- tion available, the mutation distribution should have maximal entropy (Rudolph, 1994). By this measure, a more general applicability can be expected.

Respecting these guidelines, the following operators were selected in Emmerich et al. (2000). The mutation of continuous variables is described in Algorithm 2. The new individual is obtained by adding a normally distributed random perturbation to the old values of the vector. The corresponding standard deviations are also subject to the evolution process and are thus multiplied in each step with a logarithmically distributed random number. Schwefel (1995) termed the resulting process as self-adaptive, because the adaptation of the mutation parameters is governed by an evolutionary process itself. The general idea behind self-adaptation is that if a set of different individuals is generated, all with a different probability distribution, the individual with the best object variables is also likely the one with the best probability distribution that led to the generation of these object variables. Thus the parameters of this probability distribution are also inherited by the offspring individual. We now contribute some remarks on the properties of the normal distributions, as they are responsible for the choice of this type of distribution for mutating continuous variables. Among all continuous

Evolutionary Computation Volume 21, Number 1 35

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Figure 1: 2D representation of the distribution obtained as the difference of two geo- metrical distributions for different values of ϕ.

distributions with finite variance on R, the normal distribution possesses the maximum entropy.The multidimensional normal distribution is symmetrical to its mean value and unimodal. The step sizes represent standard deviations of the multi-dimensional normal distribution for each real valued variable. By variation of these standard deviations, the impact of the mutation on the variable vector in terms of similarity can be scaled. The mutation step size parameters are mutated by multiplication with a value drawn from a log-normal distribution. This way, increments are as likely as decrements, and bias is avoided. The global and local learning rates τ and τ are set according to the recommendations of Schwefel (1995). The last step of Algorithm 2 keeps the variables within the interval boundaries. It will be discussed later in this section. As opposed to continuous variables, integer variables are less commonly used in ESs. The mutation procedure for integer variables is borrowed from Rudolph (1994), where the replacement of the normally distributed random variables by the difference between two geometrically distributed variables was suggested. Among distributions defined on integer spaces, the multidimensional geometric distribution is one of the distributions of maximum entropy and finite variance. Rudolph proposed the use of the difference Z1 − Z2 of two geometrically distributed random variables, because the original geometric distribution is single tailed. The resulting distribution is depicted for 2 the 1D case in Figure 1 and the 2D case in Figure 2. It is l1-symmetrical centered around the mean value 0, unimodal, and it has an infinite support. Thereby, symmetry and accessibility of the mutation are provided. The strength of the mutation for the integer parameters is controlled by a set of step size parameters which represent the mean value of the absolute variation of the integer object variables. Accessibility is provided in the strict sense, as every point can be reached from every other point in a single step with positive probability. The details of this mutation operator are found in Algorithm 3. Note that a geometrically distributed random value can be generated by transforming a uniformly distributed random value u,using ln(1 − u) -1 z = ,ψ= 1 − ς 1 + 1 + ς 2 (2) ln(1 − ψ) The width of the distribution can be controlled by the parameter ς, the mean value of the exponential distribution (cf. Equation (1); for a derivation, see Rudolph, 1994). Excepting

2 ∈ Zn n | | The l1-norm of a vector z is defined as i=1 zi .

36 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Figure 2: 3D representation of the distribution obtained as the difference of two geo- metrical distributions. Figure courtesy of Gunter¨ Rudolph.

the different distribution types used, it is very similar to the real valued mutation operator in Algorithm 2. Self-adaptation is used to control the width parameter(s). The mutation of the width parameter is done as in Rudolph (1994) using a global learning rate τ and local learning rate τ .Sincethemeanstepsizebelow1isnotusefulfor integer problems, the mutated mean step size is set back to 1 whenever its mutation results in a value less then 1. To keep integer and continuous parameters in their feasible interval, a transformation function is applied to the mutation operators. Given a step size of the mutation, we may consider this to be the length a particle has to travel within the interval. Starting in the direction of the original unbounded mutation, whenever it meets with an interval boundary, the direction is inverted until the total length of the unbounded mutation has been covered. The implementation of this method is described in Algorithm 4 and an illustration of how the transformation function works can be found in Figure 3. Unlike other mappings, the limiting distribution of the random walk Xt+1 = Ta,b(Xt + σN(0, 1)),t = 1, 2,... is the uniform distribution. This means that there are no preferred regions of the search space in the long term in case of neutral selection and thus bias is avoided. In order to prevent a loss of causality, the step size should be kept smaller than the interval width. We recommend a maximal step size of 0.2(b − a). Finally, a mutation of the discrete parameters is carried out with a mutation prob- ability as described in Algorithm 5. The probability is a strategy parameter for each discrete variable. Each new value is chosen randomly (uniformly distributed) out of the finite domain of values. The application of a uniform distribution is due to the principle of maximal entropy, since the assumption was made that there is no reasonable order defined on the sets of discrete values. To consider requirements such as symmetry and scalability, we need to define a distance measure on the discrete subspace. The assumption that there is no order, which can be defined on the finite domains of discrete values, leads to the application 3  d ,...,d , d ,...,d = nd H d = d of the overlap distance measure: (( 1 nd ) ( 1 nd )) i=1 ( i i )with

3For the binary case this corresponds to the Hamming distance.

Evolutionary Computation Volume 21, Number 1 37

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Algorithm 3 Mutation of integer parameters

input: z1,...,znz ,ς1,...,ςnς z ,...,z ,ς ,...,ς output: 1 nz 1 nς control parameters: nς ∈{1,nz}

Nc ← N(0, 1)

τ ← √1 ; τ ← √ 1√ 2nz 2 nz if nς = 1 then {single step size mode} ← ς1 max(1,ς1 exp(τNc)) i ∈{ ,...,n } for all 1 z do -1 ς ← ← ← − + + 1 2 u1 U(0, 1); u2 U(0, 1); ψ 1 (ς1/nz) 1 1 ( n ) z − − ← ln(1 u1) ← ln(1 u2) G1 ln(1−ψ) ; G2 ln(1−ψ) ← + − zi zi G1 G2 end for else {multiple step size mode}

for all i ∈{1,...,nz} do ς ← max(1,ς exp(τN + τ N(0, 1))) i i c -1 ς ← ← ← − + + i 2 u1 U(0, 1); u2 U(0, 1); ψ 1 (ςi /nz) 1 1 ( n ) z − − ← ln(1 u1) ← ln(1 u2) G1 ln(1−ψ) ; G2 ln(1−ψ) ← + − zi zi G1 G2 end for end if {interval boundary treatment}

for all i ∈{1,...,nz} do

z ← T min max z i zi ,zi ( i ) end for

Algorithm 4 Computation Ta,b(x), for interval boundaries a and b y = (x − a)/(b − a) if y mod 2 = 0 then y =|y − y| else y = 1 −|y − y| end if x = a + (b − a)y return x

38 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

10

8

6 T(x) 4

2

0 0 2 4 6 8 10 x

Figure 3: An illustration to show working mechanism of the transformation function (a = 4andb = 6 in this example).

H (true)= 1; H(false)= 0 as a distance measure for guiding the design of the mutation operator. A self-adaptation of the mutation probability for the discrete parameters is achieved by a logistic mutation of these parameters, generating new probabilities in the feasi- ble domain. The logistic transformation function is recommended and discussed by Schutz¨ (Schutz¨ and Sprave, 1996). The basic idea of this transformation is to keep the variables within the range [0, 1] and let the probability of increasing p be equal to the probability of decreasing p. Given an original mutation probability p ∈ [0, 1], the mutation is p using the following expression:

1 p = (3) + 1−p − 1 p exp( τ N(0, 1)) Here N(0, 1) denotes a function that returns a normally distributed random number. We

recommend employing a second transformation function (Tpmin,pmax ) that keeps the value of p in the interval [1/(3nd ), 0.5]. The upper bound of 0.5 for the mutation probability is motivated by the observation that the mutation loses its causality once the probability exceeds the value of about 0.5. The lower bound is used to prevent the mutation probability from being too close to 0, in which case the MIES becomes insensitive to changes of that parameter. In the case where p = 1/(3nd ), a discrete mutation can be expected in every third application of the mutation operator. Depending on the discrete subspace, it is beneficial to use a single mutation proba-

bility instead of many individual mutation probabilities p1,...,pnd .Incaseofasingle mutation probability, for each position of the discrete subvector, it is decided inde- pendently, but with the same probability, whether to mutate this position or not. By adapting the mutation rate, the average number of mutations on the discrete val- ues is adjusted to the mean step size if the Hamming distance is considered as the metric.

3.2 Step Size Adaptation Study Previous work has shown that self-adaptive MIESs are able to converge to optima of simple functions in arbitrary precision by using step size adaptation. However, it is an open question whether the self-adaptation is adjusting the step size close to an optimal

Evolutionary Computation Volume 21, Number 1 39

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Algorithm 5 Mutation of nominal discrete parameters

input: d1,...,dnd ,p1,...,pnp d ,...,d ,p ,...,p output: 1 nd 1 np control parameters: np ∈{1,nd }

Nc ← N(0, 1)

τ ← √1 ; τ ← √ 1√ 2nd 2 nd

if np = 1 then {single step size mode} ← 1 p + 1−p − 1 p exp( τNc) = p T1/nd ,0.5(p )

for all i ∈{1,...,np} do if U(0, 1)

choose a new element uniform distributed out of Di \{di } end if end for else {multiple step size mode}

for all i ∈{1,...,np} do 1 ← − pi 1 pi 1+ exp(−τNc−τ N(0,1)) pi = pi T1/(3nd ),0.5(pi )

if U(0, 1)

value. A theoretical analysis of the step size adaptation in the mixed integer case is difficult, even for simple models such as the sphere model. We use a semiempirical analysis by approximating a local performance measure4 at a given location in the search space statistically for different step sizes, in order to find a step sizes ˆ∗ that approximately maximizes the local progress of the MIES. This computation is repeated for different stages of the evolution and each time the empirically found optimal step sizes ˆ∗ is compared to the current step size of the MIES found by self-adaptation. This approach is very flexible and can be applied for the analysis on a wide range of test problems for which the optimum is known and the evaluation of the objective function is fast. According to Beyer (2001) and Beyer et al. (2002), the local progress rate φ(s,x)fora step size s is the expectation of the distance covered toward the optimum in one mutation step starting from position x. Consider a mutation operator muts parameterized by the step size, an objective function f : I → R with single optimum x∗ ∈ I, and a position x

4Progress rate for Study 1 and quality gain for Study 2, and study 3.

40 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

in the metric search space (I,d). Then max{0,R(x) − R(muts (x))} ∗ φ(s,x) = E ,R(x) = d(x, x )(4) R(x) In order to statistically approximate φ(s,x) for a given pair (s,x) we compute the sample mean for M mutations:

M 1 max{0,R(x) − R(muti (x))} φˆ (s,x) = s . (5) M R(x) i=1 Depending on the parameter type, a different distance measurement is used to compute d(x, x∗) (cf. Equation (6)). For instance, the Euclidean distance is applied to continuous parameters, the Manhattan distance is applied to integer parameters, and for discrete parameters, we choose the overlap distance function. ⎧ ⎪ ⎪n ⎪ ∗ 2 ⎪ (xi − x ) if xi ∈ R (Euclidean distance) ⎪ i ⎪ i ⎪ ⎨n ∗ ∗ d(x, x ) = |x − x | if x ∈ Z (Manhattan distance) (6) ⎪ i i i ⎪ i ⎧ ⎪ ∗ ⎪n ⎨ = ⎪ 0if(xi xi ) ⎪ ∗ = ∈ D ⎪ I(xi ,xi ) ⎩ if xi (overlap distance) ⎩ = ∗ i 1if(xi xi )

The quality gain (Beyer, 2001) is an alternative local performance measure and is (t) (t) (t) defined as Q¯ = E(max(0, (f (x ) − f (muts (x )))/f (x ))) (in case of minimization). It can be estimated in a similar manner as the progress rate. Both the quality gain and the progress rate measure local performance. We assume here that optimizing local performance also optimizes global performance. We believe that for many unimodal landscapes this is true, but, for example, for multimodal landscapes, this may not be the case. The overarching research question is whether the MIES can find and keep a step size that maximizes the local progress rate or quality gain. Three studies addressing this question were conducted: (1) step size adaptation for problems with only one parameter type, (2) step size adaptation on a mixed problem with a single step size per parameter type, and (3) step size adaptation for multiple step sizes per parameter type for a mixed problem with differently scaled (weighted) variables. In all three studies, the value of φˆ (s,x(t))orQ¯ (s,x(t)) is computed in every generation (t) (t) t foranequidistantsetofL points in the interval [0,smax(x )], where x is the best in- (t) (t) ∗ dividual in generation t, and the upper boundary smax(x ) = 2|x − x | for continuous (t) and integer variables and smax(x ) = 0.1 for mutation probabilities. 3.2.1 Study 1 As a first study we test the MIES self-adaptation for a single step size per parameter type and for each parameter type independently on a simple problem landscape. 3.2.1.1 Experimental Overview As a test problem, the minimization of the sum of squares of the variables is used. Continuous, integer, and discrete spaces are studied separately. We compute progress

Evolutionary Computation Volume 21, Number 1 41

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

(t) (t) Figure 4: Comparison between the graph of φˆ (si, x ) and the step size s (vertical dashed line) of the best individual x(t) for continuous (top row), integer (middle row), and discrete (bottom row) variables used by the MIES at certain generations of the evolution. The step size s(t) is found by self-adaptation within the (4, 28)-MIES (i.e., without knowledge of φˆ ). On the right-hand side, a statistical summary of the efficiency of the step size adaptation for multiple runs is shown, using the SCE measure as defined in the text.

rates for different step sizes at different stages of the evolution and compare them to the step size of the best MIES individual. The results are shown in Figure 4 where the plot shows the computed progress rates and the vertical lines indicate the step size of the current best MIES individual (obtained by self-adaptation). In addition, we compute statistical summaries of repeated runs. For this, we measure for each generation t and run r the step size control efficiency SCE(t,r) = ∗ ∗ φ(st,r)/φ(st,r), where s denotes the step size found with self-adaptation and s the optimal step size. The results of this study are shown in the last column of Figure 4. 3.2.1.2 Detailed Setup For each generation t, the value of φˆ (s,x(t)) is computed for an equidistant set of L = 40 points and M = 5,000 samples. The search space dimension is 15 and the variable range is [−1,000, 1,000] for the integer and continuous variables, and {0,...,9} for the discrete variables. A (μ = 4,λ= 28)-MIES was used with a single step size (nσ = nς = np = 1) for each parameter type. We used a learning rate of 0.5, which slightly deviates from the default setting which also worked, but takes longer to adapt. The MIES step sizes

42 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

are initialized to 10% of the interval width for continuous and 33% of the interval width for integer variables. The mutation probabilities are initialized to 0.1. On purpose, we set the initial step size to a suboptimal value, to study whether self-adaptation is able to find a good step size.

3.2.1.3 Results Selected snapshots for different parameter types at different generations are shown in Figure 4. For all three parameter types, the self-adaptation is able to find the op- timal step size and track it. The plots of the typical runs (first three columns from the left-hand side) show that the step size needs to be chosen from a relatively small interval in order to achieve progress rates of, say, above 75%. The position of the red dashed lines indicates that the ES succeeds in finding and tracking this interval. The statistics on the SCE add support for this based on repeated runs (50 runs for each plot). The box plots seen in the right column of Figure 4 for repeated runs demon- strate that the MIES finds in most runs, after less than 10 generations, step sizes for which it operates at an efficiency of more than 75% and is able to keep this efficiency. This proves that, at least for relatively simple—but nevertheless high-dimensional— problems, the self-adaptation of the step sizes is able to find and track the optimal step size. Note that the scale of the plots in Figure 4 changes during the run by orders of magnitude.

3.2.2 Study 2 Secondly, we study the optimality of the step size adaptation for all three parameter types combined using a single step size per parameter type (nσ = nς = np = 1).

3.2.2.1 Experimental Overview As a test problem, we use a mixed integer sphere problem. We use here the quality gain as a measure, as there are many possible ways of how to combine the distance functions for the three parameter types in order to compute the progress rate.

3.2.2.2 Detailed Setup The sample mean for the quality gain is computed every generation based on 200,000 samples. Grid search with a resolution of 10 (3D grid with 1,000 points) has been used to obtain the optimal combination for the three step sizes (σ, ς, p). The boundaries of the cube are chosen according to the intervals described in Study 1. Also here (μ = 4,λ= 28) was used. A plot of 20 repeated runs was computed.

3.2.2.3 Result Figure 5 shows that the step size vector that maximizes the quality gain is learned quickly and kept as the evolution passes through different orders of magnitude. This demonstrates that step sizes or mutation rates for different parameter types can be learned simultaneously.

3.2.3 Study 3 Lastly, we study the use of multiple step sizes for each parameter on a weighted mixed integer sphere problem (nσ = nr ,nς = nz,np = nd ). Here we will look at each parameter type independently.

Evolutionary Computation Volume 21, Number 1 43

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

σ σ ς ς

Figure 5: Single step size dynamics on the 15D (nr = nz = nd = 5) sphere problems. The step sizes found by the MIES using self-adaptation and the optimized step sizes using statistical optimization have been plotted.

Figure 6: Multiple step size dynamics on the 9D (nr = nz = nd = 3) weighted sphere problem. The step sizes found by the MIES using self-adaptation and the optimized step sizes using statistical optimization have been plotted for discrete variables.

3.2.3.1 Experimental Overview

As a test problem, a 9D weighted mixed integer sphere problem is used (nr = nz = 5 nd = 3). Step sizes are studied separately. The results are shown in Figures 6 and 7,

5 2 2 2 2 2 2 2 The test problem reads: (r1) + 100(r2) + 10,000(r3) + (z1) + 100(z2) + 10,000(z3) + (d1) + 2 2 1,000(d2) + 100,000(d3) .

44 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Figure 7: Multiple step size dynamics on the 9D (nr = nz = nd = 3) weighted sphere problem. The step sizes found by the MIES using self-adaptation and the optimized step sizes using statistical optimization have been plotted for continuous (left) and integer variables (right).

where both the step size optimizing the quality gain and the step sizes of the best MIES individual for continuous, integer, and discrete genes are plotted. 3.2.3.2 Detailed Setup (t) Each value of φˆ (si, x ) in the coordinate search is estimated from M = 5000 samples. For each parameter type L = 40 steps have been taken. The same interval settings for the step sizes and control parameter settings for the MIES have been used as in Study 2. To reduce the effect of outliers, median values of 50 runs were presented in the plot. 3.2.3.3 Results Figure 7 shows that the multiple step sizes MIES is capable of keeping the step sizes for the continuous and integer variables close to the optimal value. The or- dering of the MIES step sizes corresponds to the ordering of the optimal step sizes which are counter-proportional to the weight. Figure 6 shows that it may have some problems in the learning process of mutation probability when np = nd . In contrast, learning of a single mutation rate for the nominal discrete variables is possible (see Figures 4 and 5). These findings are consistent with previous work (see Schutz,¨ 1994, 86–89). As a consequence, a single step size mode for nominal discrete parameters is recommended.

3.3 Global Convergence Properties Given that certain regularity requirements are met, it is possible to prove strong prob- abilistic convergence of the MIES for t →∞toward the global optimum. The theorem generalizes a theorem on the ES for continuous spaces by Born (1978). Both the plus and the comma strategy are considered, and for the convergence analysis the best solution t foundsofar,thatis,xbest, will be considered.

DEFINITION 1: A function f : C → R is called regular, if:

Evolutionary Computation Volume 21, Number 1 45

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

(A) f is continuous, (B) C ⊆ Rn is a closed set,

(C) ∀x ∈ C : ∀ >0:the set {x ∈ C|x = x ∧ f (x) ≤ f (x ) + } is nonempty.

DEFINITION 2: Let A ⊆ A and let g : A → B a function. By g|A we denote the restriction of

the function g defined by g|A : A → B and g|A (a ):= g(a ) where a ∈ A . Given these technical preliminaries, we can state the following theorem. n THEOREM 1: Let f : R × A → R denote a mixed integer function, and f |Rn×{a} is a reg- ular function for at least one a∗ ∈ A which is optimal. Then for a (μ + λ) MIES with lower t = limit σmin > 0 for the step sizes and mutation rates, the series f (xbest)t 1,2,... converges with probability one to the global minimum of f ,thatis, t ∗ Pr lim t = 0 = 1, with t = f x − f ≥ 0(7) t→∞ best

Here t represents the number of iterations, and f ∗ denotes the global optimum.

PROOF: From the construction of the algorithm it follows that

∀t ≥ 0: t+1 ≤ t (8)

and from the definition of a global optimum we get

∀t ≥ 0: t ≥ 0(9)

With Equations (8) and (9), it follows that t (t = 1, 2,...) has a limit value

lim t = ∞ (10) t→∞

Below we show that ∞ > 0 leads to a contradiction with the proposition in Equation ∞ t ∞ = →∞ (9) and thus  0istrue.Letfbest denote the function value fbest for t ,the existence of which we have shown above. Then let = ∞ − ∗ fbest f /2 (11)

Given the assumption ∞ > 0 and hence >0, it follows that

∗ ∗ n ∗ ={ ∈ R × A ||| n ∗ − |≤ } X (r,a ) f R ×{a }(r) f (12)

∗ is a nonempty set, where a denotes an optimal setting for a ∈ A for which f |Rn×{a} is ∗ | n regular. Then X is nonempty because of the assumption of regularity of f R ×{a} for any optimal a ∈ A. n Thus there exists a closed n-dimensional ball K ={r ∈ R ||r0 − r|≤ρ} with ρ>0 ∗ ∗ ∈ Rn ∗ ={ | ∈ }⊆ and center r0 that Ka (r,a ) r K X . Now, let us compute the probability of the event that the mutation of the discrete

and continuous variables yields a point in Ka∗ , given some arbitrary parent (r ,a), where a denotes the discrete part of a solution. This can be computed as the joint probability for the following two independent events:

46 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

1. The mutation of real vector generates r ∈ K. 2. The mutation of discrete part of the solution generates a∗.

This joint probability is lower bounded by: n 1 1 T p = min pa →a∗ min √ · exp (r − r0) · (r − r0) dr > 0 (13) a ∈A r ∈Rn 2 ∈ σ 2 0 2πσmin r K 2 min

∗ for a step size σmin > 0. Here pa →a∗ is the probability of obtaining a by one mutation of discrete parameters, which is larger than 0. Now we can derive a lower bound for the probability that Ka∗ is hit at least once after q generations as (where (q − 1)λ ≤ t 0. In other words, any vector with a distance t > 0 will be improved as t →∞ with probability one. 

4 Empirical Studies on Artificial Landscapes In this section, empirical results demonstrate the convergence behavior of the proposed MIES algorithm. First, test results for scalable theoretical test functions are illustrated. This gives the reader the opportunity to compare results of this EA with that of other optimization algorithms, for instance, with standard ES. Since rather simple test func- tions have been chosen, the experiments give hints on the minimal running time for practical problems of similar dimension. The first problem—the barrier functions (Li, Emmerich, Bovenkamp, et al., 2006)—introduce multimodality in a gradual way, as for the second test problem correlation/interaction between variables of different types can be introduced in a gradual way.

4.1 Barrier Functions We designed a multimodal barrier problem generator that produces integer optimiza- tion problems with a scalable degree of ruggedness (determined by parameter C)by generating an integer array A using Algorithm 6.

Algorithm 6 A[i] = i, i = 0,...,19 for k ∈{1,...,C} do j ← uniform random number out of {0,...,18} swap values of A[j]andA[j + 1] end for

Evolutionary Computation Volume 21, Number 1 47

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Figure 8: Surface plots of the barrier test functions for two integer variables Z1 and Z2, with control parameter C = 20, 100, and 500. All other variables were kept constant at a value of zero. Z1 and Z2 values varied in the range from 0 to 19.

For C = 0 the ordering of the variable y ∈ [0, 19] values corresponds to the ordering of values of A(y). If the value of C is slightly increased, still part of the order will be preserved under the mapping A, and thus similarity information can be exploited. Then a barrier function is computed:

nr nz nd 2 2 2 fbarrier(r, z, d) = A[ ri ] + A[zi ] + Bi[di ] → min (16) i=1 i=1 i=1

n nr nr = nz = nd = 5, r ∈ [0, 19] r ⊂ R , z ∈ [0, 19]nz ⊂ Znz , d ∈{0,...,19}nd ⊂ Dnd . (17)

Here, Bi (i = 1,...,nd ) denotes a set of i permutations of the sequence 0,...,19, each of which being a random permutation fixed before the run. This construction prevents the value of the nominal value di from being quantitatively (counter)correlated with the value of the objective function f . Such a correlation would contradict the assumption that di are nominal values. Whenever a correlation between neighboring values can be assumed, it is wiser to assign them to the ordinal type and treat them accordingly. The parameter C controls the ruggedness of the resulting function with regard to the integer space. Higher values of C result in more rugged landscapes with many barriers. To get an intuition about the influence of C on the geometry of the function, we included plots for a two-variable instantiation of the barrier function in Figure 8 for C = 20, 100, and 500. Intuitively, barrier functions with a higher control parameter C may have many local optima and a search procedure can easily get trapped by them. As we can see from these plots, when the control parameter C increases, the landscape becomes more rugged. For instance, the landscape of C = 500 shows many more barriers compared to C = 20. 4.1.1 Results Suggested by our former studies (Emmerich et al., 2000; Li, Emmerich, Bovenkamp, et al., 2006; Li, Emmerich, Eggermont, and Bovenkamp, 2006), the following MIES settings are chosen for the experiments on the barrier problems: (μ = 4,λ= 28) for the 6 population and offspring sizes, and (nσ = nς = np = 1) for the step size mode. Step

6 In contrast, (nσ = nr ,nς = nz,np = nd )representsn step size mode.

48 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Figure 9: Comparison between average best fitness values found by (4, 28) MIES and (4, 28) ES for C = 20, 100, 300, 500, and 1000.

sizes were initialized to 10% of the interval width, and probabilities were initialized to 0.1. Since ESs are stochastic algorithms, in the empirical experiments we created 10 instantiations7 for each control parameter C, and for each of them we let the algorithm perform 20 repeated runs (in total 20 × 10 = 200 runs for each value of C). Figure 9 shows average best fitness values found by one step size (4,28) MIES and one step size (4, 28) ES8 on barrier functions with different control parameters C, respectively. As we can see, it is more difficult for both MIES and ES to find the global optimum on barrier functions with a higher C value. This observation supports our finding from Figure 8: the landscape with a larger C value is more rugged and it is more challenging to tackle. Based on our algorithm design in Section 3, MIES is supposed to be more efficient for exploring the mixed integer landscapes compared to a standard ES. The corresponding box plot for best fitness values found by both MIES and ES in the last generation (=100) is shown in Figure 10. When C = 20 or 100, ES performs slightly better than MIES (p = 0.245 × 10-6 and p = 0.145 × 10-4, respectively). An explanation could be that constructed landscapes with C = 20 or 100 are still simple, which gives standard ES a chance to fully explore the search space. However, in the case of higher C values, MIES shows the advantage over standard ES. For C = 300, none of the strategies outperforms the other. For C = 500 and C = 1, 000 the overall performance of MIES is significantly better than that of the standard ES (p = 2.21 × 10-9 and p = 7.64 × 10-19, respectively). The p values are based on a left-sided t-test with α = 0.05. 4.2 Mixed Integer NK Landscapes Mixed integer NK landscapes are introduced as test problems for which correlation be- tween variables of different types can be introduced in a controlled way. NK landscapes

7By using different random seeds. 8Here, all parameters are evolved as continuous variables.

Evolutionary Computation Volume 21, Number 1 49

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Figure 10: Notched box plot for fitness value (generation = 100) of barrier functions with different control parameter C = 20, 100, 300, 500, and 1000 by using standard ES, MIES, and random search.

(NKL, also referred to as NK fitness landscapes), introduced by Kauffman (Kauffman, 1987), were devised to explore the way that epistasis9 controls the ruggedness of an adaptive landscape. They are particularly useful as test problem generators for genetic algorithms (GAs) to understand the dynamics of evolutionary search. The ruggedness and the degree of interaction between variables of NKL can be easily controlled by two tunable parameters: the number of genes N and the number of epistatic links of each gene to K other genes. Moreover, for given values of N and K, a large number of NK landscapes can be created at random. Mixed-integer NK-landscapes (MI-NKL) were introduced in Li, Emmerich, Egger- mont, Bovenkamp, et al. (2006), and are an extension of NKL from the traditional binary case to a mixed variable case with continuous, nominal discrete, and integer variable types. The resulting test function generator is a suitable test model for our mixed in- teger evolution strategy. For the continuous and integer variables, the binary NKL is extended to an N-dimensional hypercube [0, 1]N . Therefore, all continuous and integer variables are normalized between [0, 1]. Whenever the continuous or integer variable takes values at the corners of the hypercube, the value of the corresponding binary NKL is returned. For values located in the interior of the hypercube or its delimiting hyperplanes, we employ a multilinear interpolation technique that achieves a continu- ous interpolation between the function values at the corners. Note that a higher order approach is also possible but a multilinear approach was chosen for simplicity and ease of programming. Moreover, the theory of multilinear models introduces intuitive notions for the effect of single variables and interaction between multiple variables of potentially different types. To introduce nominal discrete variables in an appropriate manner, a more radical change to the NKL model is needed. In this case it is not feasible to use interpolation, as

9The interaction between genes.

50 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

(Xd=0, Xi=0.4, Xr=0.8)Xi Xr Fr

Fr(1, 0)=0.7 Fr(1, 1)=0.5 Fi

Xi + F Fr(0.4, 0.8)

Fd Xd Fr(0, 0)=0.8 Fr(0, 1)=0.7 Xr

Figure 11: Example for the computation of an MI-NK landscape.

Figure 12: Example epistasis matrix (left) and fitness matrix (right).

this would imply some inherent neighborhood defined on a single variable’s domain ∈{ i i } = xi d1,...,dL , i 1,...,N, which, by definition, is not given for the nominal discrete case. Therefore, the NKL is extended from a binary to an n-ary case to take into account the special characteristics of nominal discrete variables. Initial experiments demonstrate the applicability of the mixed integer NKL problem generator. The results obtained with a mixed integer ES indicate the increase in difficulty for finding the global optimum as K grows. Instead of describing the mixed variable case in a formal manner we give an illus- trating example (cf. Figure 11). This example shows one individual with three parame- ters (one continuous, one integer, and one discrete), and each gene interacts with both other genes. We assume there are three levels for the discrete gene Xd (L = 3), so the hypercube is reduced to three parallel planes, and the value of the discrete gene decides which plane is chosen. More concretely, assuming the individual has the following values: Xd = 0,Xi = 0.4,Xr = 0.8, the value of the discrete parameter Xd determines which square is chosen (Xd = 0). The value for each corner is based on the fitness matrix in Figure 12 (bold displayed). As mentioned in the previous section, we calculate the fitness value of this individual as follows:

Fr (a, x) = a0 + a1Xr + a2Xi + a3Xi Xr

a0 = Fr (0, 0) = 0.8,a1 = Fr (0, 1) − a0 =−0.1

a2 = Fr (1, 0) − a0 =−0.1,a3 = Fr (1, 1) − a0 − a1 − a2 =−0.1

Fr (0.4, 0.8) = 0.648 4.2.1 Experimental Results on MINKL We choose a population size μ of 4, offspring size λ of 28, and comma strategy for our experiments. The maximum number of generations is set to 100. Similar to experiments for barrier functions, in order to evaluate algorithm performance, we generated 10 problem instantiations for each K ∈{1, 3, 5, 10, 14} so that it is still feasible to find the

Evolutionary Computation Volume 21, Number 1 51

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Figure 13: The error averaged over 10 mixed integer NK landscape problems with N = 15 by using standard ES and MIES. Each problem contained five continuous, five integer-valued, and five Boolean-valued variables.

global optimum by evaluating all bit strings of length N = 15. Each generated problem consists of five continuous (Nr = 5), five integer (Nz = 5), and five nominal discrete (Nd = 5) variables. The continuous variables are in the range [–10,10], the integer valued variables are also in the range [–10,10], and we used {0, 1} for the nominal discrete variables (Booleans). We ran both MIES and ES 20 times on each problem instance. To compare the results of the different experiments we define the following error measure:

error = best found fitness – best possible fitness

The results are displayed in Figure 13. The x axis shows the number of generations while the y axis shows the average error (over all experiments). As can be seen, for both algorithms an increase in K results in an increase in error, which indicates the problem difficulty increases with K. The box plot for the last generation’s errors of both MIES and ES can be seen in Figure 14. There, overall performance between standard ES and MIES algorithms is com- pared. As can be seen, in the case of K = 1, 3, 5 the MIES obtained significantly better results than the standard ES (p = 2.74 × 10-7, p = 1.19 × 10-6,andp = 7.23 × 10-4,re- spectively). For K = 10 and K = 14 none of the two strategies is significantly better. As K = 14 means that each variable is connected to all other variables, the search space becomes extremely complex and it becomes too hard for both algorithms. In this cir- cumstance, MIES and ES show almost the same performance. Again, a left-sided t-test was employed for the comparisons.

4.3 Applications to MINLP Problems As we emphasized from the very beginning of this article, the MIES is particularly capable of yielding good solutions to black box mixed integer parameter optimization problems in high-dimensional spaces. The experimental results, which are from both synthetic test problems (Section 4) and real-world applications (Section 5), do underpin the argument. In this section, we will show that by utilizing the constraint handling

52 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

0.25

0.2

0.15

0.1 Function Value

0.05

0

ES MIES ES MIES ES MIES ES MIES ES MIES K=1, L=2 K=3, L=2 K=5, L=2 K=10, L=2 K=14, L=2

Figure 14: Notched box plot for fitness value (generation = 100) of mixed integer NK landscape problems with N = 15 by using standard ES and MIES. Each problem con- tained five continuous, five integer-valued, and five Boolean-valued variables.

techniques such as penalty function method, MIES can also produce good results on some classical MINLP problems. 4.3.1 Penalty Function Method The standard mixed integer optimization problem can be generalized by using Equa- tion (1) (in Section 2).10 In practice, the penalty function method has been widely used to assist evolutionary algorithms in dealing with constraint optimization problems. In- stead of using the original objective function to evaluate individuals in a population, a penalty term is devised and included in the evaluation phase. As an immediate result of this modification, the search landscape is changed. An individual that is a good candi- date according to the original objective function might become a bad solution because of the penalty function. In this work, we adopt the penalty function as defined in Joines and Houck (1994),

ψ(x) = f (x) + r(t)φ(x) (18) with r(t) = (Ct)β m n α α φ(x) = max{0,gi (x)} + | hj (x) | i=1 j=1 where f (x) is the original objective function, φ(x) stands for the penalty function re- garding the given constraints, and r(t) should be treated as coefficients of the φ(x). More specifically, t is the iteration/generation index and thus the r(t) is not constant but grows with t. Consequently, r(t) will bring more pressure on unfeasible solutions at the end of the search process. From this perspective, this method can be called the dynamic

10For some cases, equality constraints can be approximated by inequality constraints through a small positive number δ that determines the maximal deviation from 0.

Evolutionary Computation Volume 21, Number 1 53

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Table 1: Experimental results on benchmark functions with global competitive ranking scheme. P Function Best literature Median SD Gmedian Penalty ( (α, β),Pf ) × -13 P = f1 2.0000 2.0000 1.9870 10 10 (1, 1),Pf 0.45 × -15 P = f2 7.6672 7.6672 3.5616 10 31 (1, 1),Pf 0.45 × -15 P = f3 1.0765 1.0765 3.6293 10 32 (1, 1),Pf 0.45 × -13 P = f4 4.5796 4.5796 1.5132 10 34 (1, 1),Pf 0.45 − − × -1 P = f5 17.0000 17.0000 2.3570 10 29 (1, 1),Pf 0.45 × -1 P = f6 4.5796 4.5796 6.0510 10 188 (1, 2),Pf 0.40

. The penalty component ψ(x) is controlled by C, α,andβ. These pa- rameters are constants and independent of the total number of constraints. The choice of the combination of these parameters will decide the quality of the solution. For the experimental results presented below, P(α, β) will be used to indicate the value choice of α and β.

4.3.2 Global Competitive Ranking and Selection The introduction of the penalty function makes it necessary to design a new ranking and selection scheme. The ideal situation would be that the perfect balance between objective and penalty functions can be achieved through this selection scheme. Here, we chose the global competitive ranking mechanism, which was addressed in Runarsson and Yao (2003). In short, an individual is ranked two times—one is based on the objective function and the other follows the penalty function (the function φ in Equation (18). Then the final ranking is computed by using Equation (19), and selection is executed based on this final ranking.

πf (i) − 1 πφ(i) − 1 ψ(xi ) = Pf + (1 − Pf ) (19) λ − 1 λ − 1

where πf (i)andπφ(i) represent the ranking of an individual i based on the objective and penalty functions respectively. Pf denotes the probability that a comparison is done through the objective function only. For more detailed information, we recom- mend Runarsson and Yao (2003).

4.3.3 Experimental Results For our experiments, we chose six benchmark functions from the literature and applied (100, 700)-MIES to them. The details of these functions are given in the ap- pendix. For each of the benchmark functions, 200 independent runs (with a different random seed) are performed. The experimental results are presented in Table 1 as well as corresponding settings. The function name, the best feasible objective value from the literature, the median, standard deviation, and median number of gener- ations of each run to find the best solution are displayed. As one can see from Table 2, MIES with a global competitive ranking performed well for all test func- tions and can find the global optima in most cases. For functions f1−f5,weuseP(1, 1) and Pf = 0.45 for all experiments. But for function f6, we chose the different setting P(1, 2) and Pf = 0.40 . This is because compared to other functions, constraints of function f6 are more complex and the problem is more challenging. The statistic on Gmedian, the median convergence time to the global optimum, clearly indicates that it

54 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

is harder for the MIES to find the global optima for function f6 than for the other test functions. It is not the intention of this article to argue that MINLP methods should be replaced by MIES in case of explicitly given functions. In this case, efficient MINLP methods such as Branch and Bound (Dakin, 1965) are often available that, unlike the MIES, can guarantee the optimality of solutions. Rather, it is interesting to consider the MIES in cases where constraint functions are only defined implicitly, that is, the constraint functions must be treated as a black box function by the algorithm.

5 Real-World Applications As we addressed in Section 1, mixed integer parameter optimization problems occur quite frequently in practice. This section summarizes results for three typical parameter optimization tasks: intravascular ultrasound image analysis, multilayer optical coatings and chemical plant optimization. They all satisfy the requirements for mixed integer parameter optimization which is given by definition 1. We would like to mention some important characteristics of these practical applications as follows:

• The objective function is multimodal, high-dimensional, box-constrained in the continuous part, and with a of the search space that is characterized by implicit nonlinear constraints.

• The problem is treated as a black box optimization problem, due to the complexity of the evaluation procedure (e.g., simulator based evaluation with commercial simulators).

• Because of different parameter types, a standard representation such as binary strings or real-valued vectors is difficult to apply to these problems.

5.1 Intravascular Ultrasound Image Analysis Intravascular ultrasound (IVUS) is an advanced technique used to get real-time high resolution tomographic images from the inside of coronary vessels and other arteries. To gain insight into the status of an arterial segment, a so-called catheter pullback sequence is carried out. A catheter (φ ± 1 mm) with a miniaturized ultrasound transducer at the tip is inserted into a patient’s artery and positioned downstream of the segment of interest. The catheter is then pulled back in a controlled manner, using motorized pullback (1 mm/s), during which images are acquired continuously. IVUS provides real-time high resolution tomographic images of the coronary vessels wall, and is able to show the presence or absence of compensatory artery enlargement. IVUS allows precise tomographic measurement of the lumen area and plaque size, distribution, and, to some extent, composition of the plaque. An example of an IVUS image is shown in Figure 15. 5.1.1 Problem Definition In Bovenkamp et al. (2004, 2009) a state of the art multiagent system is developed to detect lumen, vessel, shadows, sidebranches, and calcified plaques in IVUS images. IVUS image processing agents interact with other agents through communication, act on the world by controlling and adapting image processing operations and perceive that same world by accessing image processing results. Agents thereby dynamically adapt the parameters of low-level image segmentation algorithms based on knowledge of global constraints, contextual knowledge, local image information, and personal beliefs.

Evolutionary Computation Volume 21, Number 1 55

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Figure 15: Intravascular ultrasound (IVUS) image with detected features.

Figure 16: IVUS lumen image processing pipeline.

The lumen agent, for example, encodes and controls an image processing pipeline (Figure 16) which includes binary morphological operations, an ellipse fitter, and a module, and it determines all relevant parameters. Generally, agent control allows the underlying segmentation algorithms to be simpler and to be applied to a wider range of problems with a higher reliability. Although the multiagent system has been shown to offer lumen and vessel detec- tion comparable to human experts (Bovenkamp et al., 2004), it is designed for symbolic reasoning, not numerical optimization. Further, it is almost impossible for a human ex- pert to completely specify how an agent should adjust its feature detection parameters in each and every possible interpretation context. As a result, an agent has only control knowledge for a limited number of contexts and a limited set of feature detector pa- rameters. In addition, this knowledge has to be updated whenever something changes in the image acquisition pipeline. Therefore, it would be much better if this knowl- edge could be acquired by learning the optimal parameters for different interpretation contexts automatically.

56 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Table 2: Parameters for the lumen feature detector. Name Type Range Dependencies Default Maxgray Integer [2, 150] >mingray 35 Mingray Integer [1, 149] scanindexleft 7 Centermethod Discrete {0,1} 1 Fitmodel Discrete {ellipse, circle} ellipse Sigma Continuous [0.5, 10.0] 0.8 Scantype Discrete {0,1,2} 0 Sidestep Integer [0, 20] 3 Sidecost Continuous [0.0, 100] 5 Nroflines Integer [32, 256] 128

5.1.2 Parameter Optimization for Lumen Feature Detector To determine whether or not MIES can be used as an optimizer for the parameters of feature detectors in the multiagent image analysis system, we applied MIES to the lumen feature detector (Li, Emmerich, Bovenkamp, et al., 2006; Li, Emmerich, Eggermont, et al., 2006). This detector was chosen because it can produce good results in isolation without additional information about side branches, shadows, plaques, and vessels. Table 2 contains the parameters for the lumen feature detector together with their type, range, dependencies, and the default settings determined by an expert. As can be seen, the parameters are a mix of continuous, integer, and nominal discrete (includ- ing Boolean) variables. The definition of the fitness function is crucially important for a successful application of MIES and should represent the success of the image seg- mentation very well. We used a dissimilarity measure D(c1,c2) between the contour c2 found by the lumen feature detector and the desired lumen contour c1 drawn by a 11 human expert. D(c1,c2) is defined as follows: nrofpoints d(c1(i),c2) iff d(c1(i),c2) >τ D(c1,c2) = θ(i), with θ(i) = (20) i=1 0, otherwise This measure penalizes each contour point which is more than a distance τ away from c2 and is proportional to the distance. The reason for allowing a small dissimilarity between the two contours is that even an expert physician will not draw the exact same contour twice.

5.1.3 Experimental Results For the experiments we used five disjoint sets of 40 images. The images were acquired with a 20 Mhz Endosonics Five64 catheter using motorized pullback (1 mm/s). The image size is 384 × 384 pixels (8 bit grayscale) with a pixel size of 0.02602 mm2. For

11 The variable c1 is also called ground truth.

Evolutionary Computation Volume 21, Number 1 57

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Table 3: Performance of the best found MIES and ES parameter solutions when trained on one of the five datasets (MIES solution i was trained on dataset i). All parameter solutions and the (default) expert parameters are applied to all datasets. Average differ- ence (fitness) and standard deviation w.r.t. expert drawn contours are given with τ = 2.24 and nrofpoints = 128.

Dataset 1 Dataset 2 Dataset 3 Dataset 4 Dataset 5 Fitness SD Fitness SD Fitness SD Fitness SD Fitness SD Expert 395.2 86.2 400.2 109.2 344.8 66.4 483.1 110.6 444.2 90.6 MIES 1 151.3 39.2 183.6 59.0 201.0 67.1 280.9 91.9 365.5 105.9 MIES 2 160.3 45.9 181.4 58.7 206.7 70.3 273.6 74.5 372.5 99.2 MIES 3 173.8 42.1 202.9 69.1 165.6 47.2 250.7 80.2 446.4 372.8 MIES 4 154.0 51.7 243.7 67.7 198.8 80.1 186.4 59.0 171.3 57.8 MIES 5 275.7 75.6 358.4 76.9 327.7 56.7 329.1 82.0 171.8 54.5 ES 1 149.8 46.0 185.1 60.1 203.1 66.6 270.2 69.8 358.1 100.1 ES 2 157.8 45.2 183.0 57.3 202.8 68.1 270.7 69.4 364.2 98.5 ES 3 212.6 46.3 243.3 72.3 182.3 49.7 252.9 58.4 756.8 734.4 ES 4 247.2 143.0 255.7 85.4 176.4 49.2 232.9 62.4 941.0 795.2 ES 5 217.0 51.6 225.3 68.5 292.8 73.6 330.7 78.0 324.7 96.0

Figure 17: Expert-drawn lumen contours (green/dark contour) compared with expert- set parameter solution (yellow/light contour, top row) and MIES parameter solu- tion (bottom row, yellow/light contour). A color version of this figure is available at http://www.mitpressjournals.org/doi/suppl/10.1162/EVCO a 00059.

the fitness function, we took the average dissimilarity over all 40 images with τ set to 2.24 pixels and nrofpoints = 128 (see Equation (20)). For each of the five datasets, we used (4, 28) MIES and ES algorithms and limited the number of iterations to 25, which resulted in 704 fitness evaluations for each dataset. The training results are displayed in Table 3, where MIES solution 1 was trained on dataset 1 by the MIES algorithm; MIES solution 2 was trained on dataset 2, and so on. As Table 3 shows, all MIES solutions performed better than or equal to the default expert solution. Furthermore MIES solution 4 performs slightly better on dataset 5 than MIES solution 5, which was trained on this dataset. However, a Student t-test with p = .05 shows that this difference is not statistically significant. Visual inspection of the results of the application of MIES parameter solution 4 to all 200 images shows that this solution is a good approximator of the lumen contours

58 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Figure 18: Multilayer optical coating (schematic).

as can be seen in Figure 17 (bottom row). When we compare the contours generated with MIES solution 4 to the expert-drawn contours, we see that they are very similar and in some cases the MIES contours actually seem to follow the lumen boundary more precisely. Apart from being closer to the expert-drawn contours, another major difference between the MIES-found contours and the contours detected with the default parameter settings is that the MIES solutions are more smooth (see Figure 17, top and bottom row).

5.2 Further Application of MIES Further applications of the MIES in practical problem domains are reported in the literature. Next, a brief summary of these is given. 5.2.1 Optimization of Multilayer Optical Coatings The first applications of MIES were reported in the domain of multilayer optical coating (MOC) design. In this first version of the MIES, nominal discrete parameters were used to represent the colors of the parallel layers of an optical filter and the width of these layers was represented by continuous variables (cf. Figure 18). The objective of MOC design is to find a sequence of layers of certain materials and certain thicknesses, such that all unwanted frequencies are being cut off, while the wanted frequencies pass without any reflection. Due to the high degree of nonlinearity of the objective function, it is difficult to find optimal solutions with classical methods, and the MIES produced competitive results in comparison to state of the art techniques in this domain (Back¨ and Schutz,¨ 1995). 5.2.2 Optimization of Chemical Engineering Plants A classical domain where mixed integer optimization is applied is the optimization of chemical engineering plants. MIES provides for optimization in combination with commercial steady-state flowsheet simulators (ASPEN PLUS; Emmerich et al., 2000; Henrich et al., 2007). Due to the black box character of objective function evaluations in this domain, classical mixed integer optimization techniques that require model equations of the unit operations are difficult to apply here, and MIES provides a more flexible and robust approach. Optimization parameters of a chemical engineering plant comprise continuous variables (pressure differences, temperatures), integer variables

Evolutionary Computation Volume 21, Number 1 59

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Split Ratio [0.0, 1.0] Hydrogen Recycle Material stream Purge Splitter Compressor Pump Conversion [0.3, 0.98] Toluene Temperature [895.2, 978.2] K Temperature [183.2,373.2] K

Pressure [1.0, 25.0] bar Hydrogen (Methane) Mixer Heater Furnace Reactor Flash Pressure [5.0, 15.0] bar Pump Cooler

Valve Hydrogen, Methane Condenser Utilities {BRINE, WATER} Toluene Benzene Recycle Valve Pressures [1.0, 5.0] bar Numbers of trays [10,100] Numbers of feed tray [5,50] Diphenyl

Reboiler Utilities {D15, D45, D150, D250, D600} Stabilizer-Column Toluene-Column Benzene-Column

Figure 19: Flowchart of a chemical engineering plant annotated with parameters to be optimized (Emmerich et al., 2000).

(e.g., number of stages of a distillation column), and nominal discrete variables (e.g., the type of coolant used in a condenser; cf. Figure 19). Successful applications of the MIES in chemical engineering plant optimization are reported for the optimization of a hydrodesalkylation (HDA) plant in Emmerich (1999), and for the economical optimiza- tion of non-sharp separation processes (Henrich et al., 2007).

6 Conclusions and Outlook In this article, we described MIESs and their application to real-world parameter opti- mization problems. As opposed to standard ESs, which are used for continuous optimization, MIESs are designed for the optimization of combinations of continuous, integer, and nominal discrete variables. For this purpose, MIESs use specific mutation operators for the differ- ent types of decision variables that are chosen based on common EA-design guidelines such as symmetry, maximal entropy, and scalability. We prove that for regular functions, the MIES converges with probability one to the global optimum. Empirical studies show that the MIES is able to learn and simul- taneously track optimal step sizes and mutation rates for different parameter types, thereby maximizing its local progress. Studies on a weighted quadratic model show that for integer vectors and continuous vectors, the mutation step sizes can be learned independently and their ordering reflects the scaling of the variables. The only self-adaptation scheme that turned out to be problematic was the learning of multiple mutation rates for the nominal discrete case. Future research should further investigate this problem. However, the capability of the MIES to learn a single, global mutation rate for nominal parameters is confirmed by our results. To test the algorithm’s performance and determine favorable default settings, we applied the MIES to several mixed integer benchmark problems, including problems with constraints. Moreover, we compared MIES to the standard (continuous) ES using the simple truncation of continuous variables. It turns out that the MIES approach has a much higher convergence reliability. MIES have also been applied to several real-world applications: medical image pro- cessing, multilayer optical coatings and chemical plant optimization. The results show

60 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

that the MIES is a valuable technique for high-dimensional mixed integer optimization problems. It is interesting in particular for black box problems that cannot be solved with standard mathematical programming techniques. A promising direction of future research will be to evaluate approaches that enable us to integrate more problem knowledge, such as exploiting dependencies between different variables. For a first step in this direction, see Emmerich et al. (2008), where heterogeneous Bayesian networks for representing mutation probability distributions are proposed. Moreover, extensions of the approach, for example, for multi-objective, robust, or dynamic optimization, will be of interest. In this context, we also refer to recent extensions to multimodal optimization using niching (Li, Eggermont, et al., 2008) and for optimization with expensive evaluations by using mixed integer metamodels in Li, Emmerich, et al. (2008).

Acknowledgments This research was partially financed by the Netherlands Organization for Scientific Research (NWO) under project 635.100.006 SAVAGE. Michael Emmerich acknowledges financial support by the DELiVER project, STW. The authors acknowledge the support of Andre H. Deutz checking the mathematical notations and theorems. We also like to thank the reviewers for their valuable comments.

References Back,¨ T. (1996). Evolutionary algorithms in theory and practice. Oxford, UK: Oxford University Press.

Back,¨ T., and Schutz,¨ M. (1995). Evolution strategies for mixed-integer optimization of optical multilayer systems. In Proceedings of the Fourth Annual Conference on Evolutionary Program- ming, pp. 33–51. Beyer, H.-G. (2001). Theory of evolution strategies. Berlin: Springer.

Beyer, H.-G. (2007). Evolution strategies. Scholarpedia, 2(8):1965. Beyer, H.-G., Schwefel, H.-P., and Wegener, I. (2002). How to analyse evolutionary algorithms. Theoretical Computer Science, 287(1):101–130.

Borchers, B., and Mitchell, J. (1994). An improved branch and bound algorithm for mixed integer nonlinear programs. Computers and Operations Research, 21(4):359–367. Born, J. (1978). Evolutionsstrategien zur numerischen Losung¨ von Adaptationsaufgaben. Dissertation, Humboldt Universitat,¨ Berlin.

Bovenkamp, E., Dijkstra, J., Bosch, J., and Reiber, J. (2004). Multi-agent segmentation of IVUS images. , 37(4):647–663. Bovenkamp, E., Dijkstra, J., Bosch, J., and Reiber, J. (2009). User-agent cooperation in multi-agent IVUS image segmentation. IEEE Transactions on Medical Imaging, 28(1):94–105.

Bovenkamp, E., Eggermont, J., Li, R., Emmerich, M., Back,¨ T., Dijkstra, J., and Reiber, J. (2006). Op- timizing IVUS lumen segmentations using evolutionary algorithms. In The First International Workshop on for IntraVascular and IntraCardiac Imaging, pp. 74–81. Bussieck, M., and Pruessner, A. (2003). Mixed-integer nonlinear programming. SIAG/OPT Views- and-News, 14(1):19–22.

Dakin, R. (1965). A - for mixed integer programming problems. The Computer Journal, 8:250–255.

Evolutionary Computation Volume 21, Number 1 61

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

Droste, S., and Wiesmann, D. (2000). Metric based evolutionary algorithms. In EuroGP. Lecture Notes in Computer Science, Vol. 1802 (pp. 29–43). Berlin: Springer.

Duran, M., and Grossmann, I. (1986). An outer- for a class of mixed- integer nonlinear programs. Mathematical Programming, 36(3):307–339.

Emmerich, M. (1999). Optimierung verfahrenstechnischer Prozeßstrukruren mit Evolutionaren¨ Algorithmen. Technical report, Department of Computer Science, University of Dortmund.

Emmerich, M., Grotzner,¨ M., Groß, B., and Schutz,¨ M. (2000). Mixed-integer evolution strategy for chemical plant optimization with simulators. In I. Parmee (Ed.), Evolutionary Design and Manufacture—Selected Papers from ACDM (pp. 55–67). Berlin: Springer. Emmerich, M., Li, R., Zhang, A., Flesch, I., and Lucas, P. (2008). Mixed-integer Bayesian opti- mization utilizing a priori knowledge on parameter dependences. In Proceedings of the 20th Belgium–Netherlands Conference on Artificial Intelligence (BNAIC), pp. 65–72.

Floudas, C. (1995a). Nonlinear and mixed-integer optimization. Oxford, UK: Oxford University Press. Floudas, C. (1995b). Nonlinear and mixed-integer optimization fundamentals and applications. Oxford, UK: Oxford University Press.

Geoffrion, A. (1972). Generalized benders decomposition. Journal of Optimization Theory and Applications, 10(4):237–260. Groß, B. (1999). Gesamtoptimierung verfahrenstechnischer Systeme mit Evolutionaren¨ Algorithmen. Ph.D. Dissertation, Rheinisch—Westfalische¨ Technische Hochschule Aachen, Lehrstuhl fur¨ Technische Thermodynamik.

Henrich, F., Bouvy, C., Kausch, C., Lucas, K., Preuß, M., Rudolph, G., and Roosen, P. (2007). Economic optimization of non-sharp separation sequences by means of evolutionary algo- rithms. Computers and Chemical Engineering, 7(32):1411–1432. Joines, J., and Houck, C. (1994). On the use of non-stationary penalty functions to solve nonlinear problems with GA’s. In Proceedings of the Evolutionary Computation Conference, pp. 579–584.

Kauffman, S. (1987). Towards a general theory of adaptive walks on rugged landscapes. Journal of Theoretical Biology, 128(1):11–45. Kocis, G., and Grossman, I. (1988). Global optimization of nonconvex mixed-integer nonlinear programming (MINLP) problems in process synthesis. Industrial & Engineering Chemistry Research, 27(8):1407.

Li, R., Eggermont, J., Shir, O., Emmerich, M., Back,¨ T., Dijkstra, J., and Reiber, J. (2008). Mixed- integer evolution strategies with dynamic niching. In PPSN X Proceedings. Lecture Notes in Computer Science, Vol. 5199, pp. 246–255. Berlin: Springer. Li, R., Emmerich, M., Bovenkamp, E., Eggermont, J., Back,¨ T., Dijkstra, J., and Reiber, J. (2006). Mixed-integer evolution strategies and their application to intravascular ultrasound image analysis. In Proceedings of EvoWorkshops 2006 (Track: EvoIASP). Lecture Notes in Computer Science, Vol. 3907, pp. 415–426. Berlin: Springer. Li, R., Emmerich, M., Eggermont, J., and Bovenkamp, E. (2006). Mixed-integer optimization of coronary vessel image analysis using evolution strategies. In M. Cattolico (Ed.), Proceedings of the Genetic and Evolutionary Computation Conference, GECCO 2006, pp. 1645–1652.

Li, R., Emmerich, M., Eggermont, J., Bovenkamp, E., Back,¨ T., Dijkstra, J., and Reiber, J. (2006). Mixed-integer NK landscapes. In Proceedings of the 9th International Conference of Parallel Problem Solving from Nature, PPSN IX. Lecture Notes in Computer Science, Vol. 4193, pp. 42–51. Berlin: Springer.

62 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 Mixed Integer Evolution Strategies for Parameter Optimization

Li, R., Emmerich, M., Bovenkamp, E., Eggermont, J., Back,¨ T., Dijkstra, J., and Reiber, J. (2008). Metamodel-assisted mixed integer evolution strategies and their application to intravascular ultrasound image analysis. In Proceedings of the Conference on IEEE-CEC 2008.

Lohmann, R. (1992). Structure evolution and incomplete induction. In Proceedings of Parallel Problem Solving from Nature 2, PPSN-II, pp. 177–188. Oyman, A. I., Beyer, H.-G., and Schwefel, H.-P. (2000). Analysis of the (1, lambda)-ES on the parabolic ridge. Evolutionary Computation, 8(3):249–265.

Porn, R., Harjunkaski, I., and Westerlund, T. (1999). Convexification of different classes of non- convex MINLP problems. Computers and Chemical Engineering, 23(3):439–448. Rechenberg, I. (1994). Evolutionsstrategie. Stuttgart: Frommann-Holzboog.

Rudolph, G. (1994). An for integer programming. In PPSN III, International Conference on Evolutionary Computation. The Proceedings of the Third Conference on Parallel Problem Solving from Nature, Lecture Notes in Computer Science, Vol. 866, (pp. 139–148). Berlin: Springer. Rudolph, G. (1997). Convergence properties of evolutionary algorithms. Schriftenreihe Forschung- sergebnisse zur Informatik. Hamburg: Kovac.

Runarsson, T., and Yao, X. (2003). Evolutionary optimization, Vol. 48. Constrained evolutionary optimization—The penalty function approach (pp. 87–113). Berlin: Springer. Schutz,¨ M. (1994). Eine Evolutionsstrategie fur¨ gemischt-ganzzahlige Optimierungsprobleme mit variabler Dimension. Master’s thesis, FB Informatik, University Dortmund, Germany.

Schutz,¨ M., and Sprave, J. (1996). Application of parallel mixed-integer evolution strategies with mutation rate pooling. In Proceedings of , EP’96, pp. 345–354. Schwefel, H.-P. (1995). Evolution and optimum seeking. Sixth-Generation Computer Technology Series. New York: Wiley Press.

Utecht, U., and Trint, K. (1994). Mutation operators for structure evolution of neural networks. In PPSN III, International Conference on Evolutionary Computation. The Third Conference on Parallel Problem Solving from Nature, Lecture Notes in Computer Science, Vol. 866, (pp. 492–501). Berlin: Springer. Wolsey, L., and Nemhauser, G. (1999). Integer and combinatorial optimization. New York: Wiley Press.

Yuan, X., Zhang, S., and Pibouleau, L. (1988). A mixed-integer nonlinear-programming method for process design. Operations Research, 22(4):331–346.

Appendix: Test Function Suite To test the MIES algorithm on mixed-integer nonlinear programming (MINLP) prob- lems we tested it on the following MINLP problems from literature:

1. f1 = 2r1 + d1 → min (Kocis and Grossman, 1988) − 2 − ≤ + ≤ subject to: 1.25 r1 d1 0,r1 d1 1.6, with r1 ∈ [0, 1.6],d1 ∈{0, 1}.

2. f2 = 2r1 + 3r2 + 1.5d1 + 2d2 − 0.5d3 → min (Kocis and Grossman, 1988) 2 + = 1.5 + = subject to: r1 d1 1.25,r2 1.5d2 3, r1 + d1 ≤ 1.6, 1.333r2 + d2 ≤ 3, −d1 − d2 + d3, with r1,2 ∈ [1, 10],d1,2,3 ∈{0, 1}.

Evolutionary Computation Volume 21, Number 1 63

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021 R. Li et al.

2 3. f3 = 0.8 + 5(r1 − 0.5) − 0.7d1 → min (Floudas, 1995b) subject to: − exp(r1 − 0.2) − r2 ≤ 0,r2 + 1.1d1 ≤−1,r1 − 1.2d1 ≤ 0, with r1 ∈ [0.2, 1],r2 ∈ [−2.22554, −1],d1 ∈{0, 1}.

2 2 2 4. f4 = 6 + (r1 − 1) − (r2 − 2) − (r3 − 3) − d1 − 3d2 − d3 −0.693147180559945d4 → min (Yuan et al., 1988) + + + + + ≤ 2 + 2 + 2 + ≤ subject to: r1 r2 r3 d1 d2 d3 5,r1 r2 r3 d3 5.5, r1 + d1 ≤ 1.2,r2 + d2 ≤ 1.8,r3 + d3 ≤ 2.5,r1 + d4 ≤ 1.2, 2 + ≤ 2 + ≤ 2 + ≤ r2 d2 1.64,r3 d3 4.25,r3 d2 4.64, with r1,2,3 ∈ [0, 10],d1,2,3,4 ∈{0, 1}.

5. f5 =−5r1 + 3r2 → min (Porn et al., 1999) − 0.5 2 + + 2 − 0.5 ≤ − ≤ subject to: 8r1 2r1 r2 11r2 2r2 2r2 39,r1 r2 3, 3r1 + 2r2 ≤ 24,r2 − d1 − 2d2 − 4d3 = 1,d2 + d3 ≤ 1, with r1 ∈ [1, 10],r2 ∈ [1, 6],d1,2,3 ∈{0, 1}.

2 2 2 2 2 6. f6 = (r4 − 1) − (r5 − 2) − (r6 − 1) + log(1 + r7) − (r1 − 1) − (r2 − 2) 2 − (r3 − 3) → min (Yuan et al., 1988) + + + + + ≤ 2 + 2 + 2 + 2 ≤ subject to: r1 r2 r3 d1 d2 d3 5,r6 r1 r2 r3 5.5, r1 + d1 ≤ 1.2,r2 + d2 ≤ 1.8,r3 + d3 ≤ 2.5,r1 + d4 ≤ 1.2, 2 + 2 ≤ 2 + 2 ≤ 2 + 2 ≤ r5 r2 1.64,r6 r3 4.25,r5 r3 4.64, r4 − d1 = 0,r5 − d2 = 0,r6 − d3 = 0,r7 − d4 = 0, with r1,2,3 ∈ [0, 10],r4,5,6,7 ∈ [0, 1],d1,2,3,4 ∈{0, 1}.

64 Evolutionary Computation Volume 21, Number 1

Downloaded from http://www.mitpressjournals.org/doi/pdf/10.1162/EVCO_a_00059 by guest on 26 September 2021