Deb02 PDF GZ
Deb02 PDF GZ
net/publication/11019773
CITATIONS READS
665 2,386
3 authors:
Dhiraj Joshi
FX Palo Alto Laboratory
50 PUBLICATIONS 7,961 CITATIONS
SEE PROFILE
All content following this page was uploaded by Ashish Anand on 19 May 2014.
Abstract
Due to increasing interest in solving real-world optimization problems using evo-
lutionary algorithms (EAs), researchers have recently developed a number of real-
parameter genetic algorithms (GAs). In these studies, the main research effort is spent
on developing an efficient recombination operator. Such recombination operators use
probability distributions around the parent solutions to create an offspring. Some op-
erators emphasize solutions at the center of mass of parents and some around the par-
ents. In this paper, we propose a generic parent-centric recombination operator (PCX)
and a steady-state, elite-preserving, scalable, and computationally fast population-
alteration model (we call the G3 model). The performance of the G3 model with the
PCX operator is investigated on three commonly used test problems and is compared
with a number of evolutionary and classical optimization algorithms including other
real-parameter GAs with the unimodal normal distribution crossover (UNDX) and the
simplex crossover (SPX) operators, the correlated self-adaptive evolution strategy, the
covariance matrix adaptation evolution strategy (CMA-ES), the differential evolution
technique, and the quasi-Newton method. The proposed approach is found to consis-
tently and reliably perform better than all other methods used in the study. A scale-up
study with problem sizes up to 500 variables shows a polynomial computational com-
plexity of the proposed approach. This extensive study clearly demonstrates the power
of the proposed technique in tackling real-parameter optimization problems.
Keywords
Real-parameter optimization, simulated binary crossover, self-adaptive evolution
strategy, covariance matrix adaptation, differential evolution, quasi-Newton method,
parent-centric recombination, scalable evolutionary algorithms.
1 Introduction
Over the past few years, there has been a surge of studies related to real-parameter
genetic algorithms (GAs), despite the existence of specific real-parameter evolutionary
algorithms (EAs), such as evolution strategy and differential evolution. Although, in
principle, such real-parameter GA studies have been shown to have a similar theoret-
ical behavior on certain fitness landscapes with proper parameter tuning in an earlier
study (Beyer and Deb, 2001), in this paper we investigate the performance of a cou-
ple of popular real-parameter genetic algorithms and compare extensively with the
above-mentioned real-parameter EAs and with a commonly used classical optimiza-
tion method.
A good description of different real-parameter GA recombination operators can
be found in Herrara et al. (1998) and Deb (2001). Of different approaches, the unimodal
normal distribution crossover (UNDX) operator (Ono and Kobayashi, 1997), the simplex
crossover (SPX) operator (Higuchi et al., 2000), and the simulated binary crossover (SBX)
(Deb and Agrawal, 1995) are well studied. The UNDX operator uses multiple parents
and creates offspring solutions around the center of mass of these parents. A small
probability is assigned to solutions away from the center of mass. On the other hand,
the SPX operator assigns a uniform probability distribution for creating offspring in a
restricted search space around the region marked by the parents. These mean-centric
operators have been applied with a specific GA model (the minimum generation gap
(MGG) model suggested by Satoh et al. (1996)). The MGG model is a steady-state GA
in which 200 offspring solutions are created from a few parent solutions and only two
solutions are selected. Using this MGG model, a recent study (Higuchi et al., 2000)
compared both UNDX and SPX and observed that the SPX operator performs better
than the UNDX operator on a number of test problems. Since the SPX operator uses
a uniform probability distribution for creating an offspring, a large offspring pool size
(200 members) was necessary to find a useful progeny. On the other hand, the UNDX
operator uses a normal probability distribution to create an offspring, giving more em-
phasis to solutions close to the center of mass of the parents. Therefore, such a large
offspring pool size may not be optimal with the UNDX operator. Despite the exten-
sive use of these two recombination operators, we believe that adequate parametric
studies were not performed in any earlier study to establish the best parameter settings
for these GAs. In this paper, we perform a parametric study by varying the offspring
pool size and the overall population size and report interesting outcomes. For a fixed
number of function evaluations, the UNDX operator with a biased (normal) probability
distribution of creating offspring solutions around the centroid of parents works much
better with a small offspring pool size and outperforms the SPX operator, which uses a
uniform probability distribution over a simplex surrounding the parents.
Working with nonuniform probability distribution for creating offspring, it is not
intuitively clear whether biasing the centroidal region (mean-centric approach as in the
UNDX operator) or biasing the parental region (parent-centric approach as in the SBX
or fuzzy recombination) is a better approach. A previous study (Deb and Agrawal,
1995) has shown that the binary-coded GAs with the single-point crossover operator,
when applied to continuous search spaces, use an inherent probability distribution bi-
asing the parental region, rather than the centroidal region. Using variable-wise recom-
bination operators, a past study (Deb and Beyer, 2000) has clearly shown the advantage
of using a parent-centric operator (SBX) over a number of other recombination opera-
tors. Motivated by these studies, in this paper, we suggest a generic parent-centric
recombination (PCX) operator, which allows a large probability of creating a solution
near each parent, rather than near the centroid of the parents. In order to make the
MGG model computationally faster, we also suggest a generalized generation gap (G3)
model, which replaces the roulette-wheel selection operator of the MGG model with
a block selection operator. The proposed G3 model is a steady-state, elite-preserving,
and computationally fast algorithm for real-parameter optimization. The efficacy of the
G3 model with the proposed PCX operator is investigated by comparing it with UNDX
1. Population mean decision variable vector should remain the same before and after
the variation operator.
2. Variance of the intramember distances must increase due to the application of the
variation operator.
Since variation operators usually do not use any fitness function information explic-
itly, the first argument makes sense. The second argument comes from the realization
that the selection operator has a tendency to reduce the population variance. Thus,
where wi and vi are zero-mean normally distributed variables with variances σζ2 and
√
ση2 , respectively.
√ Kita and Yamamura (1999) suggested σζ = 1/ µ − 2 and ση =
0.35/ n − µ − 2, respectively and observed that µ = 3 to 7 performed well. It is inter-
esting to note that each offspring is created around the mean vector ~g. The probability
of creating an offspring diminishes with the distance from the mean vector, and the
maximum probability is assigned at the mean vector. Figure 1 shows three parents and
a few offspring created by the UNDX operator. The complexity of the above procedure
in creating one offspring is O(µ2 ), governed by the Gram-Schmidt orthonormalization
needed in the process.
The SPX operator also creates offspring uniformly around the √ mean but restricts
them within a predefined region (in a simplex similar but γ = µ + 1 times bigger
than the parent simplex). A distinguishing aspect of the SPX from the UNDX operator
is that the SPX assigns a uniform probability distribution for creating any solution in a
restricted region (called the simplex). Although, in its true sense, it is not a mean-centric
operator, because of its emphasis to solutions around the centroid of the participating
parents we put this operator in the category of the mean-centric operators. Figure 2
shows the density of solutions with three parents for the SPX operator. The compu-
tational complexity for creating one offspring here is O(µ), thereby making the SPX
operator faster than the UNDX operator.
µ
y = ~x(p) + wζ d~(p) +
X
~ wη D̄~e(i) , (2)
i=1, i6=p
where ~e(i) are the (µ − 1) orthonormal bases that span the subspace perpendicular to
d~(p) . Thus, the complexity of the PCX operator to create one offspring is O(µ), instead
of O(µ2 ) required for the UNDX operator. The parameters wζ and wη are zero-mean
normally distributed variables with variance σζ2 and ση2 , respectively. The important
distinction from the UNDX operator is that offspring solutions are centered around
each parent. The probability of creating an offspring closer to the parent is higher.
Figure 3 shows the distribution of offspring solutions with three parents. The moti-
vation of the PCX operator is as follows. Since individual parents have qualified the
“fitness test” in the selection operator, it can be assumed that solutions close to these
parents are also potential good candidates, particularly in the context of continuous
search space problems. On the contrary, it may be quite demanding to assume that the
solutions close to the centroid of the participating parents are also good, especially in
cases where parents are well sparsed in the search space. Creating solutions close to
previously found good solutions, as emphasized by the PCX operator, should make a
more reliable search. It is also intuitive that the convergence towards a local optimum
can be made faster by always choosing ~x(p) as the best parent.
100000
1
SPX
1e−05
1e−10
Best Fitness 1e−15
1e−20
1e−25
1e−30
1e−35
1e−40 UNDX
1e−45
1e−50
2 10 100 300
λ
Figure 4: Best fitness for different λ on Felp using the MGG model with SPX and UNDX
operators.
1e+10
1
SPX
1e−10
Best Fitness
1e−20
UNDX
1e−30
1e−40
1e−50
2 10 100 300
λ
Figure 5: Best fitness on Fsch using the MGG model with SPX and UNDX.
recombination operators and perform a parametric study with λ on three standard test
problems:
n
X
Felp = ix2i (ellipsoidal function) (3)
i=1
2
n
X i
X
Fsch = xj (Schwefel’s function) (4)
i=1 j=1
n−1
X
100(x2i − xi+1 )2 + (xi − 1)2 (generalized Rosenbrock’s function) (5)
Fros =
i=1
In all problems, we have used n = 20. The first two problems have their minimum at
x∗i = 0 with F ∗ = 0, and the third function has its minimum at x∗i = 1 with F ∗ = 0. In
order not to bias the search, we have initialized the population in xi ∈ [−10, −5] for all
i in all problems.
First, we fix N = 300 and vary λ from 2 to 300. All other parameters are kept as
they were used in the original study (Higuchi et al., 2000), except that in UNDX, µ = 6
is used, as this value was found to produce better results. In all experiments, we ran the
MGG model until a predefined number of function evaluations F T elapsed. We used
the following values of F T for different functions: Felp
T
= 0.5(106 ), Fsch
T
= 1(106 ), and
T 6
Fros = 1(10 ). In all experiments, 50 runs with different initial populations were taken
and the smallest, median, and highest best fitness values recorded. Figure 4 shows the
best fitness values obtained by the SPX and the UNDX operators on Felp with different
values of λ. The figure shows that λ = 50 produced the best reliable performance for
the SPX operator. Importantly, the MGG model with λ = 200 (which was suggested in
the original study) did not perform as well. Similarly, for the UNDX operator, the best
SPX
1e+10
SPX
100000
1e−10
1e−15
1e−20 UNDX
1e−25
1e−30
2 10 100 300
λ
Figure 6: Best fitness on Fros using the MGG model with SPX and UNDX.
1. From the population P , select the best parent and µ − 1 other parents randomly.
1e+06
SPX
13
Function Evaluations
9
UNDX
100000
PCX
10000
1000
2 4 10 20 50 100 300
λ
Figure 7: Function evaluations needed to find a solution of fitness 10−20 on Felp using
the G3 model with PCX, UNDX, and SPX. For PCX and UNDX operators, a population
size of 100 is used, and for the SPX operator, a population size of 300 is used.
The above G3 model can also be modified by choosing only one parent in Step 3 (in-
stead of two parents) and replacing the parent with the best of the combined subpopu-
lation of λ offspring and the parent. At first, we do not make this change and continue
with the above G3 model in order to keep the structure of the model similar to the MGG
model. Later, we shall investigate the efficacy of this modified G3 model.
1e+06
SPX
Function Evaluations
100000
UNDX
10000
PCX
1000
50 70 100 150 200 300 400 500
Population Size
Figure 8: Function evaluations to reach a fitness 10−20 versus population sizes on Felp
using the G3 model with PCX, UNDX, and SPX.
The number of runs (out of 50) where such a solution was found are marked on the
plot. For λ < 12, the SPX operator did not find the target solution in any of the 50 runs.
The best run of the SPX operator (with λ = 12) required 163,380 function evaluations.
The figure shows that with smaller offspring pool size (λ), better performance with
PCX and UNDX operators is achieved. Thus, we choose λ = 2 for these operators and
perform a parametric study with the population size. For the SPX operator, we use
λ = 15, below which satisfactory results were not found. Figure 8 shows that there
exists an optimum population size at which the performance is the best for PCX (with
5,744 function evaluations) and UNDX (with 15,914 function evaluations). For the 20-
variable ellipsoidal problem, N ∼ 100 seems to be optimum for these two operators.
It is also interesting to note that the optimal population size requirement for the PCX
operator is larger than that for the UNDX operator. However, the solution accuracy
achieved by the PCX operator is almost an order of magnitude better than that obtained
by the UNDX operator. Another interesting aspect is that for the SPX operator with
λ = 15, all runs with population size smaller than 300 did not find the desired solution.
Since SPX creates solutions within a fixed range proportional to the location of the
parents, its search power is limited. Moreover, since random samples are taken from a
wide region in the search space, the success of the algorithm relies on the presence of a
large population size.
Next, we apply the G3 model with all three recombination operators to Fsch . Fig-
ures 9 and 10 show the parametric studies with λ and the population size for the PCX
and the UNDX operators, respectively. Once again, N = 100 and λ = 2 are found
to perform the best for both operators. However, the PCX operator is able to find the
desired solution with a smaller number of function evaluations (14,643 for PCX ver-
sus 27,556 for UNDX). However, the SPX operator does not perform well at all on the
Schwefel’s function. The minimum function evaluations needed with any parameter
1e+06
Function Evaluations
UNDX
100000
PCX
10000
2 4 10 20 50 100 300
λ
Figure 9: Function evaluations needed to find a solution of fitness 10−20 on Fsch using
the G3 model with PCX and UNDX. N = 100 is used.
65000
60000
Function Evaluations
55000
50000 UNDX
45000
40000
35000
30000
25000 PCX
20000
15000
10000
50 70 100 200 300
Population Size
Figure 10: Function evaluations to reach a fitness 10−20 versus population sizes on Fsch
using the G3 model with PCX and UNDX. λ = 2 is used.
100000
SPX
Best Fitness
1e−05
1e−10 UNDX
1e−15 PCX
1e−20
0 50000 100000 150000 200000
Function Evaluations
Table 1: Three minima (or near minima) for the 20-variable Rosenbrock’s function.
f (~
x) xi , i = 1, 2, . . . , 20
0 1 1 1 1 1 1 1
1 1 1 1 1 1 1
1 1 1 1 1 1
3.986624 −0.993286 0.996651 0.998330 0.999168 0.999585 0.999793 0.999897
0.999949 0.999974 0.999987 0.999994 0.999997 0.999998 0.999999
0.999999 0.999999 0.999999 0.999997 0.999995 0.999989
65.025362 −0.010941 0.462100 0.707587 0.847722 0.922426 0.960915 0.980415
0.990213 0.995115 0.997563 0.998782 0.999387 0.999682 0.999818
0.999861 0.999834 0.999722 0.999471 0.998953 0.997907
setting of the SPX operator to find the desired solution is 414,350, which is an order of
magnitude more than the best results obtained using the PCX and the UNDX operators.
Thus, we have not presented any simulation result for the SPX operator.
Figure 11 shows the population best fitness values (of 50 runs) of Fsch with number
of function evaluations in the case of the G3 model with the best-performing parame-
ters for PCX (λ = 2 and N = 150), UNDX (λ = 2 and N = 70), and SPX (λ = 70 and
N = 300) operators. The figure shows the superiority of the PCX operator in achieving
a desired accuracy with the smallest number of function evaluations.
Next, we attempt to solve the Fros function. This function is more difficult to solve
than the previous two functions. Here, no implementation is able to find a solution
very close to the global optimum (with a fitness value 10−20 ) in all 50 runs each within
one million function evaluations. Runs with PCX and UNDX operators sometimes get
stuck to a local optimum solution. Interestingly, this function is claimed to be unimodal
in a number past of studies. However, for n > 3, this function has more than one min-
1e+06
40 44
42
UNDX 40
Function Evaluations
42 40
43 35
46
39
39 41 42
100000 44 38
43 PCX
39
38
42 43
38 33 43
36
10000
2 4 10 20 50 100 300
λ
Figure 12: Function evaluations needed to find a solution of fitness 10−20 for different
λ values on the Fros using the G3 model with PCX and UNDX operators.
imum, of which the solution xi = 1 (for all i) is the global minimum. For 20 variables,
we have identified three minima with function values 0, 3.98662, and 65.025362, re-
spectively. The corresponding variable values are tabulated in Table 1. In the figures
(to be described next) involving Fros , whenever a GA did not find the global minimum,
it was always attracted to the best local optimum (with the function value 3.98662). It
is also important to highlight that the SPX operator failed to find the required solution
in any of the 50 runs. However, Figures 12 and 13 show that the PCX operator (with a
minimum of 14,847 function evaluations) has performed much better than the UNDX
operator (with a minimum of 58,205 function evaluations). The number of times where
a run has converged near the global optimum is also marked in the figures. Once again,
a small λ (∼ 4) and an adequate population size (100 to 150) are found to produce op-
timum behavior for PCX and UNDX operators in this problem.
To compare the performance of the UNDX operator when applied with the MGG
and the G3 model, we compare the number of function evaluations needed to achieve a
solution with fitness 10−20 on all three test problems. In both cases, results are obtained
with their best parameter settings. The following table shows that the G3 model is an
order of magnitude better than the MGG model.
Function MGG G3
Felp 2,97,546 18,120
Fsch 5,03,838 30,568
Fros 9,38,544 73,380
The careful creation of new solutions near successful parents and the dependence of the
actual distance of new solutions from parents on the diversity of the parent population
allow the PCX operator to have a self-adaptive yet efficient search. The combination
of such a search operator along with a fast population alteration procedure makes the
overall approach much better than the previous real-parameter EAs.
1e+06
45
Function Evaluations
UNDX
40
42 38 43 36
100000 36 40
34
PCX
37 41 36
44 40
38 36
10000
50 70 100 150 200 300
Population Size
Figure 13: Function evaluations to reach a fitness 10−20 versus population sizes on Fros
using the G3 model with PCX and UNDX operators. For PCX and UNDX operators,
a population size of 100 is used, and for the SPX operator, a population size of 300 is
used.
Table 2: Comparison of the original and the modified G3 model on three test problems.
G3 Felp Fsch
model Best Median Worst Best Median Worst
Original 5,744 6,624 7,372 14,643 16,326 17,712
Modified G3 5,826 6,800 7,728 13,988 15,602 17,188
G3 Fros
model Best Median Worst
Original 14,847 22,368 25,797
Modified G3 16,508 21,452 25,520
Felp Fsch
EA Best Median Worst Best Median Worst
(1, 10)-ES 28,030 40,850 87,070 72,330 105,630 212,870
(15, 100)-ES 83,200 108,400 135,900 173,100 217,200 269,500
CMA-ES 8,064 8,472 8,868 15,096 15,672 16,464
Modified G3 5,826 6,800 7,728 13,988 15,602 17,188
Fros
EA Best Median Worst
(1, 10)-ES 591,400 803,800 997,500
(15, 100)-ES 663,760 837,840 936,930
CMA-ES 29,208 33,048 41,076
Modified G3 16,508 21,452 25,520
1e+06
PCX
1000 10000
5 10 50 100 150 200 300 5 10 50 100 150 200
Population Size Population Size
Figure 14: Function evaluations needed Figure 15: Function evaluations needed
with differential evolution for the ellip- with differential evolution for the
soidal problem. Schwefel’s function.
1e+07
Number of Function Evaluations
49
48 48
48 49
DE
1e+06 43
100000
40
PCX 30
3142 35 34 46
10000
20 50 100 150 200 300
Population Size
Figure 16: Function evaluations needed with differential evolution for the Rosenbrock’s
function.
are much better than those needed for the correlated self-adaptive ES, they are some-
what worse than those needed for the modified G3 model with the PCX operator. It is
worth mentioning that the CMA-ES requires setting of many parameters, some reason-
able values of which are suggested in the original study. However, in the case of the G3
model, there are two parameters ση and σζ (in Equation 2) that a user has to set. In all
studies here, we have used a constant value of 0.1 for these two parameters.
5 Differential Evolution
Differential evolution (DE) (Storn and Price, 1997) is a steady-state EA in which for every
offspring a set of three parent solutions and an index parent are chosen. Thereafter,
a new solution is created either by a variable-wise linear combination of three parent
solutions or by simply choosing the variable from the index parent with a probability.
The resulting new solution is compared with the index parent and the better of them is
declared as the offspring. We have used the C-code downloaded from the DE website
http://http.icsi.berkeley.edu/˜storn/ and used strategy 1 with all param-
eters set as suggested in the code. Moreover, this strategy is also observed to perform
better than other DE strategies in our limited experimental study.
Figures 14, 15, and 16 compare the function evaluations needed with DE to achieve
a solution with a function value of 10−20 on three test problems with the modified G3
model and the PCX operator. We have applied DE with different population sizes each
performed with 50 independent runs. It is interesting to note that in all three problems,
an optimal population size is observed. In the case of the Rosenbrock’s function, only
a few runs out of 50 runs find the true optimum with a population size smaller than
20. All three figures demonstrate that the best performance of the modified G3 with
the PCX operator is better than the best performance of the DE. To highlight the best
performances of both methods, we have tabulated corresponding function evaluations
in the following table:
Felp Fsch
Method Best Median Worst Best Median Worst
DE 9,660 12,033 20,881 102,000 119,170 185,590
Modified G3 5,826 6,800 7,728 13,988 15,602 17,188
Fros
Method Best Median Worst
DE 243,800 587,920 942,040
Modified G3 16,508 21,452 25,520
The table shows that except for the ellipsoidal function, the modified G3 requires an
order of magnitude less number of function evaluations than DE.
6 Quasi-Newton Method
The quasi-Newton method for unconstrained optimization is a popular and efficient
approach (Deb, 1995; Reklaitis et al., 1983). Here, we have used the BFGS quasi-Newton
method along with a mixed quadratic-cubic polynomial line search approach available
in MATLAB (Branch and Grace, 1996). The code computes gradients numerically and
adjusts step sizes in each iteration adaptively for a fast convergence to the minimum.
This method is found to produce the best performance among all optimization proce-
dures coded in MATLAB, including the steepest-descent approach.
In Table 4, we present the best, median, and worst function values obtained from
a set of 10 independent runs started from random solutions with xi ∈ [−10, −5]. The
maximum number of function evaluations allowed in each test problem is determined
Table 4: Solution accuracy obtained using the quasi-Newton method. FE denotes the
maximum allowed function evaluations. In each case, the G3 model with the PCX
operator finds a solution with a function value smaller than 10−20 .
1e+07
100000
1000
100
5 50 100 200 500
Number of Variables
Figure 17: Scale-up study for the ellipsoidal function using the modified G3 model and
the PCX operator.
from the best and worst function evaluations needed for the G3 model with the PCX
operator to achieve an accuracy of 10−20 . These limiting function evaluations are also
tabulated. The tolerances in the variables and in the function values are set to 10−30 .
The table shows that the quasi-Newton method has outperformed the G3 model with
PCX for the ellipsoidal function by achieving a better function value. Since the el-
lipsoidal function has no linkage among its variables, the performance of the quasi-
Newton search method is difficult to match with any other method. However, it is
clear from the table that the quasi-Newton method is not able to find the optimum
with an accuracy of 10−20 within the allowed number of function evaluations in more
epistatic problems (Schwefel’s and Rosenbrock’s functions).
1e+07
100000
slope = 1.71
10000
1000
100
5 50 100 200 500
Number of Variables
Figure 18: Scale-up study for the Schwefel’s function using the modified G3 model and
the PCX operator.
Figure 17 shows the experimental results for the ellipsoidal test problem having
as large as 500 variables. The use of two offspring (λ = 2) is found to be the best in
this case. The figure is plotted in a log-log scale. A straight line is fitted through the
experimental points. The slope of the straight line is found to be 1.88, meaning that
the function evaluations vary approximately polynomially as O(n1.88 ) over the entire
range of the problem size in n ∈ [5, 500].
Next, we do a similar scale-up study for the Schwefel’s function. Figure 18 shows
the outcome of the study with λ = 2. A similar curve-fitting finds that the number of
function evaluations required to obtain a solution with 10−10 varies as O(n1.71 ) over
the entire range n ∈ [5, 500] of problem size.
Finally, we apply the modified G3 with PCX on the Rosenbrock’s function. Be-
cause of the large number of function evaluations needed to solve a 500-variable Rosen-
brock’s function, we limit our study to a maximum of 200 variables. Figure 19 shows
the simulation results and the fitted straight line on a log-log plot. The function evalu-
ations needed to obtain a function value of 10−10 varies as O(n2 ). Interestingly, all the
above experiments on three test problems show that the modified G3 model with the
PCX operator needs a polynomially increasing number of function evaluations with an
increasing problem size.
The ellipsoidal and Schwefel’s functions have only one optimum, whereas the Rosen-
brock’s function is multimodal. To really investigate the performance of the G3 model
with PCX and UNDX operators on multimodal problems further, we have chosen the
1e+07
100000
slope = 2.00
10000
1000
5 50 100 200
Number of Variables
Figure 19: Scale-up study for the Rosenbrock’s function using the modified G3 model
and the PCX operator.
The global minimum function value is zero. Each integer corresponds to a local min-
imum. Unlike many past studies involving this function, we initialize the population
randomly at xi ∈ [−10, −5]. This initialization prevents a couple of important matters,
which are ignored in many past studies involving this function:
1. The initial population is away from the global basin, thereby making sure that an
algorithm must overcome a number of local minima to reach the global basin.
2. Such initialization prevents the advantage enjoyed by algorithms that have an in-
herent tendency to create solutions near the centroid of the parents.
Most past studies have initialized a population symmetrically about the global mini-
mum (such as by initializing xi ∈ [−5.12, 5.12]) to solve this function to global opti-
mality. We consider this unfair, as a mean-centric recombination of two solutions on
either side of xi = 0 may result in a solution close to xi = 0. Moreover, in most real-
world problems, the knowledge of the exact optimum is usually not available, and the
performance of an EA on a symmetric initialization may not represent the EA’s true
performance in solving the same problem with a different initialization or other prob-
lems.
Within the range [−10, −5] for each variable, there exist six local minima. In order
for an algorithm to reach the global minimum, it has to overcome four more local min-
ima for each variable. We have tried varying different G3 model parameters, such as
λ, population size and variances. For both PCX and UNDX operators, no solution in
the global basin is found in a maximum of one million function evaluations over mul-
tiple runs. From typical function values of the order of 103 , which exist in the initial
population, the G3 model with the PCX and UNDX operators finds best solutions with
function values equal to 15.936 and 19.899, respectively. Since these function values are
less than 20 × 1 (the best local minimum on each variable has a function value equal
to one) or 20, at least one variable (xi ∈ [−0.07157, 0.07157]) lies close to the global
optimum value of xi = 0. Although this itself is a substantial progress made by both
models despite the existence of many local minima, it would be interesting to investi-
gate if there exists a better global approach to solve this problem starting from an initial
population far away from the global optimum.
9 Conclusions
The ad-hoc use of a uniformly distributed probability for creating an offspring often
implemented in many real-parameter evolutionary algorithms (such as the use of SPX,
BLX, or arithmetic crossover) and the use of a mean-centric probability distribution
(such as that used in UNDX) have been found not to be as efficient as the proposed
parent-centric approach in this paper. The parent-centric recombination operator fa-
vors solutions close to parents, rather than the region close to the centroid of the par-
ents or any other region in the search space. Systematic studies on three 20-variable
test problems started with a skewed population not bracketing the global optimum
have shown that a parent-centric recombination is a meaningful and efficient way of
solving real-parameter optimization problems. In any case, the use of a uniform proba-
bility distribution for creating offspring has not been found to be efficient compared to
a biased probability distribution favoring the search region represented by the parent
solutions. Moreover, it has been observed that the offspring pool size used in previous
real-parameter EA studies is too large to be computationally efficient.
Furthermore, the use of an elite-preserving, steady-state, scalable, and computa-
tionally fast evolutionary model (named as the G3 model) has been found to be effec-
tive with both PCX and UNDX recombination operators. In all simulation runs, the G3
model with the PCX operator has outperformed both the UNDX and SPX operators in
terms of the number of function evaluations needed in achieving a desired accuracy.
The proposed PCX operator has also been found to be computationally faster. Unlike
most past studies on test problems, this study has stressed the importance of initializing
an EA run with a skewed population.
To further investigate the performance of the G3 model with the PCX operator, we
have compared the results with three other commonly used evolutionary algorithms:
correlated self-adaptive evolution strategy, covariance matrix adaptation (CMA-ES),
and differential evolution, and with a commonly used classical unconstrained opti-
mization method, the quasi-Newton algorithm, with the BFGS update method. Com-
pared to all these state-of-the-art optimization algorithms, the proposed model with
the PCX operator has consistently and reliably performed better (in some cases, more
than an order of magnitude better in terms of required function evaluations).
A scale-up study on three chosen problems over a wide range of problem sizes (up
to 500 variables) has demonstrated a polynomial (of a maximum degree of two) com-
putational complexity of the proposed approach. Compared to existing studies with a
number of different GAs and a number of different self-adaptive evolution strategies,
the proposed approach has shown better performance than most studies on the four
test problems studied here. However, compared to a CMA-based evolution strategy
approach (Hansen and Ostermeier, 2000), which involves an update of O(n2 ) strategy
parameters in each iteration, our G3 approach with a computationally efficient proce-
dure of creating offspring (with the PCX operator) and a simple parent replacement
strategy has shown a similar scale-up effect to all test problems studied here. This
study clearly demonstrates that the search power involved in the strategy-parameter
based self-adaptive evolution strategies (CMA-ES) in solving real-parameter optimiza-
tion problems can be matched with an equivalent evolutionary algorithm (G3 model)
without explicitly updating any strategy parameter, but with an adaptive recombina-
tion operator. A similar outcome was also reported elsewhere (Deb and Beyer, 2000)
showing the equivalence of a real-parameter GA (with the SBX operator) and the cor-
related ES. Based on the plethora of general-purpose optimization algorithms applied
to these problems over the years, we tend to conclude that the computational complex-
ities observed here (and that reported in the CMA-ES study as well) are probably the
best that can be achieved on these problems.
Based on this extensive study and computational advantage demonstrated over
other well-known optimization algorithms (evolutionary or classical), we recommend
the use of the proposed G3 model with the PCX operator on more complex problems
and on real-world optimization problems.
Acknowledgments
We thank Hajime Kita for letting us use his UNDX code. Simulation studies for the
differential evolution strategy and the CMA-ES are performed with the C code down-
References
Bäck, T. (1997). Self-adaptation. In Bäck, T., Fogel, D., and Michalewicz, Z., editors, Handbook of
Evolutionary Computation, pages C7.1:1–15, Institute of Physics Publishing, Bristol, UK, and
Oxford University Press, New York, New York.
Beyer, H.-G. (2001). The Theory of Evolution Strategies. Springer, Berlin, Germany.
Beyer, H.-G. and Deb, K. (2001). On self-adaptive features in real-parameter evolutionary algo-
rithms. IEEE Transactions on Evolutionary Computation, 5(3):250–270.
Branch, M. A. and Grace, A. (1996). Optimization toolbox for use with MATLAB, The MathWorks,
Inc., Natick, Massachusetts.
Deb, K. (1995). Optimization methods for engineering design, Prentice-Hall, New Delhi, India.
Deb, K. (2001). Multi-objective optimization using evolutionary algorithms. John Wiley and Sons,
Chichester, UK.
Deb, K. and Agrawal, R. B. (1995). Simulated binary crossover for continuous search space. Com-
plex Systems, 9(2):115–148.
Deb, K. and Beyer, H.-G. (2000). Self-adaptive genetic algorithms with simulated binary
crossover. Evolutionary Computation, 9(2):197–221.
Fogel, D. B. and Beyer, H.-G. (1995). A note on the empirical evaluation of intermediate recombi-
nation. Evolutionary Computation, 3(4):491–495.
Goldberg, D. E., Deb, K., and Clark, J. H. (1992). Genetic algorithms, noise, and the sizing of
populations. Complex Systems, 6(4):333–362.
Hansen, N. and Ostermeier, A. (1996). Adapting arbitrary normal mutation distributions in evo-
lution strategies: The covariance matrix adaptation. In Proceedings of the IEEE International
Conference on Evolutionary Computation, pages 312–317, IEEE Press, Piscataway, New Jersey.
Harik, G. et al. (1999). The gambler’s ruin problem, genetic algorithms, and the sizing of popula-
tions. Evolutionary Computation, 7(3):231–254.
Herrera, F., Lozano, M., and Verdegay, J. L. (1998). Tackling real-coded genetic algorithms: Oper-
ators and tools for behavioural analysis. Artificial Intelligence Review, 12(4):265–319.
Higuchi, T., Tsutsui, S., and Yamamura, M. (2000). Theoretical analysis of simplex crossover for
real-coded genetic algorithms. In Schoenauer, M. et al., editors, Parallel Problem Solving from
Nature (PPSN-VI), pages 365–374, Springer, Berlin, Germany.
Kita, H. and Yamamura, M. (1999). A functional specialization hypothesis for designing genetic
algorithms. Proceedings of the 1999 IEEE International Conference on Systems, Man, and Cyber-
netics, pages 579–584, IEEE Press, Piscataway, New Jersey.
Kita, H., Ono, I., and Kobayashi, S. (1999). Multi-parental extension of the unimodal normal
distribution crossover for real-coded genetic algorithms. In Porto, V. W., editor, Proceedings
of the 1999 Congress on Evolutionary Computation, pages 1581–1587, IEEE Press, Piscataway,
New Jersey.
Ono, I. and Kobayashi, S. (1997). A real-coded genetic algorithm for function optimization using
unimodal normal distribution crossover. In Bäck, T., editor, Proceedings of the Seventh Inter-
national Conference on Genetic Algorithms (ICGA-7), pages 246–253, Morgan Kaufmann, San
Francisco, California.
Patton, A. L., Goodman, E. D., and Punch, W. F. (1999). Scheduling variance loss using pop-
ulation level annealing for evolutionary computation. Proceedings of the IEEE Congress on
Evolutionary Computation (CEC-1999), pages 760–767, IEEE Press, Piscataway, New Jersey.
Rechenberg, I. (1973). Evolutionsstrategie: Optimierung Technischer Systeme nach Prinzipien der Biol-
ogischen Evolution. Frommann-Holzboog Verlag, Stuttgart, Germany.
Reklaitis, G. V., Ravindran, A., and Ragsdell, K. M. (1983). Engineering Optimization Methods and
Applications. John Wiley and Sons, New York, New York.
Salomon, R. (1999). The deterministic genetic algorithm: Implementation details and some re-
sults. Proceedings of the IEEE Congress on Evolutionary Computation (CEC-1999), pages 695–
702, IEEE Press, Piscataway, New Jersey.
Satoh, H., Yamamura, M., and Kobayashi, S. (1996). Minimal generation gap model for GAs con-
sidering both exploration and exploitation. In Yamakawa, T. and Matsumoto, G., editors,
Proceedings of the IIZUKA: Methodologies for the Conception, Design, and Application of Intelli-
gent Systems, pages 494–497, World Scientific, Singapore.
Schwefel, H.-P. (1987). Collective intelligence in evolving systems. In Wolff, W., Soeder, C. J.,
and Drepper, F., editors, Ecodynamics – Contributions to Theoretical Ecology, pages 95–100,
Springer, Berlin, Germany.
Storn, R. and Price, K. (1997). Differential evolution – a fast and efficient heuristic for global
optimization over continuous spaces. Journal of Global Optimization, 11:341–359.
Tsutsui, S., Yamamura, M., and Higuchi, T. (1999). Multi-parent recombination with simplex
crossover in real-coded genetic algorithms. In Banzhaf, W. et al., editors, Proceedings of the
Genetic and Evolutionary Computation Conference (GECCO-99), pages 657–664, Morgan Kauf-
mann, San Francisco, California.
Voigt, H.-M., Mühlenbein, H., and Cvetković, D. (1995). Fuzzy recombination for the Breeder
Genetic Algorithm. In Eshelman, L., editor, Proceedings of the Sixth International Conference on
Genetic Algorithms, pages 104–111, Morgan Kaufmann, San Francisco, California.
Wakunda, J. and Zell, A. (2000). Median-selection for parallel steady-state evolution strategies. In
Schoenauer, M. et al., editors, Proceedings of the Parallel Problem Solving from Nature, (PPSN-
VI), pages 405–414, Springer, Berlin, Germany.