Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
I. I NTRODUCTION
In recent years, Evolutionary Algorithms, in particular
Genetic Algorithms (GAs), have been applied to many different optimization problems, generally with great success
([1], [2]). The fact that these are guided search methods that
carry out optimization in conjunction with any simulation
model has allowed them to be adopted in a wide range
of disciplines. However, the successful application of a GA
requires expert knowledge of the algorithm, and how it solves
a problem.
One potential cause of poor performance of a GA is the
incorrect calibration of the GA parameters. Common GA
parameters include population size, probability of crossover,
and probability of mutation. The values used for these
parameters must be calibrated to each specific problem that is
tackled. Both the absolute value used for these parameters, as
well as their relative values, will determine how a GA finds
new solutions and, ultimately, the quality of the final solution
found. The values adopted for the GA parameters will
produce a search behavior between two possible extremes:
exploitation, where the current best solution is used as a basis
M.S. Gibbs, H.R. Maier and G.C. Dandy are at the Centre for Applied
Modelling in Water Engineering, School of Civil and Environmental Engineering, University of Adelaide, Adelaide SA 5005, Australia (e-mail:
mgibbs;hmaier;gdandy@civeng.adelaide.edu.au).
J.B. Nixon is with United Water International Pty Ltd, GPO Box 1875,
Adelaide SA 5001, Australia (e-mail: john.nixon@uwi.com.au).
0-7803-9487-9/06/$20.00/2006 IEEE
A. Empirical Studies
The simplest method for determining useful relationships
considering all aspects of a GA is to perform empirical
analyses on a range of test functions. However, there is
the potential that the results will be specific to the cases
that have been considered in the analyses, and will not hold
for the general case. [5] provide an example of this, where
test functions that have previously been used in empirical
experiments are separable functions, where there are no nonlinear interactions between decision variables. The functions
may be nonlinear in the contribution from each decision
variable to the fitness function, but without the nonlinear
interactions between the decision variables, each decision
variable can be optimized independently of the others. These
separable functions are often readily solved by local search
methods [5], and are most likely unsuitable representations
of more realistic problems that would be tackled in GA
applications. At best, empirical analyses can provide some
useful results about the classes of functions that have been
considered in the analyses, but it is unlikely that the results
will extend to more general applications of GAs.
565
B. Parameter Adaptation
A review of parameter adaptation methods is provided
in [6]. Three classes of parameter adaptation are defined:
deterministic parameter control, where the values change
in accordance with a predetermined rule; adaptive parameter control, where the values are changed based on the
performance of the algorithm; and self-adaptive parameter
control, where the parameters are built into the optimization
problem for the GA to solve along with the optimization
problem itself. While a number of these methods provide
some guidance on applying a GA to an optimization problem,
a complete calibration methodology based on parameter
adaptation is yet to be proposed. Also, only the adaptive
methods (and possibly, implicitly, the self-adaptive methods)
adjust the parameters based on the performance of the GA
on a given problem, and none of the methods analyze the
problem to determine the most appropriate GA parameters
for each case.
C. Theoretical Modeling
Numerous theoretical modeling studies have been performed to provide an insight into the inner workings of GAs.
As the interaction between GA operators is very complex,
generally the models consider the effect of only one or two
aspects of a GA, enabling meaningful results to be drawn.
Some of these studies include considering mutation alone
[7], selection and crossover [8], convergence due to selection
[9], convergence due to genetic drift [10], and using the
initial supply of building blocks and their correct selection
to determine the population size [11].
[9] and [12] have used simple models of GA convergence
to relate the number of generations before convergence
occurs to the length of the solution string. The models are
based on the fact that for selection schemes with a constant
selection intensity (e.g., tournament selection), the increase
in population mean fitness from one generation to the next
(f (g + 1) f (g)) is equal to the product of the selection
intensity (I) and the standard deviation of the population
fitness (t ) [12]:
f (g + 1) f (g) = t I.
This relationship is independent of the algorithm coding or
other operators (such as crossover or mutation), as it is only
dependent on the distribution of the fitness function under
a selection operator with constant selection intensity. [9]
considers normally distributed fitness functions to derive that
the number of
generations before convergence, gconv , is of
the order O( l), where l is the string length. As a specific
case, the Onemax problem is considered, where expressions
can be derived for f (t) and t as a proportion of the genes
that have
converged to the optimal value, 1. The relationship
gconv = 2 l is derived for this problem with a tournament
size of two, and validated experimentally with tournament
selection and uniform crossover.
[12] provides a similar analysis for arbitrarily distributed
fitness functions, where some building blocks have a higher
566
TABLE I
T HE TEST FUNCTIONS USED FOR THE PARAMETRIC STUDY.
F1
f1 (x) =
l
X
` 2
F2
f2 (x) =
l1
X
i=1
xi + xi+1
2
xi + xi+1
+ 10 cos 2
2
567
TABLE II
T HE GA PARAMETER VALUES USED
C. Parametric Study
To determine the best parameters for the GA, a large-scale
parametric study has been undertaken for each test function.
While the results for the GA parameters other than the
population size have not been included in the analyses, they
must be included in the calibration of the GA to ensure that
they are not biasing the results toward a certain population
size. The parameter values tested are given in Table II. Each
combination of parameter values was tested 30 times with
different random number sequences populations to evaluate
Parameter
Values
Population Size, n
Probability of Mutation, pm
Probability of Crossover, pc
Standard Deviation of Crossover,
Elite Solutions per Generation, e
Fig. 1. The crossover distribution used for the GA for one decision variable.
A normal distribution with standard deviation is used around p1 before the
crossover point, then around p2 after the crossover point. The distribution
shown uses = s/6.
568
(a) l = 5 Dimensions
(b) l = 10 Dimensions
(c) l = 20 Dimensions
(d) l = 30 Dimensions
Fig. 3. Optimal Population Sizes (n) for F2, the Rastrigin function with interactions between decision variables introduced, in l = 5, 10, 20, and 30
dimensions for increasing number of function evaluations, (FE). The slope of the straight line fitted between the optimal population sizes indicates the
number of generations before convergence for each problem. The box plots for the given number of function evaluations represent different combinations
of GA parameters that produce statistically similar results.
TABLE III
O PTIMAL GENERATION RESULTS FOR F2.
Fig. 2. Optimal Population Sizes (n) for F1, the original Rastrigin function,
in l = 5 dimensions for increasing number of function evaluations (FE).
g/l
g/ l
5
10
20
30
70
100
202
302
14.00
10.00
10.10
10.07
31.30
31.62
45.17
55.14
569
A. Epistasis Effects
The effect of the epistatic interactions introduced into F2
observed above can be explained by the efficient use of function evaluations. For F2, where there are interactions between
decision variables, combinations of decision variable values
must be stored and evaluated to solve the problem. For a GA,
this is done using a large population size, and better solutions
are found using a larger population size, provided there is
time for the GA to process the information in the population
and converge to a solution. Hence, as the number of function
evaluations increases, so does the optimal population size.
However, these larger population sizes do not find the best
results for fewer function evaluations, as there is insufficient
time to process all the combinations of decision variable
values in the population and thus to converge on a solution.
For F1, which does not have any interactions between
decision variables, each decision variable can be optimized
independently of the other decision variable values. Therefore, it is not necessary to store combinations of the decision
variable values, and function evaluations are wasted on poor
quality solutions stored in a large population. As can be seen
from the results in Figure 2, it is more efficient to have
a smaller population size updated more frequently through
more generations for the completely separable function.
B. Problem Size Effects
The results indicate that for F1, the size of the problem
does not affect the optimal population size, which is always
the smallest population size that has been considered. Similar
to the case described above for the epistasis effects, as F1
is a separable function, each decision variable can be optimized separately and therefore the smallest population size is
always the most efficient, and the only way to locate better
solutions is through more function evaluations. Therefore,
in this case, and possibly for all separable functions, better
results are obtained by increasing the number of generations,
as opposed to increasing the population size.
Figure 3 indicates that for F2 the problem size does have
an effect on the optimal population size, where for the same
number of function evaluations a smaller population size
performs better for the larger problem sizes compared to that
for the smaller problem sizes. The straight line fitted between
the optimal population sizes and the number of function
evaluations, seen in Figure 3, suggests that for F2 there is
an optimal number of generations to solve the problem, and
the values of the optimal number of generations for each
problem size are given in Table III. Table III also shows
570
571
[13] B. Minsker, Genetic algorithms, in Hydroinformatics: Data Integrative Approaches in Computation, Analysis and Modeling, P. Kumar,
J. Alameda, P. Bajcsy, M. Folk, and M. Markus, Eds. Florida, USA:
CRC Press, 2005, pp. 439456.
[14] D. Thierens, Analysis and design of genetic algorithms, PhD,
Katholieke Universiteit, 1995.
[15] G. R. Harik and F. G. Lobo, A parameter-less genetic algorithm, in
GECCO, Proc., W. Banzhaf, J. Daida, A. E. Eiben, M. H. Garzon,
V. Honavar, M. Jakiela, and R. E. Smith, Eds., vol. 1. Orlando,
Florida, USA: Morgan Kaufmann, 1999, pp. 258265.
[16] H. Muhlenbein and D. Schlierkamp-Voosen, Predictive models for
the breeder genetic algorithm: I. Continuous parameter optimization,
Evol. Comput., vol. 1, no. 1, pp. 2549, 1993.
[17] F. Herrera, M. Lozano, and A. M. Sanchez, A taxonomy for the
crossover operator for real-coded genetic algorithms: An experimental
study, Int. J. Intell. Syst., vol. 18, no. 3, pp. 309338, 2003.
[18] K. Deb and R. Agrawal, Simulated binary crossover for continuous
search space, Complex Syst., vol. 9, pp. 115148, 1995.
[19] H.-M. Voigt, H. Mhlenbein, and D. Cvetkovic, Fuzzy recombination
for the breeder genetic algorithm, in Proc. of the 6th Intl Conf. on
Genetic Algorithms, L. Eshelman, Ed. San Francisco, CA: Morgan
Kaufmann, 1995, pp. 104113.
572