These two factors are not independent, so we will perform an analysis using all the possible pairs (, ) of the Cartesian product of the two sets. For each pair we will perform 30 runs of the genetic algorithm with different random seeds. Table 2 shows the average value and standard deviation of the 30 runs for each experiment.
The study of the results has been made by means of an analysis of variance ANOVA II [DC74,Mil81,SC80], with the fitness of the best individuals, , as test variable. This fitness is obtained independently in 30 runs and depending on two fixed factors and their interaction. The fixed factors are: the confidence coefficient with four levels and the number of best individuals with five levels. The linear model has the form:
(14) | |||
and |
where:
The hypothesis tests try to determine the effect of each term over the fitness of the best individuals, . We have carried out tests for every factor and for the interaction among the factors. This and subsequent tests are performed with a confidence level of . The coefficient of the linear model tells us the percentage of variance of that is explained by the model.
For determining whether there are significant differences among the various levels of a factor, we perform a multiple comparison test of the average fitness obtained with the different levels of each factor. First, we carry out a Levene test [Mil96,Lev60] for evaluating the equality of variances. If the hypothesis that the variances are equal is accepted, we perform a Bonferroni test [Mil96] for ranking the means of each level of the factor. Our aim is to find the level of each factor whose average fitness is significantly better than the average fitness of the rest of the levels of the factor. If the test of Levene results in rejecting the equality of covariance matrixes, we perform a Tamhane test [TD00] instead of a Bonferroni test. Tables 9, 12, and 13 in Appendix A show the results obtained following the above methodology.
For Sphere function, the significant levels of each term of the linear model on Table 9 show that none of the factors of the linear model has a significant effect on the model built to explain the variance of the fitness . This effect is due to the fact that is easy to optimize and the fitness behaves as a singular random variable with sample variance near 0. We can see in Table 2 that the best results are obtained with the pair . The multiple comparison test of Table 12 confirms that the means obtained with the value are significatively better than the means obtained with other values. In the same way, the average fitness for is significantly the best one. The results show that, for any value of , the best value of , in general, is . Due to the simple form of , the best parameters of the crossover show a high exploitative component with a fast shifting towards the region of the best individuals.
For the unimodal and non-separable functions and , both factors and their interaction are significant in the linear model that explains the sample variance of with a determination coefficient around . Table 2 shows that the best results are obtained with ; the Tamhane test shows that the means obtained with this value of are significatively better than the means obtained with other values. The results for the value of the confidence coefficient are less conclusive. In fact, for there are no significant differences among the different values of , although the best results are obtained with . For the average fitness for is the best one, but without significant differences with . together with is the one that shows the best results. We can conclude that the feature of non-separability of the functions does not imply a notable change in the parameters of the crossover with respect to the parameters used for .
For and , which are separable and multimodal, the most adequate pair of parameters is . For , the test shows that the performance of this pair is significantly better. However, for , the best mean is obtained with with results that are significantly better than these obtained with other values, with the exception of . There are no significant differences among , and . The three factors of the linear model are significant with quite large determination coefficients of for and for. This means that the factors and their interaction explain a high percentage of the variance of the fitness .
For , the best results are obtained with the pair . The Tamhane test confirms that is the most suitable value, while there are no significant differences among , and . For the best results are obtained with the pair . The test shows that large values of are the most suitable for the optimization of this function. There are no significant differences among the performance of the different values of . For both functions the determination coefficient of the linear model is low, showing that the linear model does not explain the variance of the fitness. The lack of a linear relation among , and the fitness makes it more difficult to determine the best value of the parameters of the crossover.
The case of and is similar, as the linear model hardly gives any information about the effect of the parameters on the fitness. The most adequate pair for the optimization of these two functions is . The test shows that the best values of are and . On the other hand, there are no significant differences among the performance of the crossover for the different values of .
The overall results show that the selection of the best individuals of the population would suffice for obtaining a localization estimator good enough to guide the search process even for multimodal functions where a small value of could favor the convergence to local optima. However, if the virtual parents have a worse fitness than the parent from the population, the offspring is generated near the latter, and the domain can be explored in multiple directions. In this way, the premature convergence to suboptimal virtual parents is avoided.
However, if the best individuals are concentrated in a local optimum the algorithm will very likely converge to such optimum. That is the reason why in complex functions a larger value of may be reasonable, adding to the confidence interval individuals located in or near different optima. As an example of this, the case of for which the best results are achieved with and is noteworthy.
The confidence coefficient bounds the error in the determination of the localization parameter and is responsible for focussing the search. The multiple comparison tests show that the value is the best for 6 problems, and is, as least, no worse than the best one in the other problems. So it can be chosen as the most adequate value of the parameter.
Domingo 2005-07-11