WCCI 2012 IEEE World Congress on Computational Intelligence June, 10-15, 2012 - Brisbane, Australia
IEEE CEC
Particle Swarm Optimization with pbest Crossover Stephen Chen School of Information Technology York University Toronto, Canada
[email protected] design decisions (e.g. crossover rate, mutation rate, crossover operator, selection mechanism, population size, etc) than particle swarm optimization which has a recommended standard configuration that has been shown to be robust across a broad range of fitness landscapes [7].
Abstract—Particle swarm optimization can be viewed as a system with two populations: a population of current positions and a population of personal best attractors. In genetic algorithms, crossover is applied after selection – the goal is to create a new offspring solution using components from the best available solutions. In a particle swarm, the best available solutions are in the population of personal best attractors. Compared to standard particle swarm optimization, a modified version which periodically creates particle positions by crossing the personal best positions can achieve large improvements. These improvements are most consistent on multi-modal search spaces where the new crossover solutions may help the search process escape from local optima.
The following proposal to leverage the benefits of genetic algorithms in particle swarm optimization is relatively simple – it adds one new parameter and recommends small modifications to two existing PSO parameters. The new proposal also differs significantly from previous attempts (e.g. [5][6]) in that it crosses the personal best positions (pbest) instead of the current positions (x) of each particle. This change eliminates the need for additional selection mechanisms as the pbest positions already represent the best solutions available for mating.
Keywords-particle swarm optimization; crossover; multi-modal search spaces; exploration;exploitation
I.
INTRODUCTION
The development of particle swarm optimization (PSO) includes inspirations from “bird flocking, fish schooling, and swarming theory in particular” [1]. Each particle (e.g. a simulated bird) is attracted to its personal best position and the best position of a neighbouring member of the swarm. Rather than a simple line search between a current position and a best position, the velocity and momentum of each particle encourage a more explorative search path. Overall, particle swarm optimization is a simple and elegant nature-inspired metaphor for heuristic search that has achieved good performance across a broad range of optimization problems [2]. The development of genetic algorithms is inspired by Darwinian evolution and genetic crossover [3]. New solutions are created from components taken from the best solutions in the population. Since only the best solutions are allowed to be part of this “mating pool”, this selection process mimics survival of the fittest, and the resulting process allows better solutions to “evolve” over time. Overall, the genetic algorithm is also a simple and elegant nature-inspired metaphor for heuristic search that has achieved good performance across a broad range of optimization problems [4]. Several attempts have been made to add the advantages of genetic algorithms to the process of particle swarm optimization (e.g. [5][6]). Although these attempts have been mostly successful in terms of improving the performance of PSO, the increased complexity of these systems has limited their popularity. In general, the performance of genetic algorithms is greatly affected by many more parameters and
U.S. Government work not protected by U.S. copyright
1234
The addition of pbest crossover is tested across a broad range of benchmark functions. In general, heuristic search techniques are only needed for multi-modal search spaces since gradient descent tends to be sufficient for unimodal search spaces. Therefore, the primary functions used are set 4 (multimodal functions with adequate global structure) and set 5 (multi-modal functions with weak global structure) from the Black-Box Optimization Benchmarking (BBOB) functions [8]. The goals of these experiments are to show that pbest crossover can be more effective than crossing the current position x, and to show how this simple addition can lead to large improvements over standard PSO [7]. This presentation of the potential benefits of pbest crossover starts in Section II with a review of key ideas from genetic algorithms and particle swarm optimization. The addition of pbest crossover requires a new parameter for frequency (Section III), and it benefits from adjustments to the constriction factor (Section IV) and population size (Section V). Several features of pbest crossover help scale the PSO search process to higher dimension problems, and this is shown in Section VI. Despite the relative simplicity of the new modifications, they are shown in Section VII to be highly competitive with other more complex PSO-based systems. The paper then ends with a discussion of the key results in Section VIII and their summary in Section IX. II. BACKGROUND Several comparisons between particle swarm optimization and evolutionary computation (EC) approaches have been previously conducted (e.g. [9]). In EC, new solutions are
formed using crossover and mutation. In crossover, information from existing solutions is used to guide the search process, and this is similar to the attraction vectors in PSO which can encourage particles to explore “a region that represents the geometric mean between two promising regions” [9]. The exact points explored in a region can be influenced by mutation in EC which is generally omnidirectional, while the generation of new candidate solutions in PSO is affected by a particle’s velocity which is directional. A key feature in the development of pbest crossover is the concept of elitism. Although it is claimed that “PSO is the only evolutionary algorithm that does not incorporate survival of the fittest” [9], personal best information serves a highly analogous role. In particular, early GAs had generational replacement schemes [3][10]. Similar to natural systems, all members of a population are eventually replaced – even if it is the best solution. Most EC techniques (e.g. differential evolution [11]) are now implemented with elitism – the best solution always survives, and it is always active in guiding the search process. In a natural swarm, it is plausible that a bird might know its personal best position. However, it is more likely to be affected by a neighbour’s current position and velocity than its neighbour’s personal best position. This attraction to a neighbour’s personal best position has similar effects to elitism in EC – information about personal best positions is always maintained in the swarm memory, and this information is always active in guiding the search process. Attraction to personal best positions encourages exploitation and the particle’s momentum encourages exploration. As a particle explores away from its personal best attractor, this attraction “force” eventually draws the particle back to the area around the personal best. However, similar to a comet which spends a large amount of time very far away from its attracting star, a particle’s search trajectory can spend large amounts of time very far from its personal best position. This search process can be made more exploitative by periodically resetting the particle back to its personal best position [12]. Starting from “particle swarm optimization with resets”, it becomes very easy to add pbest crossover. III. PERIODIC CROSSOVER Crossover is a widely used mechanism in heuristic search (e.g. [3][11][13]). However, before the application of crossover, selection is used – crossover creates a new solution in a region of the search space defined by two promising solutions. The storage of pbest positions is effectively selection, so crossover can be applied to these pbest positions at the same time that a PSO with resets [12] returns each particle back to its personal best position. Compared to standard PSO, the resulting system is both more exploitative (because of the resets) and more explorative (because of the crossover). The following experiments build from the published source code of a PSO benchmark implemented by El-Abd and Kamel [14]. This benchmark implementation is for a gbest swarm using a star topology, so it requires a slight modification to become an lbest swarm with a ring topology. After this modification, the benchmark becomes an implementation of
1235
standard PSO [7] in which each dimension d of a particle’s velocity v is updated for the next iteration i+1 by vi 1,d vi ,d c1 1 pbesti ,d xi ,d c2 2 lbesti ,d xi ,d
where χ is the constriction factor, c1 and c2 are weights which vary the contributions of personal best and local best attractors, ε1 and ε2 are independent uniform random numbers in the range [0,1], x is the position of the particle, pbest is the best position found by the current particle, and lbest is the best position found by any particle communicating with the current particle (i.e. two neighbours in an LBest ring topology). Key parameters in [14] include 0.792 , * c1 * c2 1.4944 , i.e. c1 c2 1.887 , and p = 40 particles. To add periodic crossover, the standard position and velocity updates (1) are replaced instead by pbest crossover every crossiter iterations. (See Algorithm 1.) This crossover step sets the position of every particle to the midpoint of its pbest and a random pbest drawn without replacement from the population of pbest attractors. In this crossover scheme, it is possible for a solution to be mated with itself which causes the particle to be reset to its pbest position. Regardless, only the current position x of each particle is affected – pbest and v are not changed during this time step (unless the new position x is better than its pbest position). The frequency of crossover as represented by crossiter is the only new parameter introduced in the proposed implementation of pbest crossover. A large value for this parameter will naturally reduce the effects of the new modification. Small values for this parameter will increase both the use of crossover and the effects of resets. In the limit of crossiter = 1, the system will behave like a simple and extremely greedy genetic algorithm. In between the two limits of a simple GA and standard PSO, the crossiter parameter helps to define the balance between swarm-based search trajectories and a more focused search among the pbest positions. The need for balance between exploration and exploitation is most evident in multi-modal search spaces. In unimodal search spaces, highly exploitative techniques like hill climbing can be very effective at finding the (single) global optimum. For multi-modal search spaces, some exploration is required. However, a purely explorative technique (e.g. random search in Algorithm 1 Adding pbest crossover to PSO for iter = 2 : maxiterations if mod(iter,crossiter) == 0 parents = randperm(popsize); for i = 1 : popsize x(i,:) = (pbest(i,:) + pbest(parents(i,:))/2; end else v = w*v + c1*rand(popsize,DIM).*(pbest-x) + c2*rand(popsize,DIM).*(lbest-x); % remainder of velocity update from [14]
and 18). The improvement achieved with periodic crossover of the current positions (replace pbest with x in Algorithm 1) is shown for reference.
TABLE I BBOB FUNCTIONS Set
fn
Function Name
s X X X X X
Attribute u gs X X X X X X X X X
The addition of pbest crossover leads to meaningful improvements compared to standard PSO on the tested multimodal functions, but they are not that much larger than the improvements that can be achieved by crossing the current position x. However, crossover of the current positions x seems to work best with a larger value of crossiter = 200 – a value which may allow the current positions to converge and become more similar to their pbest positions at the time of crossover. This perceived need for greater convergence in x-based crossover may limit the advantages of the following parameter adjustments – first, an increase to the rate of convergence (see Section IV) and then the leveraging of this increased convergence through the use of a larger swarm size (see Section V) – which are able to improve the performance of PSO with pbest crossover.
1 Sphere 2 Ellipsoidal, original 1 3 Rastrigin 4 Büche-Rastrigin 5 Linear Slope 6 Attractive Sector 7 Step Ellipsoidal 2 8 Rosenbrock, original 9 Rosenbrock, rotated 10 Ellipsoidal, rotated X X 11 Discus X X 12 Bent Cigar X 3 13 Sharp Ridge X 14 Different Powers X 15 Rastrigin, rotated X 16 Weierstrass X 4 17 Schaffers F7 X 18 Schaffers F7, moderately ill-conditioned X 19 Composite Griewank-Rosenbrock F8F2 X 20 Schwefel 21 Gallagher’s Gaussian 101-me Peaks 5 22 Gallagher’s Gaussian 21-hi Peaks 23 Katsuura 24 Lunacek bi-Rastrigin Names and selected attributes of the 24 functions in the BBOB problem set – separable (s), unimodal (u), global structure (gs).
the limit) will not be able to conduct a finer search around the most promising regions. An effective search technique for multi-modal search spaces therefore requires explorative capabilities to find the most promising regions of the search space and exploitative capabilities to find the best solutions (e.g. the local optima) in these regions. The following analysis of particle swarm optimization with pbest crossover focuses on two sets from the Black-Box Optimization Benchmarking (BBOB) functions [8]: set 4, multi-modal functions with adequate global structure, and set 5, multi-modal functions with weak global structure. See Table I for more information on the BBOB functions. To be consistent with previous results (e.g. [15][16][17]), the following experiments perform 25 independent trials on each function (5 trials on each of the first 5 instances) with a fixed limit of 5000*D function evaluations. Initial experiments start with D = 20 dimensions. The first set of experiments measures the effect of different values for the crossiter parameter. The results for the previously described benchmark implementation of standard PSO were first reported in [16]. Compared to these results, Table II reports the difference (%-diff = (a-b)/a) between the mean performance for standard PSO across 25 independent trials for each function (a) and the mean performance for PSO with pbest crossover (b). A positive %-diff indicates that the addition of crossover has led to an improvement. These improvements are then averaged over the 10 functions (mean) to pick a single value of crossiter = 100 which produces the best overall performance. Naturally, there are some functions which can benefit from a more explorative (higher) value for crossiter (e.g. BBOB fn 19 and 20) and a more exploitative (lower) value for crossiter (e.g. BBOB fn 17
1236
IV. ADJUSTING THE CONSTRICTION FACTOR There are several functions (e.g. BBOB fn 15) in which performance improves as the value of crossiter decreases, and then the performance degrades considerably with values of crossiter that become too small. One possible explanation of this effect is that the reduced number of iterations between resets (and crossover) interferes with the ability of the swarm to fully converge to a local optimum. An alternative possibility is that the dislocations to particle positions caused by crossover can lead to larger attraction vectors (and thus velocities) which will again reduce the swarm’s overall rate of convergence. For both reasons, the following set of experiments was conducted to determine if PSO with pbest crossover should use a smaller constriction factor χ. TABLE II EFFECTS OF FREQUENCY OF CROSSOVER crossiter BBOB fn 50 100 200 400 1000 15 24.3% 45.9% 26.8% 21.2% 14.2% 16 -10.7% 16.6% 15.1% 1.2% 5.4% 17 87.3% 75.2% 56.0% 35.7% 5.9% 18 69.7% 52.3% 40.3% 13.5% -1.4% 19 -10.8% 9.6% 28.8% 24.0% 9.1% 20 -27.7% -0.1% 6.6% 8.9% -0.5% 21 -19.7% 19.0% -6.5% 7.8% 16.9% 22 -0.1% -7.7% -13.9% 17.9% -45.4% 23 -36.0% 13.5% 43.8% 32.7% 20.4% 24 8.3% 31.6% 29.7% 15.9% 11.4% mean 8.4% 22.7% 17.9% 3.6% 25.6% 15 13.6% 46.9% 39.1% 17.2% 5.7% 16 -3.5% 24.2% 16.9% 12.7% -2.3% 17 81.6% 74.2% 51.1% 35.3% 11.6% 18 54.4% 45.0% 32.6% 21.5% -2.2% 19 -7.9% 16.1% 30.1% 19.0% 15.0% 20 -24.5% -2.7% 3.7% 9.8% -8.5% 21 8.8% -7.6% -1.3% -5.1% 34.9% 22 13.0% -31.4% -8.9% -27.4% -15.3% 23 -49.0% 7.9% 43.1% 32.3% 17.1% 24 5.9% 29.9% 30.3% 17.1% 12.1% x mean 9.2% 20.3% 13.2% 6.8% 23.7% pbest crossover (top) works best with crossiter = 100 and x-based crossover (bottom) works best with crossiter = 200, but the performance difference between these two options for crossover is quite small.
TABLE III EFFECTS OF REDUCED CONSTRICTION FACTORS constriction factor reduction BBOB fn 0.95 0.90 0.85 0.80 15 53.3% 41.9% 40.7% 37.0% 16 21.7% 18.8% -3.2% -14.1% 17 60.9% 48.7% 29.0% -3.6% 18 45.2% 45.8% 19.1% -0.6% 19 28.7% 43.9% 48.0% 44.6% 20 -5.2% 1.4% -13.9% -13.0% 21 16.5% -21.4% 20.9% 34.6% 22 -1.1% 7.2% -4.1% -12.2% 23 41.6% 49.8% 46.4% 43.4% 24 45.7% 52.6% 52.4% 58.5% mean 28.9% 23.5% 17.5% 30.7% 15 41.2% 36.4% 35.5% 25.6% 16 10.6% -5.7% -14.8% -26.1% 17 37.1% 9.8% -4.3% -29.4% 18 24.2% -2.3% -14.6% -18.9% 19 39.1% 34.9% 38.4% 41.1% 20 -0.8% -7.8% -17.4% -15.0% 21 -23.2% -11.9% 3.8% 31.9% 22 3.9% 4.8% -34.4% -31.0% 23 38.2% 41.0% 39.6% 36.1% 24 37.7% 45.3% 48.2% 49.6% x mean 20.8% 14.4% 8.0% 6.4% base -3.4% -9.6% -4.7% -10.8% Only pbest crossover (top) with crossiter = 100 receives from a constriction factor reduction (of 0.95).
TABLE IV EFFECTS OF POPULATION SIZE ON PBEST CROSSOVER population size p BBOB fn 30 50 60 80 100 15 54.5% 54.6% 50.0% 51.6% 44.3% 16 17.3% 24.5% 19.7% 23.4% 25.3% 17 44.2% 83.4% 69.2% 77.7% 74.2% 18 36.3% 51.2% 49.3% 49.9% 54.8% 19 31.9% 29.0% 26.7% 20.3% 18.2% 20 -1.5% 0.4% -5.3% 0.9% -3.1% 21 -5.4% 34.3% 48.9% 64.6% 75.9% 22 -2.6% -1.9% 25.7% 39.2% 32.8% 23 45.6% 37.7% 37.3% 37.9% 33.3% 24 48.0% 43.7% 40.1% 36.0% 30.8% mean 26.8% 35.7% 36.2% 38.7% 40.1% 15 37.2% 37.3% 34.3% 23.3% 18.4% 16 18.8% 16.3% 10.6% 13.2% 4.7% 17 45.3% 51.8% 50.7% 39.5% 45.7% 18 18.3% 41.1% 32.0% 22.3% 19.9% 19 35.3% 30.6% 22.0% 20.8% 23.5% 20 11.5% 7.1% 9.2% 3.3% 9.7% 21 -18.0% 6.1% 50.8% 65.8% 56.1% 22 -16.9% 22.9% 16.6% 21.2% 32.4% 23 42.8% 42.7% 41.5% 33.0% 33.5% 24 36.2% 27.1% 22.2% 21.0% 16.4% x mean 21.1% 28.3% 26.3% 26.0% 29.0% base -7.5% 6.8% 4.4% 4.4% 8.3% pbest crossover (top) benefits more from a larger population size than x-based crossover (bottom).
0.75 32.6% -22.3% -44.5% -14.6% 49.3% -18.8% 38.9% 2.7% 44.6% 51.8% 12.0% 16.7% -45.7% -48.9% -18.4% 41.6% -20.8% 13.5% -2.6% 35.8% 47.8% 1.9% -16.5% a benefit
The results in Table III again report the %-diff between the performance of PSO with pbest crossover and standard PSO. The total modification between these two variations is pbest crossover every crossiter = 100 iterations and a reduction to the constriction factor χ produced by multiplying it with the listed values (i.e. 0.95, 0.90, 0.85, 0.80, and 0.75). Several functions (e.g. BBOB fn 19, 23, and 24) have noticeable benefits from smaller constriction factors, and it is hypothesized that this effect is related to the shape (e.g. narrowness) of the local optima wells in the fitness landscape. However, across all 10 multi-modal functions, the best overall results occur when the constriction factor is reduced by five percent (i.e. χ = 0.792 * 0.95 = 0.752).
able to converge in the allotted number of function evaluations, and more particles should also promote a more diverse exploration of a multi-modal search space. The results in Table IV have the same format as Tables II and III except that the varying parameter is the population size p and the base parameters for pbest crossover are crossiter = 100 and χ = 0.752 (i.e. 0.95 times the benchmark value). Crossover of the current positions x uses crossiter = 200, and the constriction factor of χ = 0.792 is used for both x-based crossover (x mean) and the benchmark PSO (base). The tested population sizes are p = 30, 50, 60, 80, and 100.
Similar experiments on the effects of a reduced constriction factor were also performed for x-based crossover (x mean) and the benchmark PSO (base). Since more constriction (i.e. a smaller constriction factor) will decrease the explorative capacity of the swarm, the observed decrease in performance matches expectations for the base/benchmark PSO results. The x-based crossover results (with crossiter = 200) also follow this trend which is a minor divergence from the pbest crossover results (i.e. no benefit at any reduced levels). It was expected that the dislocations to the current positions would also cause this system to require more constriction to properly converge.
For all three systems, p = 30 leads to worse results and p = 50, 60, 80, and 100 lead to better results. As expected, it is a general benefit to have more particles when exploring a multimodal search space. However, the best population size is different for all three systems. Since more particles mean fewer iterations (i.e. less time to converge), a system with features that increase swarm convergence (e.g. more constriction, resets, and more communication among particles as facilitated by crossover) should be able to use more particles effectively. For its best set of parameters, pbest crossover is 40% better on average than the original configuration of standard PSO and 10% better on average than the best implementation of crossover applied to the current positions x.
V. FINDING THE BEST POPULATION SIZE The effect of a ring topology is that communication from one particle to another particle (e.g. on the other side of the ring) can be greatly delayed. This delayed communication slows the rate of convergence. The effect of crossover between random pairs of particles is to increase the rate of communication and thus the rate of convergence. With a higher rate of convergence, a swarm with more particles should be
VI. HIGH DIMENSION SEARCH SPACES The complexity of an optimization problem (e.g. the number of local optima) can increase exponentially with dimensionality, so search techniques that are very effective on low-dimension problems may not be efficient enough to handle problems in higher dimensions. One reason why dimensionality may negatively affect the performance of
1237
TABLE V EFFECTS OF INCREASED DIMENSIONALITY D = 20 D = 200 BBOB fn x-based pbest x-based pbest 1 0.0% 0.0% 99.5% 100.0% 2 0.0% 0.0% 100.0% 100.0% 3 22.2% 25.8% 35.0% 27.4% 4 9.7% 12.6% 14.0% 7.9% 5 0.0% 0.0% 86.2% 98.2% 6 24.6% 92.9% 60.9% 92.3% 7 58.8% 67.9% 68.0% 73.5% 8 -7.0% -13.4% 48.5% 57.0% 9 -6.2% -2.8% 51.8% 64.4% 10 -84.6% -60.6% 47.7% 85.6% 11 25.7% 44.4% 22.5% 57.3% 12 -279.8% -254.1% 100.0% 100.0% 13 -68.5% -9.6% 93.8% 99.3% 14 -11.7% 8.4% 90.2% 97.0% 15 34.3% 51.6% 41.6% 51.0% 16 10.6% 23.4% 6.9% 12.6% 17 50.7% 77.7% 36.4% 44.8% 18 32.0% 49.9% 36.4% 46.7% 19 22.0% 20.3% 37.7% 40.8% 20 9.2% 0.9% 98.8% 98.8% 21 50.8% 64.6% 19.5% 32.8% 22 16.6% 39.2% 24.8% 31.5% 23 41.5% 37.9% 30.7% 18.0% 24 22.2% 36.0% 47.5% 50.3% mean -1.1% 54.1% 13.0% 62.0% The benefits of crossover tend to improve with increasing dimensionsality.
TABLE VI COMPARISON OF PBEST CROSSOVER WITH LOCUST SWARMS D = 20 D = 200 BBOB locust pbest locust pbest fn swarms crossover swarms crossover 1 -∞ 0.0% -3.2e6% 100.0% 2 -∞ 0.0% -3.0e3% 100.0% 3 62.5% 25.8% 77.7% 27.4% 4 55.6% 12.6% 80.5% 7.9% 5 0.0% 0.0% 100.0% 98.2% 6 51.3% 92.9% 98.3% 92.3% 7 61.4% 67.9% 89.3% 73.5% 8 -5.4% -13.4% 67.8% 57.0% 9 -1.8% -2.8% 68.2% 64.4% 10 51.8% -60.6% 93.0% 85.6% 11 56.6% 44.4% 72.3% 57.3% 12 -1.0e3% -254.1% -42.4% 100.0% 13 -132.9% -9.6% 80.1% 99.3% 14 -150.5% 8.4% 78.1% 97.0% 15 49.6% 51.6% 53.2% 51.0% 16 45.8% 23.4% 75.9% 12.6% 17 72.6% 77.7% 22.6% 44.8% 18 65.1% 49.9% 35.9% 46.7% 19 12.6% 20.3% 28.6% 40.8% 20 38.7% 0.9% 99.2% 98.8% 21 -101.7% 64.6% -7.4% 32.8% 22 -190.2% 39.2% -2.6% 31.5% 23 51.7% 37.9% 41.3% 18.0% 24 35.8% 36.0% 39.3% 50.3% PSO with pbest crossover has better overall performance than a more complex multi-swarm system.
standard PSO is that different dimensions can converge at different rates (e.g. [18]). A subtle advantage of crossover is that it can restimulate stalled search components (i.e. velocity components that have prematurely converged to zero).
using separate search/coordination techniques which focus more on exploration. These techniques introduce many extra parameters and design considerations (e.g. [16][22][23]), but it appears that the much simpler modifications involved with pbest crossover can be highly competitive with them.
The following experiments use the full BBOB problem set in D = 200 dimensions. The swarm parameters are taken from the best results for D = 20 dimensions (i.e. crossiter = 100, χ = 0.792*0.95 = 0.752, and p = 80 for pbest crossover, and crossiter = 200, χ = 0.792, and p = 60 for x-based crossover) except that the function evaluations grow with D to FE = 5000*D = 1*106. For reference, the results (i.e. %-diff relative to standard PSO) for D = 20 are also shown in Table V.
The performance of locust swarms is generally better than WoSP due to its specific selection of initial positions and initial velocities [16], and locust swarms appear to perform better than DMS-PSO before hybridizations and local search are added [22]. In Table VI, the results for PSO with pbest crossover and locust swarms (from [16]) are compared on the full BBOB problem set for D = 20 and D = 200 dimensions.
Comparing the results for D = 20 and D = 200 dimensions, the performance improvements with crossover are more or less consistent on the multi-modal functions (e.g. BBOB fn 15-24). However, there are exceptionally large changes in the performance as dimensionality increases on the unimodal functions (e.g. BBOB fn 10-14). Since the mid-point of two solutions on a unimodal function is unlikely to be better than both of the original solutions, and there are no local optima from which to escape, the ability to restimulate particle motions is the proposed explanation for the performance improvements of crossover on unimodal functions in higher dimensions.
Due to several exaggerated percentage changes (i.e. BBOB fn 1 and 2), the mean performance becomes meaningless, so it is not reported. Nonetheless, it can still be seen that the overall performance with pbest crossover is better than locust swarms. The primary advantages of multi-swarm systems are the ability to explore multiple optima in multi-modal search spaces and the ability to restimulate stalled search components in unimodal search spaces [15]. It appears that crossover can similarly provide both of these advantages in a much simpler overall system.
VII. COMPARISONS WITH MULTI-SWARM SYSTEMS
The recommended implementation of PSO with pbest crossover is also much simpler than previous methods of adding crossover to PSO (e.g. [5][6]). For example, mid-point crossover is simpler than linear-range crossover, and the current recommendation leaves the velocities unchanged (like a PSO with resets [12]) as opposed to also crossing them. Unreported experiments with these variations showed
VIII. DISCUSSION
Multi-swarm systems (e.g. Waves of Swarm Particles (WoSP) [19], locust swarms [20], and Dynamic Multi-Swarm Particle Swarm Optimizer (DMS-PSO) [21]) attempt to improve the balance between exploration and exploitation by having the (smaller) individual swarms focus more on exploitation (e.g. converging quickly to a local optimum) and
1238
negligible changes in performance, so these modifications are omitted from the final system for the sake of simplicity.
[3] [4]
One variation which did show an improvement was the use of directed initial velocities [16]. Since the pbest position can converge near a local optimum very quickly, the process of resets and pbest crossover is similar to starting a new subswarm in a multi-swarm system. It has been shown in multiswarm systems that launching the particle away from the previous position (on a line through the new initial position) can lead to improved performance [16][20], so this was tried with pbest crossover. However, since the resulting improvement was relatively small, this modification was also omitted from the final system for the sake of simplicity.
[5]
[6]
[7] [8]
Furthering the comparison with multi-swarm systems, PSO with pbest crossover and DMS-PSO [21] have some interesting similarities. In DMS-PSO, the swarm is divided into parallel sub-swarms which periodically regroup. This regrouping encourages exploration among the (local) optima discovered by the previous sub-swarms. The crossing of pbest positions also encourages exploration among the (local) optima discovered by each particle. The advantages of pbest crossover are that the periodic sampling can be faster (since pbest positions converge faster than the actual particles [24]) and the population of pbest positions can have more diversity than a population divided into n sub-populations that subsequently converge to n < p pbest positions. Convergence occurs in PSO when a particle with zero speed has the same position as all of its pbest attractors. The current use of pbest crossover reduces convergence by moving the position of a particle away from its pbest attractors. Other methods to reduce convergence include disallowing zero velocities [18], moving the pbest attractors [24][25], or preventing the co-location of the pbest attractors [17][26]. A full comparison of these alternate techniques is an area for further research. IX.
[9] [10] [11]
[12]
[13] [14]
[15]
[16]
[17]
[18]
SUMMARY
[19]
The addition of crossover to standard PSO can be simple and highly effective. The presented model randomly crosses all of the particles in the swarm every crossiter iterations. Performance improvements can be achieved by crossing either current positions x or the pbest positions. The best overall results occur with pbest crossover as this system benefits from being both more exploitative (e.g. more constriction and resets) and more explorative (e.g. more particles).
[20] [21] [22] [23] [24]
REFERENCES [1] [2]
J. Kennedy and R. C. Eberhart, “Particle swarm optimization,” IEEE ICNN, 1995, pp. 1942–1948. A. P. Engelbrecht, Fundamentals of Computational Swarm Intelligence. Hoboken, NJ: John Wiley & Sons, 2006.
[25] [26]
1239
J. Holland, Adaptation in Natural and Artificial Systems. Ann Arbor, MI: The University of Michigan Press, 1975. S. N. Sivanandam and S. N. Deepa, Introduction to Genetic Algorithms. Heidelberg: Springer, 2007. M. Løvbjerg, T. K. Rasmussen, and T. Krink, “Hybrid particle swarm optimiser with breeding and subpopulations,” GECCO, 2001, pp. 469– 476. M. Senthil Arumugam and M. V. C. Rao, “Particle swarm optimization algorithm with cross-over and RMS variants for computing optimal control of a class of hybrid systems,” Applied Soft Computing, vol. 8, pp. 324–336, 2008. D. Bratton and J. Kennedy, “Defining a standard for particle swarm optimization,” IEEE SIS, 2007, pp. 120–127. N. Hansen, S. Finck, R. Ros, and A. Auger, “Real-parameter black-box optimization benchmarking 2009: noiseless functions definitions,” INRIA Technical Report RR-6829, 2009. R. C. Eberhart and Y. Shi, “Comparison between genetic algorithms and particle swarm optimization,” EP VII, 1998, pp. 611–616. D. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning. Boston, MA: Addison-Wesley, 1989. R. Storn and K. Price, “Differential evolution – a simple and efficient heuristic for global optimization over continuous spaces,” J. Global Optimization, vol. 11, pp. 341–359, 1997. Y. Noa Vargas and S. Chen, “Particle swarm optimization with resets – improving the balance between exploration and exploitation,” MICAI, 2010, pp. 371–381. H.-G. Beyer and H.-P. Schwefel, “Evolution strategies: a comprehensive introduction,” Natural Computing, vol. 1 pp. 3–52, 2002. M. El-Abd and M. S. Kamel, “Black-box optimization benchmarking for noiseless function testbed using particle swarm optimization,” GECCO, 2009, pp. 2269–2273. S. Chen and Y. Noa Vargas, “Improving the performance of particle swarms through dimension reductions – a case study with locust swarms,” IEEE CEC, 2010, pp. 2950–2957. S. Chen and J. Montgomery, “Selection strategies for initial positions and initial velocities in multi-optima particle swarms,” GECCO, 2011, pp. 53–60. S. Chen and J. Montgomery, “A simple strategy to maintain diversity and reduce crowding in particle swarm optimization,” Austalasian AI, 2011, pp. 281–290. T. Hendtlass, “Particle swarm optimisation and high dimensional problem spaces,” IEEE CEC, 2009, pp. 1988–1994. T. Hendtlass, “WoSP: a multi-optima particle swarm algorithm,” IEEE CEC, 2005, pp. 727-734. S. Chen, “Locust swarms – a new multi-optima search technique,” IEEE CEC, 2009, pp. 1745–1752. J. J. Liang and P. N. Suganthan, “Dynamic multi-swarm particle swarm optimizer,” IEEE SIS, 2005, pp. 124–129. S. Chen, “An analysis of locust swarms on large scale global optimization problems,” ACAL, 2009, pp. 211–220. A. Bolufé Röhler and S. Chen, “An analysis of sub-swarms in multiswarm systems,” Austalasian AI, 2011, pp. 271–280. M. G. Epitropakis, V. P. Plagianakos, and M. N. Vrahatis, “Evolving cognitive and social experience in particle swarm optimization through differential evolution,” IEEE CEC, 2010, pp. 2400–2407. S. S. Pace and C. Woodward, “Diversity preservation using excited particle swarm optimisation,” GECCO, 2011, pp. 61–68. R. Brits, A. P. Engelbrecht, and F. Van den Bergh, “A niching particle swarm optimizer,” SEAL, 2002, pp. 692–696.