A Hierarchical Particle Swarm Optimizer and Its Adaptive Variant
A Hierarchical Particle Swarm Optimizer and Its Adaptive Variant
6, DECEMBER 2005
Abstract—A hierarchical version of the particle swarm optimiza- space according to its velocity vector. In every iteration, the ve-
tion (PSO) metaheuristic is introduced in this paper. In the new locity vector is adjusted so that prior personal successful po-
method called H-PSO, the particles are arranged in a dynamic hi- sitions (cognitive aspect) and the best position found by parti-
erarchy that is used to define a neighborhood structure. Depending
on the quality of their so-far best-found solution, the particles move cles within a specific neighborhood (social aspect) act as attrac-
up or down the hierarchy. This gives good particles that move up tors. In this paper we concentrate on PSO for continuous search
in the hierarchy a larger influence on the swarm. We introduce spaces but it should be mentioned that PSO has also been ap-
a variant of H-PSO, in which the shape of the hierarchy is dy- plied to discrete optimization problems (e.g., [13]).
namically adapted during the execution of the algorithm. Another In the original PSO algorithm, the neighborhood of a par-
variant is to assign different behavior to the individual particles
with respect to their level in the hierarchy. H-PSO and its variants ticle consists of all particles, so that the global best position,
are tested on a commonly used set of optimization functions and are i.e., the best solution found so far, directly influences its be-
compared to PSO using different standard neighborhood schemes. havior. Several authors have investigated the use of restricted
Index Terms—Author, please supply your own keywords or send neighborhoods. In [14], several fixed neighborhoods including
a blank e-mail to [email protected] to receive a list of suggested random neighborhoods have been studied. Dynamic neighbor-
keywords. hoods in which the neighborhood of a particle was defined as the
closest individuals in every iteration have been investigated in
[15]. The use of such a dynamic neighborhood is computation-
I. INTRODUCTION
ally intensive because the neighborhood has to be determined
II. PSO
Manuscript received September 7, 2004; revised December 21, 2004 and Jan-
uary 28, 2005. This work was supported by the German Research Foundation In this section, we describe the PSO method for function opti-
(DFG) through the project “Swarm Intelligence on Reconfigurable Architec-
tures”. This paper was recommended by Associate Editor M. Dorigo. mization (see also [1]). PSO is an iterative method that is based
The authors are with the Parallel Computing and Complex Systems Group, on the search behavior of a swarm of particles in a multi-
Department of Computer Science, University of Leipzig, D-04109 Leipzig, Ger- dimensional search space. In each iteration the velocities and
many (e-mail: [email protected]; [email protected]
leipzig.de). positions of all the particles are updated. For each particle ,
Digital Object Identifier 10.1109/TSMCB.2005.850530 its velocity vector is updated according to (1). The inertia
1083-4419/$20.00 © 2005 IEEE
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
JANSON AND MIDDENDORF: A HIERARCHICAL PARTICLE SWARM OPTIMIZER AND ITS ADAPTIVE VARIANT 1273
weight controls the influence of the previous velocity the diversity of the algorithm to not being trapped in a local
vector. The current position of the particle is denoted by . Pa- minimum.
rameter controls the impact of the personal best position In [15], a neighborhood scheme has been explored that is de-
, i.e., the position where the particle found the smallest func- fined by a particle’s actual position in the search space. A certain
tion value so far—assuming that the objective function has to be number of close particles are considered to be neighbors. This
minimized. Parameter determines the impact of the best method is computationally intensive, since in every iteration the
position that has been found so far by any of the particles in distances between all pairs of particles have to be calculated.
the neighborhood of particle . Usually and are set to the A fitness-distance-ratio PSO (FDR-PSO) has been proposed
same value. Random values and are drawn with uniform in [23], in which each particle is not only influenced by the
probability from [0,1] for each particle at every iteration. personal and global best position but also by a close and good
After all the particles’ velocities have been updated, the par- neighbor. This neighbor is selected as the particle that maxi-
ticles move with their new velocity to their new positions (2). mizes the quotient of fitness improvement over the respective
Then, for each particle the objective function is evaluated distance. Thus, any nearby improving neighbor of a particle
at its new position. If the personal best can be preferred to the global best particle, provided it is close
position is updated accordingly, i.e., is set to enough.
In [14] and [24], several neighborhood topologies or “so-
ciometries” have been examined for the PSO algorithm. In [24]
these topologies have also been applied to the fully informed
(1) PSO, in which each particle is influenced by all of its neigh-
(2) bors and not only its best neighbor. The information flow within
the swarm is controlled by the two parameters and , where
Several variations of this basic PSO scheme have been pro- gives the number of neighbors of a specific node in the neighbor-
posed in the literature. Commonly used variations are to restrict hood graph and is used to measure the clustering among the
the velocity of a particle by a maximal value or to linearly nodes in the neighborhood graph, i.e., to what extent the respec-
decrease over time [17]. This is done to adjust the swarm’s tive neighborhoods differ. The previously introduced neighbor-
behavior from exploration of the entire search space to exploita- hoods gbest and lbest, the star—one central node is connected
tion of promising regions. to all the other nodes—and other regular neighborhood graphs
Various mechanisms have been designed to increase the di- have been compared to randomly created neighborhood graphs
versity among the particles of a swarm. In [18] a spatial exten- with different values for and . The von Neumann neighbor-
sion is assigned to the particles and different collision strate- hood on a two-dimensional (2-D) lattice performed very good
gies are used to avoid crowding of the swarm. A charged swarm and also a three-dimensional (3-D) pyramid did perform reason-
(CPSO) is proposed in [19], where some or all the particles ably well. The common gbest, lbest and star topologies all have
hold a certain charge and a repulsion force is applied if two been rated worse.
particles get too close to each other. The ARPSO algorithm
[20] switches between phases of attraction and repulsion. If the
swarm becomes too tight, i.e., the diversity diminishes, the re- III. HIERARCHICAL PSO
pulsion phase is initiated and the swarm is scattered. In [21], a The hierarchical version of PSO (H-PSO) is introduced in this
predator particle is introduced that pursues the global best par- section. In H-PSO, all particles are arranged in a hierarchy that
ticle and thereby chases other particles away.
defines the neighborhood structure. Each particle is neighbored
to itself and its parent in the hierarchy. In this paper, we study
A. Neighborhood Topologies
regular tree like hierarchies, i.e., the underlying topology is a
Different neighborhood topologies have been investigated (nearly) regular tree. The hierarchy is defined by the height ,
for PSO. In the original PSO algorithm—here called the gbest the branching degree , i.e., the maximum number of children
model—the swarm is guided by the current global best particle, of the inner nodes, and the total number of nodes of the corre-
i.e., in (1) is the best solution found so far by the swarm. The sponding tree (for an example, see Fig. 1). In this paper we use
gbest model corresponds to a fully connected neighborhood. only hierarchies in which all inner nodes have the same number
In [22], other neighborhood topologies, varying the degree of of children, only the inner nodes on the deepest level might have
interconnections between the particles, have been introduced. a smaller number of children so that the maximum difference
In this paper, we also consider the lbest model, that uses the between the number of children of inner nodes on the deepest
local neighborhood best position to update a particle’s velocity. level is at most one.
The local neighborhood is defined by a ring topology through In order to give the best individuals in the swarm a high influ-
the particle’s index, so that particle is neighbored to particles ence, particles move up and down the hierarchy. In every iter-
and , where is the total number ation, after the evaluations of the objective function at the par-
of particles. It has been shown [22] that the relative performance ticles actual positions, but before the update of the velocities
of gbest and lbest depends on the type of the optimization func- and the determination of the new positions in the search space,
tion. In general, gbest performs better on unimodal functions, as the new positions of the particles within the hierarchy are de-
Sphere and Rosenbrock, whereas lbest is better suited for mul- termined as follows. For every particle in a node of the tree,
timodal functions, in which the optimization success relies on its own best solution is compared to the best solution found by
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
1274 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 35, NO. 6, DECEMBER 2005
the particles in the child nodes. If the best of these particles, say
particle , is better (i.e., ) then particles and
swap their places within the hierarchy. These comparisons are
performed starting from the top of the hierarchy and then pro-
ceed in a breadth-first manner down the tree. Observe that the
top-down approach implies that in one iteration an individual
can move down several levels in the hierarchy but it can move
up at most one level. The current global best particle will move
up one level of the hierarchy at every iteration. Hence, it will be
on top of the hierarchy after at most iterations—unless, a
better solution was found meanwhile. Fig. 2. Example for adapting the hierarchy of AH-PSO (m = 20) from d = 4
to d = 3.
For the update of the velocities in H-PSO, a particle is influ-
enced by its own so far best position and by the best position of
the individual that is directly above in the hierarchy. This means diameter similar to the gbest neighborhood, that, in general, op-
that for particle the value of in (1) equals where is the timizes faster than the PSO algorithm using the lbest neighbor-
particle in the parent node of particle . Only when particle is hood. On the other hand, the quality increase of the best solu-
tions found by H-PSO with a smaller might be slower in the
in the root of the tree, H-PSO uses .
beginning of the optimization process but it might improve the
Similar as in PSO, after the particles’ velocities are updated
objective function value further in the end of the optimization
and after the particles have moved in H-PSO, the objective func-
process (see Section V-A).
tion is evaluated at the new position. If the function value at this
position is better than the function value at the personal best po- These expectations have inspired the idea to dynamically
sition, the new position is stored in . change the branching degree of the hierarchy during a run. A
related idea has been used in [15] where the neighborhood of
A. Neighborhood the particles was gradually increased from lbest to gbest. Sur-
prisingly, the opposite direction, which might be more effective,
We propose a new neighborhood scheme for PSO that uses has not been tested. In the Adaptive H-PSO (AH-PSO) algo-
the particle’s so far best found function value to define the neigh- rithm that is proposed here, the branching degree is gradually
borhood relations. This approach is similar to the lbest model in decreased during a run of the algorithm. In order to decrease
the fact that only a certain fraction of the swarm is considered the branching degree from to , the hierarchy is traversed
for the velocity update of a particle. But in our algorithm the starting at the root node. This is done so that always one of
neighborhoods are constantly changing, according to the fitness the direct subtrees below the considered node is removed, if
development of the individuals. the number of children exceeds the new required branching
The changing arrangement of the particles can help pre- degree (see Fig. 2 and the next paragraph for an example). The
serving diversity in the search. In the described hierarchy decision about which subtree to remove is based on the quality
the arrangement of the particles leads to a different influence of the particles in the topmost nodes of all subtrees of the
for the particles at different positions. The particle with the considered nodes, i.e., all children of the considered node. This
currently best found solution can (indirectly) influence all the procedure is repeated for the entire tree. After this first phase
other particles after it has reached the top of the hierarchy. This of the branching degree reduction procedure the remaining tree
characteristic of H-PSO is similar to the gbest model. is of branching degree but has fewer nodes than before.
The removed nodes are then evenly inserted at the bottom of
B. Adapting the Hierarchy the hierarchy. Starting with the node on the second to last level
It is to be expected that the structure of the hierarchy, i.e., which has the least number of successors. The removed nodes
the branching degree , has a significant influence on the opti- are appended one by one so that the number of children of
mization behavior of H-PSO. For example, H-PSO with a high all nodes on the second last level differ by at most one. If all
branching degree might perform better in the beginning of the of these nodes have children, a new level is added to
optimization process because all particles are close to the top the hierarchy and the procedure is continued until all removed
particle in the hierarchy. Moreover, this tree topology has a small nodes are reinserted.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
JANSON AND MIDDENDORF: A HIERARCHICAL PARTICLE SWARM OPTIMIZER AND ITS ADAPTIVE VARIANT 1275
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
1276 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 35, NO. 6, DECEMBER 2005
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
JANSON AND MIDDENDORF: A HIERARCHICAL PARTICLE SWARM OPTIMIZER AND ITS ADAPTIVE VARIANT 1277
TABLE III
STEPS REQUIRED TO ACHIEVE A CERTAIN GOAL FOR PSO-g, PSO-l, H-PSO, H-PSO AND H-PSO. AVERAGE (AVG), MEDIAN (MED),
MAXIMUM (MAX), MINIMUM (MIN) AND EXPECTED (Exp := Avg/Succ) NUMBER OF ITERATIONS AND SIGNIFICANCE MATRIX; “ a” 0
0
AND “ b” DENOTE THE USED PARAMETER VALUES AS DESCRIBED IN SECTION IV
For evaluating the significance the Wilcoxon Rank Sum Test has samples but only the successful runs are considered, thus the
been used to compare the results for two algorithms. The results used sample sizes can differ. For two algorithms (X and Y) the
of the 100 test runs for two algorithms form two independent distribution of their results, and , are compared using
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
1278 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 35, NO. 6, DECEMBER 2005
B. Diversity
Fig. 3. Average ranks for unimodal functions for H-PSO for different branch
The hierarchy of H-PSO divides the swarm into several sub- m
levels ( = 40).
swarms of particles that are located in different subtrees. This di-
vision is not persistent, as the particles can move between these
subswarms. We study in this paper whether the different sub-
trees in a H-PSO hierarchy can still specialize to different re-
gions of the search space as intended by the design of H-PSO.
Therefore, the diversity of the particles in the subtrees has been
measured. The corresponding test runs were done with param-
eter values , , and . The resulting topology
consists of five subtrees below the root node, each subtree con-
taining six nodes. The diversity has been measured for each of
these five subtrees. Since the size of a particle set influences the
diversity values, we can not directly compare the diversity of the
particles within a subtree with the diversity of the whole swarm.
Therefore, the diversity within a subtree is also compared to the
diversity of a subset of six particles that have been selected ran-
domly with uniform probability from the whole swarm.
Determining the diversity within the swarm is based on
the “distance-to-average-point” measure given by Fig. 4. Average ranks for multimodal functions for H-PSO for different branch
(5), where is a subset of the swarm and the size of . m
levels ( = 40).
The problem dimension is denoted by and is the -th
component of . The average point of the particles in at As diversity measure of a subset of the swarm at
iteration is given as iteration measure is used.
Scaling by preserves the relations between dif-
ferent subsets, since all measures are scaled by the same value.
(5)
V. RESULTS
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
JANSON AND MIDDENDORF: A HIERARCHICAL PARTICLE SWARM OPTIMIZER AND ITS ADAPTIVE VARIANT 1279
Fig. 5. Sphere—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = 5) Fig. 7. Rastrigin—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = 5)
and AH-PSO; swarm size m = 40. and AH-PSO; swarm size m = 40.
Fig. 6. Rosenbrock—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = Fig. 8. Griewank—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = 5)
5) and AH-PSO; swarm size m = 40. and AH-PSO; swarm size m = 40.
The results for the two unimodal test functions show that
higher branch degrees lead to a better optimization behavior
over the entire duration of the test runs. Only for branching de-
grees higher than , the average ranks do not differ much
because for the hierarchies become very similar. For the
multimodal test functions, the results show a clearly different
behavior. In the beginning of the optimization process
again the H-PSO algorithms with higher branching degrees
achieved better ranks, but this tendency is inverted for later itera-
tions in which the smaller degrees are
better. This shows that a flexible topology in which the branch
degree changes during the optimization process might lead to
better results for multimodal test functions.
B. Topology
In this subsection, the H-PSO and AH-PSO algorithm (with Fig. 9. Schaffer—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = 5)
parameter values , ) are compared to PSO using the and AH-PSO; swarm size m = 40.
two neighborhood topologies gbest (PSO-g) and lbest (PSO-l).
The swarm size that has been used is . As shown in and Rosenbrock (Fig. 6) PSO-g performes better than PSO-l.
the last subsection, the optimization success of a neighborhood On the other hand, PSO-l achieves better results than PSO-g
topology (gbest, lbest) highly depends on the kind of function for Rastrigin (Fig. 7), Griewank (Fig. 8), Schaffer (Fig. 9) and
to be optimized (unimodal or multimodal). For Sphere (Fig. 5) Ackley (Fig. 10).
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
1280 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 35, NO. 6, DECEMBER 2005
C. Specialization
In this section, H-PSO and its weighted variants are com-
pared with PSO-l and PSO-g. The comparison is based on the
number of iterations required to reach a certain goal for each of
the test functions. The algorithms all used swarm size .
In Table III, the average, median, maximum and minimum it-
erations required to achieve the goal value are shown. Also, the
success rate and the expected number of iterations (average/suc-
cess rate) to reach the goal are given.
is always among the fastest algorithms to achieve
the desired goal. Only for the Rastrigin function, PSO-g-a does
require an average of 104.0 iterations, where takes
184.4. For all other test functions, obtains the lowest
average and expected number of iterations required to reach the
goal.
Fig. 10. Ackley—solution quality for PSO-g, PSO-l, H-PSO (h = 3, d = 5) The success rate of and is in general
and AH-PSO; swarm size m = 40. very high. An exception is the Ackley function for which
and (with ) achieve success rates of
The results show that the optimization behavior of H-PSO only 0.01 and 0.13, respectively. Also, PSO-g has a very low
does not depend so much on the specific function type. H-PSO success rate for this function in contrast to PSO-l which has a
produces competitive results for all the considered functions, high success rate. Therefore, and have
regardless of whether they are unimodal or multimodal. For all also been tested for a small branching degree of (these
test functions, H-PSO performs clearly better than the worst values are included in the table). For the small branching degree
one of PSO-l and PSO-g. On some test functions (Rastrigin and has a very high success rate and finds a solution of
Schaffer), it performs even better than both PSO versions. It the required quality on average in less iterations than the PSO-l
should be noted that these results have been obtained with the algorithms.
same fixed branch degree . Hence, the value can be The significance comparison shows that only for the Rast-
used as a recommendation for parameter choice. Nevertheless, rigin function PSO-g-a is significantly better than .
it can be expected that the results of H-PSO for a specific test performs better for the Sphere, Rosenbrock and
function can in general be improved by adapting the value Griewank function and also for the Ackley function, considering
of . that PSO-g-a only reached the goal once in all 100 test runs.
Using the adaptive H-PSO (AH-PSO) algorithm the results of The results show that can reach a required goal
H-PSO could be improved. For every test function the minimum significantly faster than PSO and H-PSO. This demonstrates that
branch degree of AH-PSO is set to the optimal branch H-PSO is a very promising algorithm. The results also indicate
degree of the respective function (determined by the experi- that a heterogenous swarm of particles with different values for
ments in Section V-A). The AH-PSO parameters are displayed seems advantageous.
as AH-PSO( , , ) for the minimum branch de- In order to explain the different optimization behavior of
gree , the decrease frequency and the decrease step and , recall that the parameter values
size . of are ranging from 0.4 to 0.729. Since the majority of the
For the unimodal test functions—Rosenbrock and swarm is located in lower levels of the hierarchy, most of the
Sphere—the AH-PSO algorithm that decreased the branch individuals of use . The average
degree only every 1000 iterations by 1 performs best. The for , with , and , is 0.692,
branch degree is only reduced to until the end of the compared to 0.437 for . This smaller impact of
algorithm. AH-PSO(2,1000,1) performs better than the best the previous velocity compared to the personal best and
performing H-PSO for the Rosenbrock function neighborhood best attractors in could explain why
and very similar to the best H-PSO for the Sphere converges faster than .
function. For all multimodal test problems, the best performing Another possible explanation can be based on the investiga-
AH-PSO algorithms reduce the branch degree to the optimum tions of the convergence behavior of a deterministic version of
degree very fast. In case the Rastrigin function AH-PSO(4,5,4) PSO. In [29] it was shown for the deterministic PSO that the
reaches the optimum degree after 20 iterations and it update equation is stable only if . Thus,
performs better than the best H-PSO . Also, for the for and , the particle tra-
other multimodal test functions, AH-PSO performs better than jectory would diverge (respectively converge). Hence, must
H-PSO with the optimum branch degree. It can be concluded be sufficiently large for a given value of and or the trajec-
that a fast adaption of the hierarchy during the initial phase tory is not stable. Assuming that the results of the deterministic
of a run seems to be advantageous for multimodal functions, PSO can be transferred, the particle on top of would
whereas optimization of unimodal functions can benefit from a not have a smooth trajectory but would jump around more ran-
slow decrease of over the entire running time. domly compared to the top particle of . But care
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
JANSON AND MIDDENDORF: A HIERARCHICAL PARTICLE SWARM OPTIMIZER AND ITS ADAPTIVE VARIANT 1281
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.
1282 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 35, NO. 6, DECEMBER 2005
VI. CONCLUSION [15] P. N. Suganthan, “Particle swarm optimizer with neighborhood oper-
ator,” in Proc. Congr. Evolutionary Computation (CEC 1999), 1999, pp.
In this paper we have introduced a hierarchical version of 1958–1962.
PSO, called H-PSO, in which the particles are arranged in a [16] J. Kennedy, “Stereotyping: improving particle swarm performance with
cluster analysis,” in Proc. Congr. Evolutionary Computation (CEC
dynamic hierarchy. This hierarchy gives the particles different 2000), 2000, pp. 1507–1512.
influence on the rest of the swarm with respect to their current [17] Y. Shi and R. C. Eberhart, “Parameter selection in particle swarm opti-
mization,” in Proc. Evolutionary Programming VII, vol. 1447, 1998, pp.
fitness. Hierarchies with different tree topologies have been 591–600.
studied for a standard set of test functions (Sphere, Rosenbrock, [18] T. Krink, J. S. Vesterstrøm, and J. Riget, “Particle swarm optimization
Rastrigin, Griewank, Schaffer’s f6, Ackley), and have been with spatial particle extension,” in Proc. IEEE Congr. Evolutionary
Computation (CEC 2002), 2002, pp. 1474–1479.
compared to several variants of PSO. The H-PSO algorithm [19] T. M. Blackwell, “Swarms in dynamic environments,” in Proc. Genetic
performed good on all of the considered test functions regard- and Evolutionary Computation (GECCO 2003), vol. 2723, 2003, pp.
less of their type (unimodal or multimodal). 1–12.
[20] J. Riget and J. S. Vesterstrøm, “A Diversity-Guided Particle Swarm Op-
Moreover, a variant of H-PSO (AH-PSO) with a dynamically timizer—The ARPSO,” Dept. Comput. Sci., Univ. of Aarhus, Aarhus,
changing branching degree of the tree topology has been intro- Denmark, Tech. Rep. 2002-02, 2002.
[21] A. Silva, A. Neves, and E. Costa, “An empirical comparison of particle
duced which could improve the performance of H-PSO. An- swarm and predator prey optimization,” in Proc. 13th Irish Int. Conf. Ar-
other extension of H-PSO is to use different values for the inertia tificial Intelligence and Cognitive Science, vol. 2464, 2002, pp. 103–110.
weight of the particles according to their level in the hier- [22] J. Kennedy, “Small worlds and mega-minds: effects of neighborhood
topology on particle swarm performance,” in Proc. Congr. Evolutionary
archy. It has been shown that this algorithm is able to reach a Computation (CEC 1999), vol. 3, 1999, pp. 1931–1938.
specified goal for every test function (except the Rastrigin func- [23] K. Veeramachaneni, T. Peram, C. Mohan, and L. A. Osadciw, “Op-
tion) faster than all other variants of PSO. timization using particle swarms with near neighbor interactions,” in
Proc. Genetic and Evolutionary Computation (GECCO 2003), vol.
In order to better understand the observed optimization be- 2723, 2003, pp. 110–121.
havior of H-PSO, the diversity of different subswarms of H-PSO [24] J. Kennedy and R. Mendes, “Neighborhood topologies in fully-informed
and best-of-neighborhood particle swarms,” in Proc. IEEE Int. Work-
and the number of swaps that occur within the hierarchy have shop on Soft Computing in Industrial Applications, 2003, pp. 45–50.
been examined. These observations indicate that, in spite of the [25] J. S. Vesterstrøm, J. Riget, and T. Krink, “Division of labor in particle
dynamic nature of the hierarchy, the subtrees are able to spe- swarm optimization,” in Proc. Congr. Evolutionary Computation (CEC
2002), 2002, pp. 1570–1575.
cialize to certain regions of the search space in case of multi- [26] M. Løvbjerg and T. Krink, “Extending particle swarm optimizers with
modal functions. self-organized criticality,” in Proc. Congr. Evolutionary Computation
One interesting topic for future research is to identify condi- (CEC 2002), 2002, pp. 1588–1593.
[27] J. Kennedy and W. M. Spears, “Matching algorithms to problems: an ex-
tions that could trigger the decrease of the branching degree for perimental test of the particle swarm and some genetic algorithms on the
AH-PSO so that it could autonomously adapt to the state of the multimodal problem generator,” in Proc. Int. Conf. Evolutionary Com-
putation, 1998, pp. 78–83.
optimization process. [28] I. C. Trelea, “The particle swarm optimization algorithm: convergence
analysis and parameter selection,” IPL: Inform. Process. Lett., vol. 85,
REFERENCES 2003.
[29] F. van den Bergh, “An Analysis of Particle Swarm Optimizers,” Ph.D.
[1] J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proc. dissertation, Dept. Comput. Sci., Univ. Pretoria, Pretoria, South Africa,
IEEE Int. Conf. Neural Networks, vol. 4, 1995, pp. 1942–1947. 2002.
[2] Handbook of Metaheuristics, F. Glover and G. Kochenberger, Eds.,
Kluwer , Norwell, MA, 2002.
[3] V. Cerný, “A thermodynamical approach to the traveling salesman
problem,” J. Optim. Theory Applicat., vol. 45, no. 1, pp. 41–51, 1985. Stefan Janson received the diploma in computer
[4] S. Kirkpatrick, C. D. Gelatt Jr., and M. P. Vecchi, “Optimization by sim- science from the University of Karlsruhe, Karlsruhe,
ulated annealing,” Science, vol. 220, pp. 671–680, 1983. Germany, in 2002.
[5] J. Holland, Adaptation in Natural and Artificial Systems. Ann Arbor, He is currently working on the German Research
MI: Univ. of Michigan Press, 1975. Council (DFG) project “Methods of Swarm Intelli-
[6] L. J. Fogel, A. J. Owens, and M. J. Walsh, Artificial Intelligence Through gence on Reconfigurable Architectures” at the Uni-
Simulated Evolution. New York: Wiley, 1966. versity of Leipzig, Leipzig, Germany.
[7] I. Rechenberg, Evolutionsstrategie—Optimierung technischer Systeme
nach Prinzipien der biologischen Information. Freiburg, Germany:
Fromman Verlag, 1973.
[8] H.-P. Schwefel, Numerical Optimization of Computer Models. Chich-
ester, U.K.: Wiley, 1981.
[9] M. Dorigo, V. Maniezzo, and A. Colorni, “Ant system: optimization by
a colony of cooperating agents,” IEEE Trans. Syst., Man, Cybern. B, vol.
Martin Middendorf (M’98) received the diploma
26, no. 1, pp. 29–41, Feb. 1996.
[10] M. Dorigo, G. Di Caro, and L. M. Gambardella, “Ant algorithms for degree in mathematics and the Dr.rer.nat. degree
discrete optimization,” Artif. Life, vol. 5, no. 2, pp. 137–172, 1999. from the University of Hannover, Hannover, Ger-
[11] M. Dorigo and T. Stützle, Ant Colony Optimization. Cambridge, MA: many, in 1988 and 1992, respectively. He also
MIT Press, 2004. received the Habilitation degree in 1998 from the
[12] M. Guntsch and M. Middendorf, “A population based approach for University of Karlsruhe, Karlsruhe, Germany.
ACO,” in Proc. 2nd Eur. Workshop on Evolutionary Computation in He has previously been with the University of
Combinatorial Optimization (EvoCOP-2002), vol. 2279, 2002, pp. Dortmund, Dortmund, Germany, and the University
72–81. of Hannover as a Visiting Professor of Computer
[13] J. Kennedy and R. C. Eberhart, “A discrete binary version of the particle Science. He was a Professor of Computer Science
swarm optimization,” in Proc. Conf. Systems, Man, and Cybernetics, at the Catholic University of Eichsttt, Eichsttt,
1997, pp. 4104–4109. Germany. Currently he is a Professor for Parallel Computing and Complex
[14] J. Kennedy and R. Mendes, “Population structure and particle swarm Systems with the University of Leipzig, Leipzig, Germany. His research
performance,” in Proc. Congr. Evolutionary Computation (CEC 2002), interests include reconfigurable architectures, parallel algorithms, algorithms
2002, pp. 1671–1676. from nature, and bioinformatics.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on March 10,2023 at 03:27:14 UTC from IEEE Xplore. Restrictions apply.