Professional Documents
Culture Documents
R. Rastegar
M. R. Meybodi
Abstract:
1. Introduction
Particle Swarm Optimization (PSO) technique was
first proposed by Kennedy and Eberhart [6] in 1995.
This technique is inspired by choreography of bird flock
and can be regarded as a distributed behavior that
perform multidimensional search. According to PSO,
the behavior of each particle is affected by either the
best local or the best global particle to help it fly
through a search space. Moreover, a particle can learn
from its past experience to adjust its flying speed and
direction. Therefore, by observing the behavior of the
flock and memorizing their flying histories all particle
in swarm can quickly converge to near optimal
geographical with a well preserved population density
distribution [7]. PSO is considered as an evolutionary
computation approach in that it possesses many
characteristics that is used by evolutionary algorithms
such as, initializing with a population of random
solutions, searching for optima by updating generations,
the adjustment of particles and evaluating particles by a
fitness function. However unlike evolutionary
algorithms, the updates of particles are not
accomplished by crossover or mutation. The particle
swarm algorithms reported in the literatures are
classified into two groups: discrete PSO and continuous
PSO [10][4][6]. In continuous PSO the particles operate
K. Badie
Information Technology
Department
Iran Telecom. Research Center
Tehran, Iran
k badie@itrc.ac.ir
_
Else
Vmax
xij=O
End if
End for
1. Pseudo
code of DPSO
3. Learning Automata
Learning Automata are adaptive decision-making
devices operating on unknown random environments.
The Learning Automaton has a finite set of actions and
each action has a certain probability (unknown for the
automaton) of getting rewarded by the environment of
the automaton. The aim is to learn to choose the optimal
action (i.e. the action with the highest probability of
being rewarded) through repeated interaction on the
system. If the learning algorithm is chosen properly,
then the iterative process of interacting on the
environment can be made to result in selection of the
optimal action. Figure 1 illustrates how a stochastic
automaton works in feedback connection with a random
environment. Learning Automata can be classified into
two main families: fixed structure learning automata
and variable structure learning automata (VSLA)
[12][14]. In the following, the variable structure
learning automata is described.
A VSLA is a quintuple <a,p,p,T(a,p,p) >, where a,
p, p are an action set with s actions, an environment
response set and the probability set p containing s
probabilities, each being the probability of performing
every action in the current internal automaton state,
respectively. The function of T is the reinforcement
algorithm, which modifies the action probability vector
p with respect to the performed action and received
response. Let a VSLA operate in an environment with
Pi(t + 1)
P j;'i(t+l)
(1-a)Pj(t)
IffJ(t)=l (Penalty).
(4)
Environment
Learning Automata
Figure
pi(t+l)=(1-b)pi(t)
P j*i(t+l)=(bjs-l)+(1-b)Pj(t)
a(n)
...
actions
While not done
FIgure
/3.
lj
{O
if LBi)
GBi)
xi)
(5)
1 otherwise
5. Simulations
This section presents simulation results and compares
the LAPSO with DPSO, in terms of solution quality, the
number of function evaluations taken and the speed of
finding the best solution for a given population size.
Each quantity of the results reported is the average
taken over 20 runs. The parameters used for DPSO are
the same as those used in [9], i.e. V is set to 6, and c l
and c2 are equal to 1. The population size varies from 2
to 50 with increments of two. The proposed Algorithm
is tested for different learning algorithms: LRI, and LRP
For the sake of convenience in presentation, we use
LA(automata)-PSO to refer to the LA-PSO algorithm
when it uses Learning automata automata. The
algorithm terminates when the number of iterations
succeeds 500. The proposed algorithms are tested on
five different standard functions to be minimized. These
functions are given below are and are taken from De
Jong's dissertation [5].
max
Ll ix: + Gaus.(O,l)
t
)1
-1.28 Xi 1.28
-65.536$.;,; $.65.536
Experiment 1:
6. Conclusion
In this paper a learning automata based discrete
particle swarm optimization algorithm (LA-PSO) was
proposed. In the proposed algorithm the learning automata
is used to determine the position of a particle that
confirms with the global goal of the group of the particles.
This happens by updating the actions probabilities of the
set of learning automata associated to the particle in such
a manner that the group goal will be achieved. Simulation
results showed the effectiveness of the proposed algorithm
in solving the optimization problems.
References
[I] Abido, M. A., "Particle swarm optimization for
multimachine power system stabilizer design", Power
Engineering Society Summer Meeting, Vol. 3, PP. 1346-1359,
2001.
[2] Blackwell, T. and Bentley, P. J., "Improvised music with
swarms", Proceedings of the IEEE Congress on Evolutionary
Computation 2002 Honolulu, Hawaii USA, 2002b.
[3] Ciuprina, G., loan, D., and Munteanu, I., "Use of intelligent
particle swarms optimization in electromagnetics", IEEE
Transactions on Magnetics, Vol. 38, No.2, PP. 1037-1040.
2002.
[4] Clerc, M., and Kennedy, J., "The Particle Swarm-Explosion,
Stability, and Convergence in a Multidimensional Complex
Space", IEEE Transaction on Evolutionary Computation, Vol. 6,
No. I, PP. 58-73, February 2002.
[5] De Jong, K. A., "The Analysis of the behavior of a class of
genetic adaptive systems" Ph.D. dissertation, University of
Michigan, Ann Arbor, 1975.
[6] Eberhart, R. C. and Kennedy, J., "Particle Swarm
Optimization", in Proceedings of IEEE International Conference
on Neural Networks, Vol 4, PP. 1942-1948, IEEE Service
Center, Piscataway, NJ, 1995.
[7] Gary, G, Y., and Haiming, Lu,. "Dynamical Population
Strategy assisted Particle Swarm Optimization", Proceedings of
the 2003 IEEE International Symposium on Intelligent Control,
Houston, Texas, PP. 697-702, October 2003.
[8] Hu, X., and Eberhart, R. C., "Solving constrained nonlinear
optimization problems with particle swarm optimization",
Proceedings of the Sixth World Multiconference on Systemic,
Cybernetics and Informatics 2002 (SCI 2002), Orlando, USA,
2002.
[9] Kennedy, J., and Eberhart, R. C., "A Discrete Binary Version
of The Particle Swarm Algorithm", in Proceedings of
Conference on Systems, Man, and Cybernetics, PP. 4104-4108,
IEEE Service Center, Piscataway, NJ, 1997.
[10] Kennedy, J., "The Particle Swarm: Social Adaptation of
Knowledge", in Proceedings of IEEE International Conference
on Neural Networks (Indianapolis, Indiana), pp. 303-308, IEEE
Service Center, Piscataway, NJ, 1997.
[11] Laskari, E. C., Parsopoulos, K. E., and Vrahatis, M. N.,
"Particle swarm optimization for minimax Problems",
Proceedings of the IEEE Congress on Evolutionary computation
2002 Honolulu, Hawaii USA, 2002b.
[12] Narendra, K. S., and Thathachar, M. A. L., Learning
Automata: An Introduction, Printice-Hall Inc, 1989.
[13] Schoofs, L., and Naudts, B., "Swarm Intelligence on the
Binary Constraint Satisfaction problem", Proceedings of the
IEEE Congress on Evolutionary Computation 2002 Honolulu,
Hawaii USA, 2002.
[14] Thathachar, M. A. L., Sastry, P. S., "Varieties of Learning
Automata: An Overview", IEEE Transaction on Systems, Man,
and Cybernetics-Part B: Cybernetics, Vol. 32, No. 6, PP. 711722, 2002.
40
12000
35
10000
8000
30
6000
25
20
15
4000
:::00
""'<6
==-
10
2000
==-
"u:
:is'
c::>
5
a
='
=
at
Figure 4. Effect of the population size and the reward parameter on the effectiveness of LA(LRI)-PSO for function F1 (a) the number of function evaluations
taken in 500 iterations to obtain the best solution. (b) the average of the best solutions obtained.
4500
200
180
160
140
120
100
80
60
40
20
a
4000
3500
3000
2500
2000
1 500
1000
-rift\J:lcmW
500
o
Da
Figure 5. Effect of the population size and the reward parameter on the effectiveness of LA(LRP)-PSO for function F2 (a) the number of function evaluations
taken in 500 iterations to obtain the best solution. (b) the average of the best solutions obtained.
10000
400
9000
350
8000
c:::::>
(a)
7000
6000
5000
4000
3000
2000
1000
a
-=
300
c::>
=
""'<6
=c:u
-=
c::>
"'0
-=
..2
250
200
150
100
50
=';==';==;--'+- a
at
(b)
Figure 6. Effect of the population size and the reward parameter on the effectiveness of LA(LRP)-PSO for function F5 (a) the number of function evaluations
taken in 500 iterations to obtain the best solution. (b) the average of the best solutions obtained.
12000
009
0.08
'"
0
l'l
'"
;l
..
'"
:,C4;;............................................................................
2000+
:Ii'
0
0.07
0.00
005
0.04
0.03
002
0.01
0
001
'.
population $11:.
population silt<!
Figure 7. Comparison of the DPSO, LA(LR/)-PSO, and LA(LRP)-PSO algorithms for function FI (a) the number of function evaluations taken to obtain the best
solution in 500 iterations. (b) the solution quality obtained at the end of the runs.
0.4 5
7000
0.4
BOOO
co
:s
'iil
"
5000
j"
3000
'1a
'"
'"
'"
.2
ti
c
0.35
03
.. 0.25
::!
4000
;l
'"
2000
:Ii'
0
1000
0.2
0.15
0.1
0. 05
0
...
population size
0/
population size
(a)
(b)
Figure 8. Comparison of the DPSO, LA(LR/)-PSO, and LA(LRP)-PSO algorithms for function F2 (a) the number of function evaluations taken to obtain the best
solution in 500 iterations. (b) the solution quality obtained at the end of the runs.
35
30000
25000
i'"
..
25 +
. .. ..............................................................................................................................................................................................
20000
1
Ji!
00 h.............................................................................................................................................................
+ .....................................................................................................................................................................................::
.............................................................................
;;;.:'..':;
:"; ",.".
,.
al
i 10 +\.........
."
10000
...
".
....................... ;
::
20+,
........................................................................................................................................
15 1-,
------
;"' ......................................................................................
F;:l:Rli:;;;;iiiQiTI
'c .......,:.:....:;..:......................!
5000
o
...
0/
population size
(a)
...
0/
population $ia..
(b)
Figure 9. Comparison of the DPSO, LA(LRI)-PSO, and LA(LRP)-PSO algorithms for function F4 (a) the number of function evaluations taken to obtain the best
solution in 500 iterations. (b) the solution quality obtained at the end of the runs.
14
14000
12000
i
Jl.
+ ................................................................................................................................................ "........./'.
10000 t
......................................................,KI
,'.:..
12 t," ,
;:
10+
r!l
'Wi
8000
6000 t
/."..;; ; "'....................................................................
.;-."....
4 000+ ;Ik:...........................................(,=,=c,.,
2000 +
: .",...,..........................................................................................
population size
(a)
:...................................................................................................
-------------------_.
....................................................................................................................................
8 +.." .
(\ ,
. ...
.
population sin
(b)
Figure 10. Comparison of the DPSO, LA(LRI)-PSO, and LA(LRP)-PSO algorithms for function F5 (a) the number of function evaluations taken to obtain the best
solution in 500 iterations. (b) the solution quality obtained at the end of the runs.