Uploaded by 锋锋

TDSC.2020.2980255

advertisement
Received July 16, 2016, accepted August 8, 2016, date of publication September 1, 2016, date of current version January 27, 2017.
Digital Object Identifier 10.1109/ACCESS.2016.2604738
Convergence Analysis and Improvement of the
Chicken Swarm Optimization Algorithm
DINGHUI WU, SHIPENG XU, AND FEI KONG
Key Laboratory of Advanced Process Control for Light Industry, Ministry of Education, School of Internet of Things Engineering,
Jiangnan University, Wuxi 214122, China
Corresponding author: D. Wu (wdh123@jiangnan.edu.cn)
This work was supported by the National Natural Science Foundation of China under Grant 61572237 and Grant 61573167.
ABSTRACT In this paper, the convergence analysis and the improvement of the chicken swarm
optimization (CSO) algorithm are investigated. The stochastic process theory is employed to establish the
Markov chain model for CSO whose state sequence is proved to be finite homogeneous Markov chain and
some properties of the Markov chain are analyzed. According to the convergence criteria of the random
search algorithms, the CSO algorithm is demonstrated to meet two convergence criteria, which ensures the
global convergence. For the problem that the CSO algorithm is easy to fall into local optimum in solving
high-dimensional problems, an improved CSO is proposed, in which the relevant parameters analysis and
the verification of optimization capability are made by lots of test functions in high-dimensional case.
INDEX TERMS Chicken swarm optimization, Markov chain, state transition, global convergence,
benchmark function.
I. INTRODUCTION
The swarm intelligence optimization algorithm is a metaheuristic algorithm based on population, which finds the
optimal solution of the problem through cooperation and
competition between individuals within populations [1], [2].
The swarm intelligence optimization algorithm as a new
evolutionary computation technology has become the focus
of a growing number of scholars, and lots of researchers
have proposed many swarm intelligence algorithms for optimization problems, such as cat swarm optimization [3], the
firefly algorithm [4], the wolf search algorithm [5], the monkey search algorithm [6] and the bat algorithm [7]. These
algorithms can be used to solve most optimization problems. The application field of these algorithms has been
extended to neural network training [8]–[10], multi-objective
optimization [11]–[13], data clustering [14], pattern recognition [15], [16], robot control [17], [18], dada mining [19], [20]
and system identification [21]–[23] etc. However, the basis of
these algorithms comes from the simulation of the biological
community, which lacks relevant theoretical analysis. What
is more, the parameter settings are empirically determined
without precise theoretical basis. So it is important to make
the theoretical research on the swarm intelligent optimization
algorithms.
The chicken swarm optimization is a meta-heuristic algorithm based on population, which mimics the behaviors of
the chicken swarm [24]. The whole swarm can be divided
9400
into several groups, each of which consists of one rooster
and many hens and chicks. Different chickens follow different
laws of motions. In particular hierarchy order, there is a competition between the various subgroups. But so far, no literature was found about the relevant theoretical research (such
as convergence analysis) on the chicken swarm optimization
algorithm. According to the problem that the chicken swarm
optimization is easy prone to premature convergence in solving high-dimensional optimization problems, an improved
chicken swarm optimization is proposed and the optimization
ability is tested by eight benchmark functions [25]. But the
relevant parameter analysis was not made, and test functions
used were too little, which cannot fully validate the optimization ability of the proposed algorithm. So there is an urgent
need for analyzing the convergence of the chicken swarm
optimization. In addition, based on the work in [25], the relevant parameters analysis and the verification of optimization
capability by lots of test functions are also necessary.
The Markov chain is a random process that plays an important role and has universal significance. It has a wide range of
applications. The Markov chain theory has a strong capability
in terms of convergence analysis of randomized algorithms
and probabilistic analysis, and has been successfully applied
to the simulated annealing [26], the PSO [27], the ant colony
algorithm [28] and the artificial bee colony algorithm [29].
In this paper, the Markov chain is applied to convergence
analysis of the chicken swarm optimization, by establishing
2169-3536 2016 IEEE. Translations and content mining are permitted for academic research only.
Personal use is also permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
the Markov chain model of the chicken swarm optimization to
study the state metastatic behavior of the chicken swarm optimization. Convergence performance of the chicken swarm
optimization was also analyzed according to randomized
algorithms convergence criteria. Based on the work in [25],
the relevant parameter analysis and the verification of the
optimization capability are also made by many test functions
in high-dimensional case.
The rest of this paper is organized as follows. In Section 2,
the chicken swarm optimization is introduced. The details
about the Markov chain model of the chicken swarm optimization are discussed in Section 3. The convergence analysis
of the chicken swarm optimization is presented in Section 4.
Section 5 gives the details of the improved chicken swarm
optimization and the improved chicken swarm optimization
for test functions is described in Section 6. In Section 7, some
conclusions and discussions are given.
II. THE CHICKEN SWARM OPTIMIZATION
The chicken swarm optimization mimics the hierarchal order
in the chicken swarm and the behaviors of the chicken swarm,
which stems from the observation of the birds foraging behavior. The chicken swarm can be divided into several groups,
each of which consists of one rooster, many hens, and several
chicks. When in foraging, the roosters can always find food
preferentially. The hens always follow the roosters to look for
food, and the chicks followed their mother in search of food.
The different individuals within the chicken swarm follow
different laws of motions. There exist competitions between
the different individuals within the chicken swarm under the
particular hierarchal order.
The position of each individual within the chicken swarm
represents a feasible solution of the optimization problem.
First define the following variables before describing location
update formula of the individuals within the chicken swarm:
RN , HN , CN and MN are the number of roosters, hens, chicks
and mother hens, respectively; N is the number of the whole
chicken swarm, D is the dimension of the search space;
xi,j (t) (i ∈ [1, · · · , N ], j ∈ [1, · · · , D]) is the position of each
individuality time t.
In the chicken swarm, the best RN chickens would be
assumed to be the roosters, while the worst CN ones would
be regarded as the chicks. The rest of the chicken swarm is
viewed as the hens. Roosters with best fitting values have
priority for food access than the ones with worse fitting
values. The position update equations of the roosters can be
formulated below:
xi,j (t + 1) = xi,j (t)(1 + Randn(0, σ 2 )),
(1)

1,
fi ≥ fk , k ∈ [1, N ], k 6 = i,
fk − fi
σ2 =
exp(
), otherwise,
|fi | + ε
(2)
where Randn(0, σ 2 ) is a Gaussian distribution with mean 0
and standard deviation σ , ε is the small constant to avoid
VOLUME 4, 2016
zero-division-error, and k is a rooster’s index, which is
randomly selected from the roosters group (k 6 = i), fi is the
fitness value of particle i.
As for the hens, they can follow their group-mate roosters
to search for food and randomly steal the food found by other
the individuals. The position update equations of the hens are
as follows:
t+1
t
t
t
Xi,j
= Xi,j
+ C1 Rand(Xr1,j
− Xi,j
)
t
t
− Xi,j
),
+ C2 Rand(Xr2,j
(3)
C1 = exp((fi − fr1 )/(|fi | + ε)),
(4)
C2 = exp((fr2 − fi )),
(5)
where Rand is a uniform random number over [0, 1], r1 is an
index of the rooster, which is the ith hen’s group-mate, and
r2 is the index of the chicken (rooster or hen), which is
randomly chosen from the chicken swarm r1 6 = r2 .
With respect to the chicks, they follow their mother to
forage for food. The position update equation of the chicks
is formulated below:
t+1
t
t
t
Xi,j
= Xi,j
+ F(Xm,j
− Xi,j
).
(6)
where xm,j (t) is the position of the chick’s mother,
F ∈ [0, 2] is the follow coefficient, which indicates that the
chick follows its mother to forage for food.
III. THE MARKOV CHAIN MODEL OF THE
CHICKEN SWARM OPTIMIZATION
For convenience, the dimension of the particle is not considered here. The simplified position update formulas are as
follows. The position update formula of the roosters is
xi (t + 1) = xi (t)(1 + R1 ).
(7)
where R1 is the fixed value between 0 and 1. The position
update formula of the hens is
xi (t + 1) = xi (t) + C1 R2 (xr1 (t) − xi (t))
+ C2 R3 (xr2 (t) − xi (t)),
(8)
where R2 and R3 are the fixed values between 0 and 1,
C2 < 1 < C1 .
The position update formula of the chicks is
xi (t + 1) = xi (t) + F(xm (t) − xi (t)).
(9)
Several definitions and mathematical description to illustrate
the Markov chain model of the chicken swarm optimization
are given in the following [31].
Definition 1 (The State of the Chicken and the State Space
of the Chicken): The chicken state consists of the location
of the chicken in the chicken swarm, denoting x. x ∈ A,
where A is the feasible solution space. All the possible states
of the chickens constitute the state space of chicken, denoting
X = {x|x ∈ A}.
Definition 2 (The State of the Chicken Swarm and the State
Space of the Chicken Swarm): The states of all the chickens
in the chicken flock constitute the state of the chicken swarm,
9401
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
denoting s = (x1 , x2 , · · · , xN ), where xi is the state of the
ith chicken and N is the total number of the individuals in the
chicken swarm. All the states of the chicken swarm constitute
the state space of the chicken swarm, and it is remembered as
S = {s = (x1 , x2 , · · · , xN )|xi ∈ A, 1 ≤ i ≤ N }.
Definition P
3 (The State Equivalence): For s ∈ S and x ∈ s,
let ϕ(s, x) = N χ|x|(xi ) , where χA is the characteristic function of event A and ϕ(s, x) are the individuals in the chicken
swarm state s contained in the state x. For two chicken swarm
states s1 , s2 ∈ S, x ∈ A, if ϕ(s1 , x) = ϕ(s2 , x), s1 is called to
be equivalent s2 , denoting s1 ∼ s2 .
Definition 4 (The State Equivalence Class): The chicken
swarm state equivalence class induced by equivalence relation in S calls L = S/ ∼ or chicken swarm equivalence class
for short.
The chickens equivalence class has the following
properties.
1) Any chicken swarm states in the equivalence class L
are equivalent each other.
2) Any one in is not equivalent with the one outside L.
3) There is no intersection between one equivalence class
and another.
Definition 5 (The State Transition of the Chicken): In the
chicken swarm optimization, for xi ∈ s, xj ∈ s, the state of
the chicken xi one step transmits to xj , denoting Ts (xi ) = xj .
Theorem 1: In the chicken swarm optimization, transition
probability of the state of the chicken xi one step transition
to xj , namely
p(Ts (xi ) = xj )


pr (Ts (xi ) = xj )
= ph (Ts (xi ) = xj )


pc (Ts (xi ) = xj )
achieved by the roosters,
(10)
achieved by the hens,
achieved by the chicks.
Proof: The chicken swarm is a set of points in hyperspace, so the chicken location update process is a transition between a set of points in hyperspace. According to
Definition 5 and the geometry nature of the chicken swarm
optimization, we can get one-step transition probability of the
roosters from xi to xj , namely

 1 , x ∈ [x , x + x R ],
j
i i
i 1
(11)
pr (Ts (xi ) = xj ) = |xi R1 |

0,
otherwise.
One-step transition probability of the hens from xi to xj is
ph (Ts (xi ) = xj ) =
1
,
|C1 R2 (xr1 − xi ) + C2 R3 (xr2 − xi )|
(12)
for xj ∈ [xi , xi + C1 R2 (xr1 − xi ) + C2 R3 (xr2 − xi )], 0 for
otherwise. One-step transition probability of the chicks from
xi to xj is
pc (Ts (xi ) = xj )

1

, xj ∈ [xi , xi + F(xm − xi )],
= |F(xm − xi )|
(13)

0,
otherwise.
9402
Definition 6 (The State Transition Probability of Chicken
Swarm Optimization): In the chicken swarm optimization,
for si ∈ S, sj ∈ S, the state of the chicken swarm si one
step transmits to sj , denoting TS (si ) = sj . The transition
probability of the state of the chicken swarm si one step
transition to sj is
p(Ts (si ) = sj ) =
N
Y
p(Ts (xin ) = xjn ),
(14)
n=1
where N is the number of the individuals in the chicken
swarm. That is to say, the transition probability of the state
of the chicken swarm si one step transition to sj is the sum
of the state transition probability of each individual of Si one
step transition to each individual of Sj .
Definition 7 (The Transition Probability of the State
Equivalence Class of Chicken Swarm): Assume that
Li = (si1 , si2 , · · · , sih ) and Lj = (si1 , si2 , · · · , sik ) are
the chicken swarm state equivalence class induced by the
equivalence relation in S. Li one-step transmits to Lj , denoting
TL (Li ) = Lj , one-step transition probability of which is
p(TL (Li ) = Lj ) =
k
h X
X
p(TS (Si f ) = Sj e).
(15)
f =1 e=1
It shows that the transition probability of the chicken swarm
equivalence class is the sum of the state transition probability
of each chicken swarm of Li one step transition to Lj .
IV. THE CONVERGENCE ANALYSIS OF THE
CHICKEN SWARM OPTIMIZATION
Definition 8 (The Markov Chain): Suppose that there is
a random process xn , n ∈ T , and the parameter set T is a
discrete time sequence, namely T = 0, 1, 2, · · ·. All possible
values x substitute the discrete state space I = {i0 , i1 , i2 , · · · }.
If the conditional probability p(xn+1 = in+1 |x0 = i0 ,
x1 = i1 , · · · , xn = in ) = p(xn+1 = in+1 |xn = in ) is met
for an arbitrary integer n ∈ T and {i0 , i1 , i2 , · · · , in+1 } ∈ I ,
{xn , n ∈ T } is called a Markov chain.
Definition 9 (The Finite Markov Chain): If the state space I
is finite, we call the chain the finite Markov chain.
Definition 10 (The Homogeneous Markov Chain):
p(xn+1 = in+1 |xn = in ) is the conditional probability that
the system state in in time n shifts to the new state in+1 . If the
probability is independent with time, we call the chain the
homogeneous Markov chain.
Theorem 2: The population sequence generated by the
chicken swarm optimization is a finite homogeneous Markov
chain.
Proof: First, prove the chicken swarm sequence is a
Markov process and then prove that the Markov chain satisfies the homogeneous and finiteness. According to Definition 6, among the chicken swarm state sequences s(t), t ≥ 0,
s(t) ∈ S and s(t + 1) ∈ S, the transition probabilities
p(TS (s(t)) = s(t + 1)) depends on the transition probabilities p(TS (x(t)) = x(t + 1)) of all the chickens in the
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
chicken swarm. From Theorem 1, the state transition probability p(TS (x(t)) = x(t + 1)) of each individual of the
chicken swarm is only related to the state x(t) of time t,
the random numbers R1 , R2 and R3 belong to [0, 1], the
random number F belongs to [0, 2], C1 , C1 , xr1 , xr2 and xm ,
so p(TS (x(t)) = x(t + 1)) is only associated with the
state at time t, not related to time t. In addition, according
to Definition 8, the population sequence generated by the
chicken swarm optimization possesses Markov sex. The sincere chicken swarm state space S is limited, according to
Definition 9, the population sequence {s(t), t ≥ 0} generated
by the chicken swarm optimization is a finite Markov chain.
From Theorem 1, p(TS (x(t)) = x(t + 1)) is only related to the
state at time t, and not associated with time t, so the population sequence produced by the chicken swarm optimization
is limited a homogeneous Markov chain. QED.
A. THE CONVERGENCE RULE
The chicken swarm optimization is a random search algorithm, so the convergence behaviors of the chicken swarm
optimization can be judged by the convergence rule of the
random search algorithm in [32]. For the optimization problem hA, f i, there exists a stochastic optimization algorithm D,
the iteration result of kth is xk , and the next iteration value is
xk+1 = D(xk , ζ ), where A is feasible solution space, f is the
fitting function and ζ is the iteration solution searched already
for the chicken swarm optimization.
Define the search infimum
α = inf{t|v(x ∈ A|f (x) < t) > 0},
(16)
where v(x) is the Lebesgue measure of the set x. Define the
optimal solution region
(
x ∈ A|f (x) < α + ε, α < ∞,
Rε,M =
(17)
x ∈ A|f (x) < −C,
α = −∞.
where ε > 0 and C is a sufficiently large positive. If the random algorithm D finds a point located in Rε,M , the algorithm
is thought to be approximate global optimal or acceptable
global optimal.
Condition H1 If ζ ∈ A, then f (D(x, ζ )) ≤ f (x).
Condition H2 For B ∈ A with v(B) > 0, there is
Q∞
k=0 (1−uk (B)) = 0, where uk (B) is the probability measure
of the kth iteration search solution of the algorithm D situated
in the set B.
Theorem 3 (The global convergence of the random
algorithm): Suppose that f is measurable and A is a measurable subset of Rn , if Conditions H1 and H2 are satisfied
and {xk }∞
k=0 is a sequence generated by the algorithm D, the probability measure P(xk ∈ Rε,M ) satisfies
limk→∞ P(xk ∈ ε, M ), where Rε,M is the optimal area. Then
the algorithm D is of the global convergence.
B. THE CONVERGENCE OF THE CHICKEN
SWARM OPTIMIZATION
Theorem 4: The chicken swarm optimization algorithm
satisfies Condition H1.
VOLUME 4, 2016
Proof: The chicken swarm optimization must update the
current best position of the individuals for iteration, namely
(
xi (t − 1), f (xi (t)) > f (xi (t − 1)),
xi (t) =
(18)
xi (t),
f (xi (t)) ≤ f (xi (t − 1)).
So that the chicken swarm optimization saves the best
position of the group for iteration and Condition H1 is
satisfied. QED.
Definition 11 (The Optimal State Set of Chicken Swarm):
Suppose that the optimal of the optimization problem hA, f i
is g∗ , define the optimal state set of chicken swarm as
G = {s = (x)|f (x) = f (g∗ ), s ∈ S}.
If G equals S, each solution in the feasible solution space
is not only a feasible solution but also an optimal one. At this
time, optimization is meaningless, so the following discussion is based on G ⊂ S.
Definition 12 (The Absorption State Markov Chain): Based
on the Markov chain {s(t), t ≥ 0} of the population sequence
of the chicken swarm optimization and the optimal state set
G ⊂ S, if {s(t), t ≥ 0} satisfies the conditional probability
p{xk+1 ∈
/ G|xk ∈ G} = 0, the Markov chain is called the
absorbing state Markov chains.
Theorem 5: For the chicken swarm optimization, the optimal state set G is a closed set of the state space S.
Proof: In the chicken swarm optimization, the best individual retention mechanism is adopted for the update strategy
of the optimal individual. That is to say, when the fitting
value of the current best individual is better than that of
the previous best individual, the previous individual will be
replaced, as shown in (18). Thus, this ensures that the new
individual is not worse than the old one for iteration of the
chicken swarm optimization, so the conditional probability
p(xk+1 ∈
/ G|xk ∈ G) = 0 is satisfied. That is, the population
sequence {s(t), t ≥ 0} generated by the algorithm is the
absorption state Markov chains. QED.
Theorem 6: For the chicken swarm optimization, the optimal chicken swarm state set G is a closed set of the state
space S.
Proof: For si ∈ G, si ∈
/ G, sj ∈ S, the state transiQN
tion probability TS (si ) = sj is p(TS (si ) = sj ) =
k=1 p
(TS (xik ) = xjk ). Thus at least one particle state is optimal.
Suppose that g∗ ∼ xi0 k is the optimal state. That is to say,
there exists at least xi0 k ∈ G such that p(TS (xi0 k ) = xjk ) = 0.
At this time p(TS (xi0 k ) = xjk ) = 0, so the optimal chicken
swarm state set G is a closed set of the state space S. QED
Theorem 7:
T There do not exist a non-empty closed set M
such that M G = ∅ in the chicken swarm state space S.
Proof: Suppose
that there is a non-empty closed set
T
subject to M G = ∅. Suppose that xi = (g∗ , g∗ , · · · ,
g∗ ) ∈ G and xj = (xj1 , xj2 , · · · , xjN ) such that
f (xjc ) > f (g∗ ). According to the Chapman-Kolmogorov
equation, it is concluded that
X
X
plSj ,Si =
···
p(TS (Sj ) = Sr1 )
Sr1 ∈S
Srl−1 ∈S
× p(TS (Sr1 ) = Sr2 ) · · · p(TS (Srl−1 ) = Si ).
(19)
9403
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
By limited iterations for the chicken swarm optimization,
Equations (11)–(13) in Theorem 1 are satisfied. So when
the step size is large enough, each item one-step transition
probability p(Ts (xrc+j )) = xrc+j+1 of the expansion in (19)
satisfies Equation (10), namely p(Ts (xrc+j )) = xrc+j+1 > 0, so
does plSj ,Si . The conclusion can be achieved that is not a closed
set and contradicts the assumption. Therefore, the Markov
chain of the chicken swarm state is irreducible, and the state
space does not contain the non-empty closed set. QED.
Theorem 8 [33]: For Markov chain, suppose that there
is non-empty closed set E, but thereTdoes not exist another
non-empty closed set O makes E O = ∅. If j ∈ E,
then limk→∞ p(xk = j) = 5j , and if j ∈
/ E, then
limk→∞ p(xk = j) = 0.
Theorem 9: When the internal iteration of the chicken
swarm approaches to infinity, the chicken swarm state
sequence will enter the optimal state set.
Proof: By using Theorems 6 to 8, we can draw the
conclusion that Theorem 9 is true. QED.
Theorem 10: The chicken swarm optimization converges
to a global optimal solution.
Proof: Based on Theorem 4, the chicken swarm optimization satisfies Condition H1. Moreover, from Theorem 9,
the probability that the chicken swarm optimization does not
search a global
Q optimal solution after countless times is zero,
so there is ∞
k=0 (1 − uk [B]) = 0, where uk [B] is the probability measure of the kth iteration search solution of the algorithm situated in the set B. The chicken swarm optimization
satisfies Condition H2. In the chicken swarm optimization,
the best individual retention mechanism is adopted for the
update strategy of the optimal individual. That is to say, when
the fitting value of the current best individual is better than
that of the previous best individual, the previous individual
will be replaced with (18), otherwise, the original individual
is retained. That is, when the iteration goes to infinity, we
have limk→∞ P(xk ∈ Rε,M ) = 1, where {xk }∞
k=0 is an
iterative sequence produced by the chicken swarm optimization. According to Theorem 3, we can draw the conclusion
that the chicken swarm optimization is a global convergence
algorithm. QED.
V. THE IMPROVED CHICKEN SWARM OPTIMIZATION
A. THE PERFORMANCE ANALYSIS FOR THE
CHICKEN SWARM OPTIMIZATION
In [24], the standard functions are only tested in lowdimensional cases, without involving high-dimensional
cases. Here, the chicken swarm optimization (CSO) algorithm is adopted to test four high-dimensional benchmark
functions and the comparison with other two algorithms is
made. The results are shown in Table 1, where the boldfaces
indicate a minimal value obtained for three algorithms.
From Table 1, we can see that although the results
obtained by the chicken swarm optimization are better than
that by PSO and BA, it has significantly deviated from
the global optimum, especially for Rosenbrock, the algorithm obviously falls into a local optimum. From the above
discussion, we can draw the conclusion that the chicken
swarm optimization is easy to prone to premature convergence when in solving high-dimensional optimization
problems.
B. THE IMPROVED CHICKEN SWARM OPTIMIZATION
In order to overcome the premature convergence of the
chicken swarm optimization in solving high-dimensional
optimization problems, the location update formula of the
individuals within the chicken swarm should be improved.
For equation (6), the chicks only follow their mothers to
search for food, but do not follow the rooster of the subgroup. Therefore, the chicks can only obtain the position
information of their own mother. The chicks will fall into a
local optimum when their mothers fall into a local optimum.
So the position update equation for chicks must be modified.
The modified position update equation of the chicks is as
follows:
xi,j (t + 1) = wxi,j (t) + F(xm,j (t) − xi,j (t))
+ C(xr,j (t) − xi,j (t)).
(20)
TABLE 1. The experiment results for three algorithms at D = 100.
9404
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
where m is the index of the mother of chick i, r is the index
of the rooster in the subgroup, C is the learning factor, which
indicates that the chicks learn from the rooster in the subgroup
and w is a self-learning coefficient for the chicks, which is
very similar to the inertia weight in PSO.
1) THE DETAILS FOR THE IMPROVED CHICKEN
SWARM OPTIMIZATION
The detail process for improved chicken swarm optimization
is formulated below:
Step 1 Initialize the chicken swarm x and the related
parameter (RN , HN , CN , MN , · · · ).
Step 2 Evaluate the fitting values of the chicken swarm x,
and initialize the personal best position pbest and the global
best position gbest.t = 1.
Step 3 If t mod G is 1, sort the fitting values of the individuals within the chicken swarm, and build the hierarchal
order of the chicken swarm; divide the whole chicken swarm
into several subgroups and ensure the relationship between
the hens and the chicks.
Step 4 Renew the position of the roosters, the hens and
the chicks using equations (1), (3) and (7), respectively, and
calculate the fitting values of the individuals.
Step 5 Update the personal best position pbest and the
global best position gbest.
Step 6 t = t + 1, if the iteration stop condition is
met, halt iteration and export the global optimum; otherwise,
go to step 3.
2) THE CONVERGENCE ANALYSIS OF THE IMPROVED
CHICKEN SWARM OPTIMIZATION
About the earlier work, Trelea analyzed the convergence
and parameter selection of the particle swarm optimization
algorithm [34]. Compared with the original chicken swarm
optimization, the position update equation of the chicks for
the improved chicken swarm optimization is modified. The
structure of the chicken swarm and domination relation of the
individuals has not changed, so the Markov chain model of
the improved chicken swarm optimization is the same as the
original chicken swarm optimization. The original chicken
swarm optimization has been proved to converge to the global
optimum by Theorem 10.
3) THE PARAMETER ANALYSIS
Firstly, four inertia weight (denote as w) update strategies are
described in [35], namely linear decreasing, opening upward
parabola descending, opening downward parabola decreasing
and exponentially descending. The update equations of the
inertia weight use equations (21)–(24). In the same iterations,
the algorithm produces different optimization results adopting different inertia weight update policies. Therefore, the
classic test functions namely Sphere, Rosenbrock, Griewank
and Rastrigrin are selected to verify the optimization performance of the algorithm when adopting different inertia
weight strategies. Parameter settings are as follows: D = 100,
Ngen = 500, Npop = 50, G = 10, Nrun = 50, wmax = 0.9,
wmin = 0.4, C = 0.4, RN = 0.2, HN = 0.6, MN = 0.1,
CN = N − RN − HN . The experiment results are shown
in Table 2.
As can be seen from Table 2, for Sphere, Rosenbrock
and Griewank, the improved chicken swarm optimization can
obtain a smaller mean and standard deviation and the average
run time is shortened by using exponentially descending
inertia weight strategy. For the test function Rastrigrin, the
improved chicken swarm optimization can get a smaller mean
and standard deviation by adopting exponential decreasing
inertia weight strategy, but the average run time of the algorithm is slightly more than the case of using other strategies.
The proposed algorithm can improve the convergence precision without affecting the convergence speed by employing
TABLE 2. The parameter estimates and errors (σ 2 = 0.102 , L = 2000).
VOLUME 4, 2016
9405
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
the exponential decreasing inertia weight strategy, so the
exponentially decreasing strategy is adopted to update the
inertia weight. The inertia weight w is exponentially decreasing from 0.9 to 0.4 with the increase of iterations. Four update
equations of w are as follows:
w = wmax − (wmax − wmin )gen/Ngen,
w = (wmax − wmin )(t/tmax )
(21)
2
+ (wmin − wmax )(2t/tmax ) + wmax ,
(22)
w = −(wmax − wmin )(gen/Ngen) + wmax ,
(23)
w = wmin (wmax /wmin )(1/(1+10t/Ngen)) ,
(24)
2
where wmin is the final value of learning coefficients, wmax is
the initial value of learning coefficients, t is the current value
of iterations, Ngen is the maximum of iterations.
Secondly, the parameter G should be taken to be an appropriate value for different optimization problems. If G is too
large, the algorithm cannot quickly converge to the global
optimum. On the contrary, if G is too small, the algorithm may
fall into a local optimum. In [24], G is given in a reasonable
range, but it did not give the details for the experiments. To do
this, the classic test functions Sphere, Rosenbrock, Griewank,
Rastrigrin and Ackley are selected to study the influence of
the different values of G on the optimization performance of
the algorithm. Since the values of G have a very wide range,
it is impossible to select all possible values of G to verify
the optimization performance of the algorithm. Here, G = 2,
G = 6, G = 10, G = 13, G = 16 and G = 20 are chosen to
validate the optimization performance of the algorithm.
The parameter settings for the improved chicken swarm
optimization are as follows: D = 100, Ngen = 200,
Npop = 50, Nrun = 50, wmax = 0.9, wmin = 0.4, C = 0.4,
RN = 0.2, HN = 0.6, MN = 0.1, CN = N − RN − HN . The
exponential decreasing strategy is adopted for inertia weight
of the improved chicken swarm optimization. The average
convergence curves of the five test functions are shown in
Figures 1 to 5, respectively, in which the abscissa represents
the maximum number of iterations and the ordinate denotes
the common logarithm of the fitting value.
As can be seen from Figure 1, when G = 10, the algorithm has faster convergence rate and higher convergence
FIGURE 1. The average convergence curve of Sphere function versus G.
9406
FIGURE 2. The average convergence curve of Rosenbrock function
versus G.
FIGURE 3. The average convergence curve of Griewank function versus G.
FIGURE 4. The average convergence curve of Rastrigrin function versus G.
accuracy. In Figure 2, synthesizing the convergence rate and
accuracy of the algorithm, when G = 10, the algorithm has
a higher convergence accuracy in ensuring the convergence
rate. In Figure 3, when G = 10, the algorithm has a faster
convergence speed and higher convergence accuracy. From
Figure 4, we can see that when G = 13, the algorithm
has a higher convergence accuracy. In Figure 5, in order
to obtain higher convergence accuracy, G should take 10.
In conclusion, taking into account all the standard test functions, and considering both the convergence rate and convergence accuracy of the algorithm, when G = 10, the improved
chicken swarm optimization shows better optimization capabilities for most test functions.
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
FIGURE 5. The average convergence curve of Ackley function versus G.
Furthermore, the learning factor (denote as C) has a significant impact on the optimization performance of the algorithm. If C is too large, the algorithm cannot quickly converge
to the global optimum. On the contrary, if C is too small,
the convergence accuracy of the algorithm will decrease.
Therefore, the classic test functions Sphere, Rosenbrock,
Griewank and Ackley are selected to study the influence of
the different values of C on the optimization performance of
the algorithm. In the paper, C = 0.2, C = 0.4, C = 0.6,
C = 0.8 and C = 1.0 were selected to validate the optimization performance of the algorithm.
The parameter settings for the improved chicken swarm
optimization are as follows: D = 10, Ngen = 200,
Npop = 50, Nrun = 50, wmax = 0.9, wmin = 0.4, G = 10,
RN = 0.2N , HN = 0.6N , MN = 0.1HN , CN = N −RN −HN .
An exponential decreasing strategy is adopted for the inertia
weight of the improved chicken swarm optimization. The
average convergence curves of the four test functions are
shown in Figures 6 to 9, where the abscissa represents the
maximum number of iterations and the ordinate denotes the
common logarithm of the fitting values.
FIGURE 7. The average convergence curve of Rosenbrock function
versus C.
FIGURE 8. The average convergence curve of Griewank function versus C.
FIGURE 9. The average convergence curve of Ackley function versus C.
FIGURE 6. The average convergence curve of Sphere function versus C.
As can be seen from Figure 6, when C = 0.4, the algorithm
has a higher convergence rate. In Figure 7, synthesizing convergence rate and accuracy of the algorithm, when C = 0.4,
the algorithm has a higher convergence accuracy. In Figure 8,
when C = 0.4, the algorithm has a faster convergence speed
and higher convergence accuracy. From Figure 9, we can see
VOLUME 4, 2016
that when C = 0.4, the algorithm has a higher convergence
accuracy. On the whole, taking into account all the standard
test functions, and considering both the convergence rate and
convergence accuracy of the algorithm, when C = 0.4, the
improved chicken swarm optimization show better optimization capabilities for the all test functions.
VI. THE IMPROVED CHICKEN SWARM OPTIMIZATION
FOR TEST FUNCTIONS
A. THE PARAMETER SETTING
To test the performance of the improved chicken swarm optimization (CSO) algorithm, fourteen canonical benchmark
9407
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
TABLE 3. Fourteen benchmark functions.
FIGURE 10. The average convergence curve for F1.
test functions shown in Table 3 are used here for comparison [36]. In this table, all of test functions are minimum problems, n is the dimension of test functions, the first function is
a simple modal function (it has a single local optimum that is
a global optimum) while the other functions are multimodal
functions with many local optima. The amount of local optimum or saddles increases with increasing complexity of the
function s, i.e. with increasing dimension. Therefore, these
test functions with different features are used to testify the
effectiveness of the improved chicken swarm optimization
(ICSO) algorithm.
The comparison with the standard PSO [37], BA [7]
and the chicken swarm optimization [24] is also performed.
To be fair, the four algorithms run 50 times independently
for each function respectively, and the size of each swarm
is 50. The maximum number of iterations for each algorithm
is 1000. All benchmark functions are tested at D = 150.
Other parameters for four algorithms are listed in Table 4.
FIGURE 11. The average convergence curve for F2.
B. THE SIMULATION ANALYSIS
Four algorithms run independently 50 times for each test
function respectively, to obtain the best value, the worst value,
the mean value, the standard deviation and the mean run
time of 50 times. The experiment results obtained are shown
in Table 5. In addition, four algorithms run independently
FIGURE 12. The average convergence curve for F3.
50 times respectively for fourteen test functions at D = 150,
and the average convergence curves obtained as shown
in Figs. 10 to 23.
TABLE 4. The related parameter values for four algorithms.
9408
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
TABLE 5. The comparison of the results of improved chicken swarm optimization and other algorithms for D = 150.
Boldface indicates a minimal value obtained for four
algorithms.
The best value and the average value may reflect
the convergence accuracy and optimization capabilities.
Table 5 shows that for the most standard test functions,
the improved chicken swarm optimization has a high conVOLUME 4, 2016
vergence accuracy, and is significantly better than the
other three algorithms. Especially for F1, F2, F4, F5,
F7, F8, F10, F11, F12, F13 and F14, the convergence
value obtained by the improved chicken swarm optimization algorithm are less than that of PSO, the chicken
swarm optimization and BA algorithm. Therefore, the the
9409
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
FIGURE 13. The average convergence curve for F4.
FIGURE 17. The average convergence curve for F8.
FIGURE 14. The average convergence curve for F5.
FIGURE 18. The average convergence curve for F9.
FIGURE 15. The average convergence curve for F6.
FIGURE 19. The average convergence curve for F10.
FIGURE 16. The average convergence curve for F7.
FIGURE 20. The average convergence curve for F11.
chicken swarm optimization algorithm has a good advantage in processing multimodal and high-dimension complex
functions.
The worst value and the standard deviation can reflect the
robustness of the algorithm and the ability of confrontation
local optimum. From Table 5, we can see that for F6 and F7,
9410
VOLUME 4, 2016
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
For the optimization problems that are less requirement care
computation time, the improved chicken swarm optimization
algorithm has obvious advantages because of its high convergence accuracy.
As is shown in Figures 10 to 23, the convergence accuracy
of the improved chicken swarm optimization is better than
that of the chicken swarm optimization, and much better than
that of PSO and BA. As for convergence rates, the improved
chicken swarm optimization has higher convergence rates
compared with the chicken swarm optimization, PSO and BA
for most test functions.
FIGURE 21. The average convergence curve for F12.
FIGURE 22. The average convergence curve for F13.
VII. CONCLUSIONS AND FUTURE WORK
Based on the chicken swarm optimization, the state spaces of
the chicken and chicken swarm are described. The Markov
chain model of the chicken swarm optimization is established
by defining the chicken swarm state transition sequence, and
a detailed analysis of the properties of the Markov chain is
made. Note that it is a finite homogeneous Markov chain. The
final transfer status of the chicken swarm state sequence is
analyzed, and the global convergence of the chicken swarm
optimization is verified by the convergence criteria of a
random search algorithm. According to the problem that
the chicken swarm optimization is easy to fall into a local
optimum in solving high-dimensional problems, an improved
chicken swarm optimization is proposed. The relevant parameter analysis and the verification of the optimization capability by test functions in high-dimensional case are made.
In our future work, the in-depth study of the theory will
be conducted for the chicken swarm optimization, such as
designing hybrid chicken swarm optimization algorithm for
high-dimensional complex optimization problems and further studying convergence of the algorithm by martingale
sequence theory.
REFERENCES
FIGURE 23. The average convergence curve for F14.
the worst values obtained by the improved chicken swarm
optimization algorithm are closer to the ideal optimum value,
and the standard deviations are minimal. Especially for F1,
F4, F7, F9, F11, F12, F13 and F14, the worst values gained
by the improved chicken swarm optimization algorithm are
approximately equal to the theoretical optimal value, and
the standard deviations are approximately zero. Thus the
improved chicken swarm optimization algorithm is robust in
the function optimization process.
From the average run time perspective, for all standard
test functions, the average run time of the improved
chicken swarm optimization is slightly longer than that of
PSO and BA, but less than the chicken swarm optimization.
VOLUME 4, 2016
[1] I. Fister, Jr., X.-S. Yang, I. Fister, J. Brest, and D. Fister, ‘‘A brief review
of nature-inspired algorithms for optimization,’’ ElektrotehniskiVestnik/Electrotech. Rev., vol. 80, no. 3, pp. 1–7, 2013.
[2] Z. Beheshti and S. M. Shamsuddin, ‘‘A review of population-based metaheuristic algorithms,’’ Int. J. Adv. Soft Comput. Appl., vol. 5, no. 1,
pp. 1–35, 2013.
[3] S. C. Chu, P. W. Tsai, and J. S. Pan, ‘‘Cat swarm optimization,’’ in Proc.
Pacific Rim Int. Conf. Artif. Intell., vol. 6. 2006, pp. 854–858.
[4] I. Fister, I. Fister, Jr., X.-S. Yang, and J. Brest, ‘‘A comprehensive review of
firefly algorithms,’’ Swarm Evol. Comput., vol. 13, no. 1, pp. 34–46, 2013.
[5] R. Tang, S. Fong, X. S. Yang, and S. Deb, ‘‘Wolf search algorithm
with ephemeral memory,’’ in Proc. Int. Conf. Digit. Inf. Manage., 2012,
pp. 165–172.
[6] A. Mucherino and O. Seref, ‘‘Monkey search: A novel metaheuristic
search for global optimization,’’ Data Mining, Syst. Anal. Optim. Biomed.,
vol. 953, no. 1, pp. 162–173, 2007.
[7] X. S. Yang, ‘‘A new metaheuristic bat-inspired algorithm,’’ in Studies in
Computational Intelligence, vol. 284. Cham, Switzerland: Springer, 2010,
pp. 65–74.
[8] S. N. Qasem and S. M. Shamsuddin, ‘‘Memetic Elitist Pareto differential
evolution algorithm based radial basis function networks for classification
problems,’’ Appl. Soft Comput., vol. 11, no. 8, pp. 5565–5581, 2011.
[9] S. N. Qasem and S. M. Shamsuddin, ‘‘Radial basis function network based
on time variant multi-objective particle swarm optimization for medical
diseases diagnosis,’’ Appl. Soft Comput., vol. 11, no. 1, pp. 1427–1438,
2011.
9411
D. Wu et al.: Convergence Analysis and Improvement of the CSO Algorithm
[10] N. S. Jaddi, S. Abdullah, and A. R. Hamdan, ‘‘Optimization of neural network model using modified bat-inspired algorithm,’’ Appl. Soft Comput.,
vol. 37, pp. 71–86, Dec. 2015.
[11] P. M. Pradhan and G. Panda, ‘‘Solving multiobjective problems using cat
swarm optimization,’’ Expert Syst. Appl., vol. 39, no. 3, pp. 2956–2964,
2012.
[12] X. S. Yang, ‘‘Multiobjective firefly algorithm for continuous optimization,’’ Eng. Comput., vol. 29, no. 2, pp. 1–10, 2013.
[13] R. Mallick, R. Ganguli, and M. S. Bhat, ‘‘Robust design of multiple
trailing edge flaps for helicopter vibration reduction: A multi-objective bat
algorithm approach,’’ Eng. Optim., vol. 47, no. 9, pp. 1243–1263, 2014.
[14] Y. Kumar and G. Sahoo, ‘‘A hybrid data clustering approach based on
improved cat swarm optimization and K -harmonic mean algorithm,’’ AI
Commun., vol. 28, no. 4, pp. 751–764, 2015.
[15] R. Senaratne, S. Halgamuge, and A. Hsu, ‘‘Face recognition by extending
elastic bunch graph matching with particle swarm optimization,’’ J. Multimedia, vol. 4, no. 4, pp. 204–214, 2009.
[16] K. Cao, X. Yang, X. J. Chen, J. M. Liang, and J. Tian, ‘‘A novel ant colony
optimization algorithm for large-distorted fingerprint matching,’’ Pattern
Recognit., vol. 45, no. 1, pp. 151–161, 2012.
[17] N. Nedic, V. Stojanovic, and V. Djordjevic, ‘‘Optimal control of hydraulically driven parallel robot platform based on firefly algorithm,’’ Nonlinear
Dyn., vol. 82, no. 3, pp. 1–17, 2015.
[18] M. Rahmani, A. Ghanbari, and M. M. Ettefagh, ‘‘Robust adaptive control
of a bio-inspired robot manipulator using bat algorithm,’’ Expert Syst.
Appl., vol. 56, pp. 164–176, Sep. 2016.
[19] T. Sousa, A. Silva, and A. Neves, ‘‘Particle swarm based data mining
algorithms for classification tasks,’’ Parallel Comput., vol. 30, nos. 5–6,
pp. 767–783, 2004.
[20] A. A. Freitas and J. Timmis, ‘‘Revisiting the foundations of artificial
immune systems for data mining,’’ IEEE Trans. Evol. Comput., vol. 11,
no. 4, pp. 521–540, Aug. 2007.
[21] G. Panda, P. M. Pradhan, and B. Majhi, ‘‘IIR system identification using
cat swarm optimization,’’ Expert Syst. Appl., Int. J., vol. 38, no. 10,
pp. 12671–12683, 2011.
[22] Y. Mao and F. Ding, ‘‘A novel parameter separation based identification algorithm for Hammerstein systems,’’ Appl. Math. Lett., vol. 60,
pp. 21–27, Oct. 2016.
[23] Y. Wang and F. Ding, ‘‘Novel data filtering based parameter identification
for multiple-input multiple-output systems using the auxiliary model,’’
Automatica, vol. 71, pp. 308–313, Sep. 2016.
[24] X. Meng, Y. Liu, X. Gao, and H. Zhang, ‘‘A new bio-inspired algorithm:
Chicken swarm optimization,’’ in Advances in Swarm Intelligence (Lecture
Notes in Computer Science), vol. 8794. Cham, Switzerland: Springer,
2014, pp. 86–94.
[25] D. Wu, F. Kong, W. Gao, and Y. Shen, ‘‘Improved chicken swarm optimization,’’ in Proc. IEEE Int. Conf. Cyber Technol. Autom., Control, Intell.
Syst., Jun. 2015, pp. 681–686.
[26] B. Gidas, ‘‘Nonstationary Markov chains and convergence of the annealing
algorithm,’’ J. Statist. Phys., vol. 39, nos. 1–2, pp. 73–131, 1985.
[27] Z. H. Ren, J. Wang, and Y. L. Gao, ‘‘The global convergence analysis of
particle swarm optimization algorithm based on Markov chain,’’ Control
Theory Appl., vol. 28, no. 4, pp. 462–466, 2011.
[28] Z. P. Su, J. G. Jiang, and C. Y. Liang, ‘‘An almost everywhere strong
convergence proof for a class of ant colony algorithm,’’ Acta Electron. Sin.,
vol. 37, no. 8, pp. 1646–1650, 2009.
[29] A. P. Ning and X. Y. Zhang, ‘‘Convergence analysis of artificial bee colony
algorithm,’’ Control Decision, vol. 28, no. 10, pp. 1554–1558, 2013.
[30] N. Li, ‘‘Analysis and application of particle swarm optimization,’’
Huazhong Univ. Sci. Technol., 2006.
[31] W. M. Qian, H. Y. Liang, and G. Q. Yang, Application of Stochastic
Process. Beijing, China: Higher Edu. Press, 2014.
[32] F. J. Solis and R. J.-B. Wets, ‘‘Minimization by random search techniques,’’
Math. Oper. Res., vol. 6, no. 1, pp. 19–30, Feb. 1981.
9412
[33] W. X. Zhang and Y. Liang, Math Foundation to GA. Xi’an, China: Xi’an
Jiaotong Univ. Press, 2003, pp. 67–87.
[34] I. C. Trelea, ‘‘The particle swarm optimization algorithm: Convergence
analysis and parameter selection,’’ Inf. Process. Lett., vol. 85, no. 6,
pp. 317–325, 2003.
[35] G. Chen, J. Jia, and Q. Han, ‘‘Study on the strategy of decreasing inertia
weight in particle swarm optimization algorithm,’’ J. Xi’an Jiao Tong Univ.,
vol. 40, no. 1, pp. 53–61, 2006.
[36] S. Y. Ho, L. S. Shu, and J. H. Chen, ‘‘Intelligent evolutionary algorithms
for large parameter optimization problems,’’ IEEE Trans. Evol. Comput.,
vol. 8, no. 6, pp. 522–541, Jun. 2004.
[37] M. Clerc and J. Kennedy, ‘‘The particle swarm—Explosion, stability, and
convergence in a multidimensional complex space,’’ IEEE Trans. Evol.
Comput., vol. 6, no. 1, pp. 58–73, Feb. 2002.
DINGHUI WU was born in Hefei, Anhui
Province, China. He received the Ph.D. degree in
control science and engineering from the School of
Internet of Things Engineering, Jiangnan University, Wuxi, China, in 2010. He was a Post-Doctoral
Fellow in the Textile Engineering in 2014. From
2014 to 2015, he was a Visiting Scholar with the
School of Computer and Electronic Engineering,
University of Denver, USA. He is currently an
Associate Professor and a Master Tutor with the
School of Internet of Things Engineering, Jiangnan University. His current
research interests include wind power generation technology, robot technology, embedded systems, and other aspects of the research and the teaching
of power electronics.
SHIPENG XU was born in Pingdingshan, Henan
Province, China. He received the B.Sc. degree
from the Zhengzhou University of Light Industry,
Zhengzhou, China, in 2014. He is currently pursuing the M.Sc. degree with the School of Internet
of Things Engineering, Jiangnan University, Wuxi,
China. His interests include intelligent optimization and shop scheduling.
FEI KONG was born in Hefei, Anhui Province,
China. He received the M.Sc. degree from
the School of Internet of Things Engineering,
Jiangnan University, Wuxi, China, in 2016.
His interests include intelligent optimization and
shop scheduling.
VOLUME 4, 2016
Download