Electronic Journal of Differential Equations, Vol. 2007(2007), No. 89, pp.... ISSN: 1072-6691. URL: or

advertisement
Electronic Journal of Differential Equations, Vol. 2007(2007), No. 89, pp. 1–12.
ISSN: 1072-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu
ftp ejde.math.txstate.edu (login: ftp)
GLOBAL ROBUST DISSIPATIVITY OF
INTEGRO-DIFFERENTIAL SYSTEMS MODELLING NEURAL
NETWORKS WITH TIME DELAY
BAOTONG CUI, XUYANG LOU
Abstract. In this paper, the global robust dissipativity of integro-differential
systems modeling neural networks with time delay is studied. Several sufficient
conditions are derived to ensure the global robust dissipativity of the neural
networks proposed by using proper Lyapunov functionals and some analytic
techniques. The results are shown to improve the previous global dissipativity
results derived in the literature. Some examples are given to illustrate the
correctness of our results.
1. Introduction
Stability of neural networks with delays has been a popular problem in this
decade because the stability properties of equilibrium points of neural systems play
an important role in some practical problems, such as optimization solvers, associative memories, image compression, speed detection of moving objects, processing of moving images, and pattern classification. Therefore, the stability of
neural networks with or without delay has received much attention in the literature
[2, 3, 4, 5, 9, 10, 11, 14, 15]. As pointed out in [1, 6, 8], the global dissipativity
is also an important concept in dynamical neural networks. The concept of global
dissipativity in dynamical systems is a more general concept and it has found applications in the areas such as stability theory, chaos and synchronization theory,
system norm estimation, and robust control [8]. The global dissipativity of several
classes of neural networks were discussed, and some sufficient conditions for the
global dissipativity of neural networks with constant delays are derived in [8]. The
authors of [1] analyzed the global dissipation of continuous-time dynamical neural
networks with time delay. However, parameter fluctuation in neural network implementation on very large scale integration (VLSI) chips is unavoidable. To the best
of our knowledge, there does not seem to be much (if any) study on global robust
2000 Mathematics Subject Classification. 45J05, 92B20.
Key words and phrases. Neural network; integro-differential system; global dissipativity;
Lyapunov functional.
c
2007
Texas State University - San Marcos.
Submitted March 22, 2007. Published June 15, 2007.
Supported by grants 60674026 from the National Natural Science Foundation of China,
and 107058 from the Key Research Foundation of Science and Technology of the
Ministry of Education of China.
1
2
B. CUI, X. LOU
EJDE-2007/89
dissipativity of delayed integro-differential systems modelling neural networks with
uncertainties.
Motivated by above discussion, in this paper, we will study the global robust
dissipativity of delayed integro-differential systems modelling neural networks with
uncertainties and derive some criteria for the delayed neural network with uncertainties via Lyapunov functionals and analysis techniques.
The organization of this paper is as follows. In Section 2, problem formulation
and preliminaries are given. In Section 3, some new results are given to ascertain
the global robust dissipativity of the delayed neural networks based on Lyapunov
method and we give concluding remarks of results. Section 4 gives some examples
to illustrate the effectiveness of our results. A conclusion is drawn in Section 5.
2. Problem formulations and preliminaries
Consider the model of continuous-time neural networks described by the following integro-differential systems
n
n
X
X
dxi (t)
= −di xi (t) +
aij fj (xj (t)) +
bij fj (xj (t − τ ))
dt
j=1
j=1
Z t
n
X
cij
+
Kij (t − s)fj (xj (s))ds + ui , i = 1, 2, . . . , n
j=1
(2.1)
−∞
or
dx(t)
= −Dx(t) + Af (x(t)) + Bf (x(t − τ ))
dt
Z t
+C
K(t − s)f (x(s))ds + u,
(2.2)
−∞
where n denotes the number of the neurons in the network, xi (t) is the state of
the ith neuron at time t, f (x(t)) = [f1 (x1 (t)), f2 (x2 (t)), . . . , fn (xn (t))]T denote the
activation functions of the j-th neuron at time t, D = diag(d1 , d2 , . . . , dn ) > 0 is
a positive diagonal matrix, A = (aij )n×n , B = (bij )n×n and C = (cij )n×n are the
feedback matrix and the delayed feedback matrix, respectively, τ is the transmission
delay, u = (u1 , u2 , . . . , un )T ∈ Rn be a constant external input vector.
In our analysis, we will employ the following two classes of activation functions:
(1) The general set of monotone nondecreasing activation functions is defined as
G = {f (x)|D+ fi (xi ) ≥ 0, i = 1, 2, . . . , n}.
(2) The set of Lipschitz-continuous activation functions is defined as
L = {f (x)|0 ≤
fi (xi ) − fi (yi )
≤ Li , Li > 0, }
xi − yi
for all xi , yi ∈ R, xi 6= yi , i = 1, 2, . . . , n.
K(·) = (Kij (·))n×n , i, j = 1, 2, . . . , n are the delay kernels which are assumed to
satisfy the following conditions simultaneously:
(i)
(ii)
(iii)
(iv)
Kij : [0, ∞) → [0, ∞);
ij are bounded and continuous on [0, ∞);
RK∞
Kij (s)ds = 1;
0
R∞
there exists a positive number ε such that 0 Kij (s)eεs ds < ∞.
EJDE-2007/89
GLOBAL ROBUST DISSIPATIVITY
3
Some examples of functions that meet the above conditions can be found in [12].
Clearly, if one assume that cij = 0 (i, j = 1, 2, . . . , n), the system (2.1) reduces
to the dynamical neural networks model
n
n
X
X
dxi (t)
= −di xi (t) +
aij fj (xj (t)) +
bij fj (xj (t − τ )) + ui ,
dt
j=1
j=1
(2.3)
whose global dissipativity was investigated in [1].
The initial conditions associated with the system (2.1) are
xi (s) = φi (s),
−τ ≤ s ≤ 0, i = 1, 2, . . . , n,
(2.4)
where φi (·) is bounded and continuous on [−τ, 0].
In electronic implementation of neural networks, the values of the constant and
weight coefficients depend on the resistance and capacitance values which are subject to uncertainties. This may lead some deviations in the values of di , aij , bij and
cij . Hence, it is important to ensure the global robust dissipativity of the designed
network against such parameter deviations. Since these deviations are bounded
in practice, the quantities di , aij , bij and cij may be intervalized as follows: for
i = 1, 2, . . . n,
DI := [D, D] = D = diag(di ) : D ≤ D ≤ D,
i.e., ci ≤ ci ≤ ci ,
AI := [A, A] = A = (aij )n×n : A ≤ A ≤ A,
i.e., aij ≤ aij ≤ aij ,
BI := [B, B] = B = (bij )n×n : B ≤ B ≤ B,
i.e., bij ≤ bij ≤ bij ,
CI := [C, C] = C = (cij )n×n : C ≤ C ≤ C,
i.e., cij ≤ cij ≤ cij .
Definition 2.1. The neural network defined by (2.1) or (2.2) is said to be a dissipative system, if there exists a compact set S ⊂ Rn , such that ∀x0 ∈ Rn , ∃T > 0,
when t ≥ t0 + T , x(t, t0 , x0 ) ⊆ S, where x(t, t0 , x0 ) denotes the solution of (2.1)
from initial state x0 and initial time t0 . In this case, S is called a globally attractive
set. A set S is called positive invariant if ∀x0 ∈ S implies x(t, t0 , x0 ) ⊆ S for t ≥ t0 .
Definition 2.2. The neural network defined by (2.1) or (2.2) is globally robust
dissipative if the system is globally dissipative for all D ∈ DI , A ∈ AI , B ∈ BI ,
C ∈ CI .
Definition 2.3 ([15]). A vector v = [v1 , . . . , vn ]T > 0, if and only if every vi >
0, i = 1, . . . , n.
Definition 2.4. If R → R is a continuous function, then the upper right derivative
D + f (t)
of f (t) is defined as
dt
D+ f (t) = lim+
θ→0
f (t + θ) − f (t)
.
θ
(2.5)
3. Global robust dissipativity
In this section, we present the following results:
Theorem 3.1. Let f (·) ∈ G, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. the neural
network defined by (2.1) is a globally robust dissipative system and the set S1 =
4
B. CUI, X. LOU
EJDE-2007/89
{x : |xi (t)| ≤ |ui |/di , i = 1, 2, . . . , n} is a positive invariant and globally attractive
set, if there exist positive constants pi > 0, i = 1, 2, . . . , n such that
n
X
pi (−aii − b∗ii − c∗ii ) −
pj (a∗ji + b∗ji + c∗ji ) ≥ 0,
(3.1)
j=1,j6=i
where i = 1, . . . , n, a∗ji = max(|aji |, aji ), b∗ji = max(|bji |, bji ), c∗ji = max(|cji |, cji ).
Proof. Let us use the positive and radially unbounded Lyapunov functional
n
n Z t
X
X
V (x(t)) =
pi |xi (t)| +
b∗ij |fj (xj (s))|ds
i=1
n
X
+
j=1
c∗ij
j=1
∞
Z
t−τ
Z
Kij (s)
0
(3.2)
t
|fj (xj (ξ))|dξ ds .
t−s
Calculating the upper right derivative D+ V of the Lyapunov functional V along
the solution of (2.1), we have that
D+ V (x(t))
≤−
n
X
{pi di |xi (t)| − pi (aii + b∗ii )|fi (xi (t))| −
i=1
− pi
+ pi
n
X
j=1
n
X
Z
|cij |
≤−
−
≤−
−
Kij (t − s)|fj (xj (s))|ds −
−∞
c∗ij
j=1
n
X
t
Z
j=1,j6=i
n
X
pi
c∗ij
j=1
Z
∞
Kij (s)|fj (xj (t))|ds
0
∞
Kij (s)|fj (xj (t − s))|ds − pi |ui |}
0
pi (a∗ij + b∗ij + c∗ij )|fj (xj (t))| − pi |ui |}
j=1,j6=i
n
X
n
X
i=1
i=1
pi (di |xi (t)| − |ui |) −
n
X
{pi (−aii − b∗ii − c∗ii )
pj (a∗ji + b∗ji + c∗ji )}|fi (xi (t))|
j=1,j6=i
n
X
pi di |xi (t)| −
i=1
pi (|aij | + b∗ij )|fj (xj (t))|
{pi di |xi (t)| − pi (aii + b∗ii + c∗ii )|fi (xi (t))|
i=1
n
X
≤−
n
X
|ui | < 0,
di
when xi ∈ Rn \S1 ; i.e., x∈S1 . Which implies that for all x0 ∈ S1 holds x(t, t0 , x0 ) ⊆
S1 , t ≥ t0 . For x∈S1 , there exists T > 0 such that
x(t, t0 , x0 ) ⊆ S1 ,
∀t ≥ t0 + T,
meaning that the neural network defined by (2.1) is a dissipative system and the
set S1 is a positive invariant set and globally attractive set.
EJDE-2007/89
GLOBAL ROBUST DISSIPATIVITY
5
Remark 3.2. As we can see, if vary the parameters pi > 0, (i = 1, 2, . . . , n)
properly, one can obtain a series of corollaries. Specially, If choose pi = 1, the
following corollary can be derived.
Corollary 3.3. Let f (·) ∈ G, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. the neural
network defined by (2.1) is a globally robust dissipative system and the set S1 =
{x||xi (t)| ≤ |ui |/di , i = 1, 2, . . . , n} is a positive invariant and globally attractive
set, if the following condition holds
n
X
−aii − b∗ii − c∗ii −
(a∗ji + b∗ji + c∗ji ) ≥ 0,
(3.3)
j=1,j6=i
where i = 1, . . . , n,
a∗ji
= max(|aji |, aji ), b∗ji = max(|bji |, bji ), c∗ji = max(|cji |, cji ).
Theorem 3.4. Let f (·) ∈ G, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. If the
following condition holds
T
A + A + (kB ∗ k∞ + kB ∗ k1 + kC ∗ k∞ + kC ∗ k1 )I ≤ 0,
∗
(b∗ji )n×n , C ∗
(3.4)
(c∗ji )n×n ,
where B =
=
then, the neural network defined by (2.2) is a
robust dissipative system and the set S2 = {x||xi (t)| ≤ |ui |/di , i = 1, 2, . . . , n} is a
positive invariant and globally attractive set.
Proof. Consider radially unbounded Lyapunov functional
n Z xi (t)
n X
n Z t
X
X
V (x(t)) = 2
fi (s)ds +
b∗ji fi2 (xi (s))ds
+
i=1 0
n X
n
X
i=1 j=1
i=1 j=1
c∗ji
Z
∞
Z
Kji (s)
0
t
t−τ
fi2 (xi (ξ))dξ ds.
t−s
Calculating the upper right derivative D+ V of the Lyapunov functional V along
the solution of (2.1), we have that
n
n X
n
X
X
D+ V (x(t)) = −2
di fi (xi (t))xi (t) + 2
aij fi (xi (t))fj (xj (t))
+2
+
i=1 j=1
n X
n
X
i=1 j=1
bij fi (xi (t))fj (xj (t − τ )) + 2
n
X
fi (xi (t))ui
i=1
Z
t
Kij (t − s)fj (xj (s))ds
−∞
i=1 j=1
n X
n
n X
n
X
X
∗ 2
bji fi (xi (t)) −
b∗ji fi2 (xi (t − τ ))
i=1 j=1
i=1 j=1
Z
n
n
∞
XX
c∗ji
Kji (s)[fi2 (xi (t)) − fi2 (xi (t − s))]ds
0
i=1 j=1
n
n X
n
X
X
+2
+
i=1
n X
n
X
≤ −2
+2
cij fi (xi (t))
di |fi (xi (t))||xi (t)|
i=1
n X
n
X
+2
aij fi (xi (t))fj (xj (t))
i=1 j=1
|bij ||fi (xi (t))||fj (xj (t − τ ))| + 2
i=1 j=1
n
X
i=1
|fi (xi (t))||ui |
6
B. CUI, X. LOU
+
Z
∞
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
0
i=1 j=1
n X
n
n X
n
X
X
b∗ji fi2 (xi (t)) −
b∗ji fi2 (xi (t − τ ))
i=1 j=1
i=1 j=1
Z
n
n
∞
XX
c∗ji
Kji (s)[fi2 (xi (t)) − fi2 (xi (t − s))]ds.
0
i=1 j=1
+2
+
n X
n
X
EJDE-2007/89
|cij |
Since
2|fi (xi (t))||fj (xj (t − τ ))| ≤ fi2 (xi (t)) + fj2 (xj (t − τ )),
Z ∞
2
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
0
Z ∞
Z ∞
2
≤
Kij (s)fi (xi (t))ds +
Kij (s)fj2 (xj (t − s))ds.
0
(3.5)
(3.6)
0
Substitute (3.5) and (3.6) into the inequality above these two, we obtain
D+ V (x(t))
aij fi (xi (t))fj (xj (t))
di |fi (xi (t))||xi (t)| + 2
i=1 j=1
i=1
n
n
n X
X
X
b∗ij fi2 (xi (t)) + 2
|fi (xi (t))||ui |
i=1
i=1 j=1
Z ∞
n X
n
X
∗
cij
Kij (s)fi2 (xi (t))ds
0
i=1 j=1
Z ∞
n
n X
n
n X
X
X
c∗ji
Kji (s)fi2 (xi (t))ds
b∗ji fi2 (xi (t)) +
0
i=1 j=1
i=1 j=1
≤ −2
+
+
+
n X
n
X
n
X
≤ −2
n
X
di |fi (xi (t))||xi (t)| + 2
i=1
n
X
|fi (xi (t))||ui |
i=1
T
+ f T (x(t)) A + A + (kB ∗ k∞ + kB ∗ k1 + kC ∗ k∞ + kC ∗ k1 )I f (x(t))
≤ −2
n
X
di |fi (xi (t))||xi (t)| + 2
i=1
n
n
X
|fi (xi (t))||ui | < 0,
i=1
when xi ∈ R \ S2 . This implies that the set S2 is a positive invariant and globally
attractive set.
Theorem 3.5. Let f (·) ∈ L, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. If the
following condition holds
T
T
A + A + BB + (1 + kC ∗ k∞ + kC ∗ k1 )I ≤ 0,
∗
(b∗ji )n×n , C ∗
(c∗ji )n×n ,
(3.7)
where B =
=
then, the neural network defined by (2.2) is a
robust dissipative system and the set
S3 = {x||fi (xi (t))| ≤ Li |ui |/di , i = 1, 2, . . . , n}
is a positive invariant and globally attractive set.
EJDE-2007/89
GLOBAL ROBUST DISSIPATIVITY
7
Proof. Let us use the following positive definite and radially unbounded Lyapunov
functional
n Z xi (t)
n Z t
X
X
V (x(t)) = 2
fi (s)ds +
fi2 (xi (s))ds
+
i=1 0
n X
n
X
i=1
c∗ji
Z
∞
t
fi2 (xi (ξ))dξ
Kji (s)
ds.
t−s
0
i=1 j=1
t−τ
Z
Calculating the upper right derivative D+ V of the Lyapunov functional V along
the solution of (2.1), we have that
D+ V (x(t)) = −2
+2
+
di fi (xi (t))xi (t) + 2
i=1
n X
n
X
i=1 j=1
n X
n
X
n X
n
X
aij fi (xi (t))fj (xj (t))
i=1 j=1
bij fi (xi (t))fj (xj (t − τ )) + 2
n
X
fi (xi (t))ui
i=1
Z
t
Kij (t − s)fj (xj (s))ds
−∞
i=1 j=1
n
n
X
X
fi2 (xi (t)) −
fi2 (xi (t − τ ))
j=1
j=1
Z ∞
n X
n
X
c∗ji
Kji (s)[fi2 (xi (t)) − fi2 (xi (t − s))]ds
0
i=1 j=1
+2
+
n
X
−2
cij fi (xi (t))
(3.8)
n
n
X
X
di 2
fi (xi (t)) + 2
|fi (xi (t))||ui |
L
i=1
i=1 i
× f T (x(t))(A + AT )f (x(t)) + 2f T (x(t − τ ))B T f (x(t))
Z ∞
n
n X
X
∗
cij
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
+2
0
i=1 j=1
T
+ f (x(t))f (x(t)) − f T (x(t − τ ))f (x(t − τ ))
Z ∞
n X
n
X
+
c∗ji
Kji (s)[fi2 (xi (t)) − fi2 (xi (t − s))]ds.
i=1 j=1
0
It is easy to see that
− f T (x(t − τ ))f (x(t − τ )) + 2f T (x(t − τ ))B T f (x(t))
= −[f (x(t − τ )) − B T f (x(t))]T [f (x(t − τ )) − B T f (x(t))] + f T (x(t))BB T f (x(t)),
from which it follows that
−f T (x(t − τ ))f (x(t − τ )) + 2f T (x(t − τ ))B T f (x(t)) ≤ f T (x(t))BB T f (x(t)).
Also we have
∞
Z
2
0
Z
≤
0
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
Z ∞
∞
2
Kij (s)fi (xi (t))ds +
Kij (s)fj2 (xj (t − s))ds.
0
8
B. CUI, X. LOU
EJDE-2007/89
Using the above inequality in (3.8) results in:
D+ V (x(t)) ≤ −2
n
n
X
X
di 2
fi (xi (t)) + 2
|fi (xi (t))||ui |
L
i=1 i
i=1
+ f T (x(t))(A + AT + BB T + I)f (x(t))
Z ∞
n X
n
X
+
c∗ij
Kij (s)fi2 (xi (t))ds
+
i=1 j=1
n X
n
X
0
c∗ji
i=1 j=1
n
X
di
≤ −2
i=1
Li
∞
Z
Kji (s)fi2 (xi (t))ds
0
fi2 (xi (t)) + 2
n
X
|fi (xi (t))||ui |
i=1
T
T
+ f T (x(t)) A + A + BB + (1 + kC ∗ k∞ + kC ∗ k1 ) I f (x(t))
≤ −2
n
n
X
X
di 2
fi (xi (t)) + 2
|fi (xi (t))||ui | < 0,
L
i=1 i
i=1
when xi ∈ Rn \ S3 . This implies that the set S3 is a positive invariant and globally
attractive set.
Theorem 3.6. Let f (·) ∈ L, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. If the matrix
Q given by
Q = P (A − L−1 D) + (A − L−1 D)T P + P BB T P + (1 + kP C ∗ k∞ + kP C ∗ k1 )I (3.9)
is negative definite, then the neural network defined by (2.2) is a robust dissipative
system and the set
pi u i 2
pi u i 2
] ≤
, i = 1, 2, . . . , n
S4 = x : [fi (xi (t)) +
λM (Q)
λM (Q)
is a positive invariant and globally attractive set, where L = diag(L1 , L2 , . . . , Ln ),
P = diag(p1 , p2 , . . . , pn ) and λM (Q) is the maximum eigenvalue of the matrix Q.
Proof. Let us use the following positive definite and radially unbounded Lyapunov
functional
Z xi (t)
n
n Z t
X
X
V (x(t)) = 2
pi
fi (s)ds +
fi2 (xi (s))ds
+
0
i=1
n
n
XX
pi c∗ji
i=1 j=1
t−τ
i=1
Z
∞
Z
Kji (s)
0
t
fi2 (xi (ξ))dξ ds.
t−s
Calculating the upper right derivative D+ V of the Lyapunov functional V along
the solution of (2.1), it follows
D+ V (x(t)) = −2
+2
n
X
pi di fi (xi (t))xi (t) + 2
i=1
n X
n
X
i=1 j=1
n X
n
X
pi aij fi (xi (t))fj (xj (t))
i=1 j=1
pi bij fi (xi (t))fj (xj (t − τ )) + 2
n
X
i=1
pi fi (xi (t))ui
EJDE-2007/89
GLOBAL ROBUST DISSIPATIVITY
+
Z
t
Kij (t − s)fj (xj (s))ds
−∞
i=1 j=1
n
n
X
X
fi2 (xi (t)) −
fi2 (xi (t − τ ))
i=1
i=1
Z ∞
n X
n
X
pi c∗ji
Kji (s) fi2 (xi (t)) − fi2 (xi (t − s)) ds
0
i=1 j=1
+2
+
n X
n
X
9
≤ −2
n
X
pi
i=1
T
pi cij fi (xi (t))
n
X
di 2
fi (xi (t)) + 2
pi fi (xi (t))ui
Li
i=1
+ f (x(t))(P A + AT P )f (x(t)) + 2f T (x(t − τ ))B T P f (x(t))
Z ∞
n X
n
X
∗
+2
pi cij
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
0
i=1 j=1
T
+ f (x(t))f (x(t)) − f T (x(t − τ ))f (x(t − τ ))
Z ∞
n X
n
X
∗
+
pi cji
Kji (s)[fi2 (xi (t)) − fi2 (xi (t − s))]ds.
(3.10)
0
i=1 j=1
We have
− f T (x(t − τ ))f (x(t − τ )) + 2f T (x(t − τ ))B T P f (x(t))
T = − f (x(t − τ )) − B T P f (x(t))
f (x(t − τ )) − B T P f (x(t))
+ f T (x(t))P BB T P f (x(t)),
from which it follows that
−f T (x(t − τ ))f (x(t − τ )) + 2f T (x(t − τ ))B T P f (x(t)) ≤ f T (x(t))P BB T P f (x(t)).
Also we have
∞
Z
Kij (s)|fi (xi (t))||fj (xj (t − s))|ds
Z ∞
∞
2
Kij (s)fi (xi (t))ds +
Kij (s)fj2 (xj (t − s))ds.
2
0
Z
≤
0
0
Using the above inequality in (3.10), it results in
D+ V (x(t)) ≤ 2
n
X
T
pi fi (xi (t))ui + f T (x(t))P BB P f (x(t)) + f T (x(t))f (x(t))
i=1
+ f T (x(t))(P (A − L−1 D) + (A − L−1 D)T P )f (x(t))
Z ∞
n X
n
X
∗
+
|pi cij |
Kij (s)fi2 (xi (t))ds
+
i=1 j=1
n X
n
X
i=1 j=1
=2
n
X
i=1
0
|pi c∗ji |
Z
∞
Kji (s)fi2 (xi (t))ds
0
pi fi (xi (t))ui + f T (x(t))Qf (x(t))
10
B. CUI, X. LOU
≤2
n
X
pi fi (xi (t))ui +
i=1
n
X
EJDE-2007/89
λM (Q)fi2 (xi (t))
i=1
= λM (Q)
n
X
fi (xi (t)) +
i=1
pi u i 2 pi ui 2
−
<0
λM (Q)
λM (Q)
n
when xi ∈ R \ S4 . This imiplies that the set S4 is a positive invariant and globally
attractive set.
Corollary 3.7. Let f (·) ∈ L, f (0) = 0 and fi (xi ) → ∞ as |xi | → ∞. If the matrix
Q = (A − L−1 D) + (A − L−1 D)T + BB T + (1 + kC ∗ k∞ + kC ∗ k1 )I
(3.11)
is negative definite, then the neural network defined by (2.2) is a robust dissipative
system and the set
ui 2
u i 2
S5 = x : fi (xi (t)) +
≤
, i = 1, 2, . . . , n
λM (Q)
λM (Q)
is a positive invariant and globally attractive set, where L = diag(L1 , L2 , . . . , Ln ),
and λM (Q) is the maximum eigenvalue of the matrix Q.
Remark 3.8. In this paper, we derive several different conditions to check the
global robust dissipativity of integro-differential systems modeling neural networks
and the conditions in Theorems 3.1–3.6 are independent of each other.
Remark 3.9. For system (2.1) or (2.2), when the parameters are certain, the global
robust dissipativity of the system has been studied in [13]. However, parameter
fluctuation in neural network implementation on very large scale integration (VLSI)
chips is unavoidable and it is important to ensure that system be dissipative with
respect to these uncertainties in the design and applications of neural networks.
Therefore, their conclusions can be included in our results as special cases.
Remark 3.10. If delay kernel functions kij (t) are of the form
kij (t) = δ(t − τij ),
i, j = 1, 2, . . . , n,
(3.12)
then system (2.1) reduces to a system with time-varying delays which has been
lucubrated in [1, 8]. And many crucial results for dynamics of the system have
been obtained. Therefore the discrete delays can be included in our models by
choosing suitable kernel functions.
Remark 3.11. When cij = 0 (i, j = 1, 2, . . . , n), the system (2.1) reduces to system
(2.3) which was studied in [1, 8]. In this case, Theorems 3.1–3.6 turn out to be
generalized results for those in [1, 8]. Moreover, the results are less conservative
and more extensive than those in [1], which will be illustrated in Example 4.2.
4. Examples
In this section, we give illustrative examples for our results.
Example 4.1. Consider the system (2.1) or (2.2) with constant delays: τ = 2 for
i, j = 1, 2, with the initial values of the system as follows:
φ(s) = 0.3,
t ∈ [−2, 0).
EJDE-2007/89
GLOBAL ROBUST DISSIPATIVITY
11
Let f (·) ∈ G, and all the kernel properties given in (i)–(iv) are satisfied.
0.6 0
1 0
−2 − 31
D=
, D=
, A= 1
,
0 1.5
0 2
−3
5
5
1
−4
0.25 −0.5
0.5 0.25
4
A= 2
, B=
, B=
,
−0.2 −0.7
0 −0.5
− 73
5
u1 = −3,
C = C = 0,
Hence
u2 = 2.
−a11 − b∗11 − c∗11 − (a∗21 + b∗21 + c∗21 ) = 0.15 > 0,
−a22 − b∗22 − c∗22 − (a∗12 + b∗12 + c∗12 ) = 0.80 > 0,
(4.1)
Then the conditions of Corollary 3.3 are satisfied, therefore the model (2.1) is a
globally robust dissipative system, and the set
4
S1 = (x1 (t), x2 (t))||x1 (t)| ≤ 5, |x2 (t)| ≤
3
is a positive invariant and globally attractive set. However, since the parameters
are uncertain, the results in [13] are not applicable for this example.
Example 4.2. Consider the system (2.1) or (2.2) with constant delays: τ = 1 for
i, j = 1, 2, with the initial values of the system as follows:
φ(s) = 0.5,
t ∈ [−1, 0).
Let f (·) ∈ G, and all the kernel properties given in (i-iv) are satisfied. And here we
let
0.6 0
1 0
−2 − 31
D=
, D=
, A= 1
,
0 1.5
0 2
−3
5
7
1
−
0.25 −0.25
0.25 0.25
4
A = 14
=
,
B
=
,
,
B
−0.25 −0.25
0.25 0.25
− 54
4
0.25 −0.25
0.25 0.25
C=
, C=
, u1 = 1.5, u2 = −4.
−0.25 −0.25
0.25 0.25
Hence
−1.5 0.5
A + A + (kB k∞ + kB k1 + kC k∞ + kC k1 )I =
≤ 0,
0.5 −0.5
T
∗
∗
∗
∗
(4.2)
Then, the conditions of Theorem 3.4 are satisfied, therefore the model (2.1) is a
globally robust dissipative system, and the set
8
S1 = (x1 (t), x2 (t))||x1 (t)| ≤ 2.5, |x2 (t)| ≤
3
is a positive invariant and globally attractive set. However, when cij = 0 (i, j =
1, 2, . . . , n), since
3
T
5
A+A
−4
4
+I = 5
Q11 =
,
− 41
2
4
1
B∗
− 81
Q12 =
+ I = 81
,
− 8 − 81
2
−1 0
Q22 = −I =
,
0 −1
12
B. CUI, X. LOU
EJDE-2007/89
so we get that
Q11 Q12
QT12 Q22
is not negative semidefinite. So the condition of Theorem 3.5 in [1] is not satisfied,
one can not determine the dissipativity of the neural network. Therefore, our
obtained criteria for the global dissipativity of integro-differential neural networks
with time delay are new.
Q=
Conclusion. The global robust dissipativity problem of integro-differential neural
networks with time delay is discussed. Several results are presented to characterize
the global dissipation together with their sets of attraction, which might have an
impact in the studying the uniqueness of equilibria, the global asymptotic stability,
the instability and the existence of periodic solutions. Our results extend the earlier
works. Two examples are given to illustrate the correctness of our results.
References
[1] S. Arik; On the global dissipativity of dynamical neural networks with time delays, Physics
Letters A, 326 (2004) 126-132.
[2] J. D. Cao, J. Wang; Global asymptotic and robust stability of recurrent. neural networks with
time delays, IEEE Trans. Circuits Syst. I, Fundam. Theory Appl, 52 (2) (2005) 417-426.
[3] J. Cao, D. Huang, Y. Qu; Global Robust Stability of Delayed Recurrent Neural. Networks,
Chaos, Solitons and Fractals 23 (2005) 221-229.
[4] J. D. Cao; Global exponential stability of Hopfield neural networks, Int. J. Syst. Sci. 32 (2)
(2001) 233-236.
[5] J. D. Cao; Global stability conditions for delayed CNNs, IEEE Trans. Circuits Syst. I 48 (11)
(2001) 1330-1333.
[6] J. D. Cao, K. Yuan, Daniel W. C. Ho, J. Lam; Global point dissipativity of neural networks
with mixed time-varying delays, Chaos, 16 (2006) 013105-1-013105-9.
[7] A. P. Chen, J. D. Cao, L. H. Huang; Periodic solution and global exponential stability for
shunting inhibitory delayed cellular neural networks, Electron. J. Diff. Eqns., 2004 (29) (2004)
1-16.
[8] X. Liao, J. Wang; Global dissipativity of continuous-time recurrent neural networks with time
delay, Phys. Rev. E, 68 (2003) 1-7.
[9] X. Y. Lou, B. T. Cui; Global asymptotic stability of delay BAM neural networks with impulses, Chaos, Solitons and Fractals, 29 (4) (2006) 1023-1031.
[10] X. Y. Lou, B. T. Cui; Stochastic exponential stability for Markovian jumping BAM neural
networks with time-varying delays, IEEE Trans. Systems, Man and Cybernetics, Part B, 37
(7) (2007) 713-719.
[11] X. Y. Lou, B. T. Cui; Delay-dependent stochastic stability of delayed Hopfield neural networks
with Markovian jump parameters, Journal of Mathematical Analysis and Applications, 328
(1) (2007) 316-326.
[12] S. Mohamad, K. Gopalsamy; Dynamics of a class of discrete-time neural networks and their
continuous-time counter parts, Math. Comput. Simulation, 53 (2000) 1-39.
[13] Q. K. Song, Z. J. Zhao; Global dissipativity of neural networks with both variable and unbounded delays, Chaos, Solitons and Fractals, 25 (2005) 393-401.
[14] H. Y. Zhao; Global exponential stability and periodicity of cellular neural networks with
variable delays, Physics Letters A 336 (4-5) (2005) 331-341.
[15] Y. Zheng, T. Chen; Global exponential stability of delayed periodic dynamical systems,
Physics Letters A, 322 (2004) 344-355.
College of Communication and Control Engineering, Jiangnan University, 1800 Lihu
Rd., Wuxi, Jiangsu 214122, China
E-mail address, B. Cui: btcui@vip.sohu.com
E-mail address, X.Lou: louxuyang28945@163.com
Download