Hindawi Publishing Corporation Journal of Applied Mathematics Volume 2012, Article ID 829594, 12 pages doi:10.1155/2012/829594 Research Article Robust Stochastic Stability Analysis for Uncertain Neutral-Type Delayed Neural Networks Driven by Wiener Process Weiwei Zhang1 and Linshan Wang2 1 2 College of Information Science and Engineering, Ocean University of China, Qingdao 266100, China Department of Mathematics, Ocean University of China, Qingdao 266100, China Correspondence should be addressed to Linshan Wang, wangls@ouc.edu.cn Received 9 July 2011; Revised 20 September 2011; Accepted 27 September 2011 Academic Editor: Shiping Lu Copyright q 2012 W. Zhang and L. Wang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The robust stochastic stability for a class of uncertain neutral-type delayed neural networks driven by Wiener process is investigated. By utilizing the Lyapunov-Krasovskii functional and inequality technique, some sufficient criteria are presented in terms of linear matrix inequality LMI to ensure the stability of the system. A numerical example is given to illustrate the applicability of the result. 1. Introduction In the past few years, neural networks and their various generalizations have drawn much research attention owing to their promising potential applications in a variety of areas, such as robotics, aerospace, telecommunications, pattern recognition, image processing, associative memory, signal processing, and combinatorial optimization 1–3. In such applications, it is of prime importance to ensure the asymptotic stability of the designed neural networks. Because of this, the stability of neural networks has been deeply investigated in the literature 4–14. It is known that time delays and stochastic perturbations are commonly encountered in the implementation of neural networks, and may result in instability or oscillation. So it is essential to investigate the stability of delayed stochastic neural networks 15, 16. Moreover, uncertainties are unavoidable in practical implementation of neural networks due to modeling errors and parameter fluctuation, which also cause instability and poor performance 15, 17, 18. Therefore, it is significant to introduce such uncertainties into delayed stochastic neural networks. 2 Journal of Applied Mathematics On the other hand, because of the complicated dynamic properties of the neural cells in the real world, it is natural and important that systems will contain some information about the derivative of the past state. Practically, such phenomenon always appears in the study of automatic control, circuit analysis, chemical process simulation, and population dynamics, and so forth. Recently, there has been increasing interest in the study of delayed neural networks of neutral type, see 6–15, 18–24. In 6, 8, the authors developed the global asymptotic stability of neutral-type neural networks with delays by utilizing the Lyapunov stability theory and LMI technique. In 9, 10, the global exponential stability of neutral-type neural networks with distributed delays is studied. However, the stochastic perturbations were not taken into account in those delayed neural networks 6–10. In 23, 24, the authors discussed the robust stability for uncertain stochastic neural networks of neutral-type with time-varying delays. However, the distributed delays were not taken into account in the models. So far, there are only a few papers that not only deal with the stochastic stability analysis for delayed neural networks of neutral-type, but also consider the parameter uncertainties. To the best of our knowledge, there are very few results on the stochastic stability analysis for uncertain neutral-type neural networks with both discrete and distributed delays driven by Wiener process. This motivates the research in this paper. In this paper, a class of uncertain neutral-type delayed neural networks driven by Wiener process is considered. By constructing a suitable Lyapunov functional, some new stability criteria to guarantee the system to be stochastically asymptotically stable in the mean square are given, which are less conservative than some existing reports. The structure of the addressed system is more general than in the other papers. The criteria can be checked easily by the LMI control toolbox in MATLAB. Moreover, a numerical example is given to illustrate the effectiveness and improvement over some existing results. 2. Preliminaries Notations. A < 0 denotes that A is a negative definite matrix. The superscript “T ” stands for the transpose of a matrix. Ω, F, P denotes a complete probability space, E· stands for the mathematical expectation operator. · stands for the Euclidean norm. I is the identity matrix of appropriate dimension, and the symmetric terms in a symmetric matrix are denoted by ∗. Consider the following class of uncertain neutral-type delayed neural networks driven by Wiener process: dxt − Cxt − ht −Atxt Btfxt − τt Dt H0 txt H1 txt − τtdwt, xt0 s ϕs, s ∈ t0 − ρ, t0 , t t−rt fxsds dt 2.1 where x x1 , x2 , . . . , xn T is the neuron state vector, At A ΔAt, Bt B ΔBt, Dt D ΔDt, H0 t H0 ΔH0 t, H1 t H1 ΔH1 t, A diagai n×n is a positive diagonal matrix, B, C, D ∈ Rn×n are the connection weight matrices, H0 , H1 ∈ Rn×n are known real constant matrices, ΔAt, ΔBt, ΔDt, ΔH0 t, ΔH1 t represent the time-varying parameter uncertain terms. fx f1 x1 , f2 x2 , . . . , fn xn T is the neuron Journal of Applied Mathematics 3 activation function with f0 0. wt w1 t, w2 t, . . . , wn tT is an n-dimensional Wiener process defined on a complete probability space Ω, F, P . rt, τt, ht are nonnegative, bounded, and differentiable time varying delays satisfying 0 < rt ≤ r < ∞, 0 < τt ≤ τ < ∞, τ̇t ≤ η1 < ∞, 0 < ht ≤ h < ∞, ḣt ≤ η2 < ∞. 2.2 The admissible parameter uncertain terms are assumed to be the following form: ΔAt, ΔBt, ΔDt, ΔH0 t, ΔH1 t UFtM1 , M2 , M3 , M4 , M5 , 2.3 where U, Mi , i 1, . . . , 5 are known real constant matrices, Ft is the time-varying uncertain matrix satisfying FT F ≤ I. 2.4 Suppose that f· is bounded and satisfies the following condition: fx ≤ Gx, 2.5 where G ∈ Rn×n is a known constant matrix. Assume that the initial value ϕ : −ρ, 0 → Rn is F0 -measurable and continuously differentiable, we introduce the following norm: ⎧ ⎫ ⎨ 2 ⎬ 2 2 ϕ max sup Eϕi s , sup Eϕ s < ∞, i ρ ⎩−α≤s≤0 ⎭ 2.6 −h≤s≤0 where ρ max{τ, h, r}, α max{τ, r}. Under the above assumptions, it is easy to verify that there exists a unique equilibrium point of system 2.1 see 25. Definition 2.1. The equilibrium point of 2.1 is said to be globally robustly stochastically asymptotically stable in the mean square, if the following condition holds: 2 lim Ex t, t0 , ϕ 0, t → ∞ t ≥ t0 , 2.7 where xt, t0 , ϕ is any solution of model 2.1 with initial value ϕ. Lemma 2.2 Schur complement 26. Given constant matrices Ω1 , Ω2 , Ω3 with appropriate dimensions, where ΩT1 Ω1 and ΩT2 Ω2 > 0, then Ω1 ΩT3 Ω−1 2 Ω3 < 0, 2.8 4 Journal of Applied Mathematics if and only if Ω1 ΩT3 ∗ < 0, −Ω2 −Ω2 Ω3 or ∗ < 0. Ω1 2.9 Lemma 2.3 see 26. Given matrices D, E, and F with FT F ≤ I and a scalar ε > 0, then DFE ET FT DT ≤ εDDT ε−1 ET E. 2.10 Lemma 2.4 see 27. For any constant matrix M ∈ Rn×n , M MT > 0, a scalar γ > 0, vector function xt : 0, γ → Rn such that the integrations are well defined, then γ T xsds 0 γ M xsds ≤ γ 0 γ xT sMxsds. 2.11 0 3. Main Results Theorem 3.1. System 2.1 is globally robustly stochastically asymptotically stable in the mean square, if there exist symmetric positive definite matrices P, Q, R, S, U1 , U2 and positive scalars δ, ε1 , ε2 > 0 such that LMI holds: ⎛ T T Γ1 A PC PB − ε1 MT1 M2 PD − ε1 MT1 M3 ε2 MT4 M5 H0 P ⎜ ⎜∗ ⎜ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ Λ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎝ ∗ −PU Γ2 −CT PB −CT PD 0 0 −CT PU ∗ Γ3 ε1 MT2 M3 0 0 0 ∗ ∗ Γ4 0 0 0 0 ∗ ∗ ∗ Γ5 T H1 P ∗ ∗ ∗ ∗ −P 0 ∗ ∗ ∗ ∗ ∗ −ε1 I ∗ ∗ ∗ ∗ ∗ ∗ T 0 ⎞ ⎟ 0 ⎟ ⎟ ⎟ ⎟ 0 ⎟ ⎟ ⎟ 0 ⎟ ⎟ < 0, 3.1 ⎟ 0 ⎟ ⎟ ⎟ PU ⎟ ⎟ ⎟ 0 ⎟ ⎠ −ε2 I where Γ1 −PA − A P Q R rGT SG ε1 MT1 M1 ε2 MT4 M4 , Γ2 −U1 − 1 − η2 R, Γ3 −δI ε1 MT2 M2 , Γ4 −r −1 S ε1 MT3 M3 , Γ5 −U2 − 1 − η1 Q δGT G ε2 MT5 M5 . Journal of Applied Mathematics 5 Proof. Using Lemma 2.2, the matrix Λ < 0 implies that ⎞ ⎛ T T Γ1 A PC PB − ε1 MT1 M2 PD − ε1 MT1 M3 ε2 MT4 M5 H0 P ⎟ ⎜ ⎟ ⎜ T T ⎜∗ Γ2 −C PB −C PD 0 0 ⎟ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ T ε1 M2 M3 0 0 ⎟ ∗ Γ3 ⎜∗ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜∗ 0 0 ∗ ∗ Γ 4 ⎟ ⎜ ⎟ ⎜ T ⎟ ⎜ ⎜∗ H1 P⎟ ∗ ∗ ∗ Γ5 ⎠ ⎝ ∗ ∗ ∗ ∗ ∗ −P ⎛ PU 0 ⎞ ⎜ ⎟ ⎜−CT PU 0 ⎟ ⎜ ⎟ ⎜ ⎟⎛ ⎞⎛ ⎞ ⎜ ⎟ T T ⎜ 0 ⎟ ε1−1 I 0 U P −U PC 0 0 0 0 0 ⎜ ⎟⎝ ⎠⎝ ⎠ ⎜ ⎟ ⎜ 0 ⎟ T −1 P 0 0 0 0 0 U 0 ∗ ε2 I ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ 0 0 ⎟ ⎝ ⎠ 0 PU ⎛ T Φ1 A PC ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜ ⎜∗ ⎝ ∗ ⎛ Γ2 ∗ ∗ ∗ ∗ MT4 T ⎞ ⎛ ⎞T ⎞⎛ 0 H0 P −MT1 −MT1 ⎟ ⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎟ ⎟⎜ −CT PB −CT PD 0 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎟ ⎜ ⎜ ⎟ ⎟ ⎟ ⎜ T ⎟⎜ T ⎟ ⎜ ⎜ ⎟ ⎟ M −δI 0 0 0 ⎟ M ⎟ 2 ⎟⎜ 2 ⎟ ⎟ ε1 ⎜ ⎜ ⎜ ⎟ ⎟ ⎟ ⎜ MT ⎟ ⎜ MT ⎟ ∗ −r −1 S 0 0 ⎟ ⎜ 3 ⎟⎜ 3 ⎟ ⎟ ⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎟ ⎟⎜ T ⎟ 0 ⎜ ⎜ ⎟ ⎟ 0 ∗ ∗ Φ2 H1 P⎟ ⎝ ⎠ ⎠⎝ ⎠ 0 0 ∗ ∗ ∗ −P ⎞⎛ PB MT4 ⎞T 3.2 PD ⎛ PU ⎞⎛ PU ⎞T ⎛ 0 ⎞⎛ 0 ⎞T ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜−CT PU⎟⎜−CT PU⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ −1 ⎜ −1 ⎜ ε2 ⎜ ⎟⎜ ⎟ ε1 ⎜ ⎟⎜ ⎟ ε2 ⎜ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ T ⎟⎜ T ⎟ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ M5 ⎟ ⎜ M5 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎝ ⎝ ⎝ ⎠⎝ ⎠ ⎠⎝ ⎠ ⎠⎝ ⎠ PU PU 0 0 0 0 < 0, T where Φ1 −PA − A P Q R rGT SG, Φ2 −U2 − 1 − η1 Q δGT G. 6 Journal of Applied Mathematics From 2.3, 2.4, using Lemma 2.3, we have ⎛ −PΔAt − ΔAT tP ΔAT tPC PΔBt PΔDt ⎜ ⎜ ∗ 0 −CT PΔBt −CT PΔDt ⎜ ⎜ ⎜ ∗ ∗ 0 0 ⎜ ⎜ ⎜ ∗ ∗ ∗ 0 ⎜ ⎜ ⎜ ∗ ∗ ∗ ∗ ⎝ ∗ ∗ ∗ ∗ ⎛ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎝ −MT1 ⎞ ⎛ PU ⎞T ⎛ ⎛ ⎞ PU ⎜ ⎟ ⎟ ⎜ T ⎟ ⎜ T ⎜ ⎜−C PU⎟ ⎜−C PU⎟ 0 ⎟ ⎜ ⎟ ⎟ ⎜ ⎟ ⎜ ⎜ ⎟ ⎟ ⎜ ⎟ ⎜ T ⎟ ⎜ ⎟ ⎟ ⎜ ⎜ M2 ⎟ T ⎜ 0 ⎟ ⎜ ⎜ 0 ⎟ Ft F t ⎜ ⎟ ⎟ ⎜ ⎟ ⎜ T ⎟ ⎜ ⎟ ⎟ ⎜ ⎜ M3 ⎟ ⎜ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ ⎟ ⎟ ⎜ ⎟ ⎜ ⎜ ⎜ 0 ⎟ ⎜ 0 ⎟ 0 ⎟ ⎝ ⎠ ⎠ ⎝ ⎠ ⎝ 0 0 0 ⎛ MT4 ⎞ ⎛ 0 ⎞T ⎛ 0 ⎞ ⎛ MT4 −MT1 ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ≤ ε1 ⎜ ⎜ ⎜ ⎜ ⎜ ⎝ −MT1 0 MT2 MT3 0 ⎛ MT4 0 0 ⎞⎛ MT4 ⎞T ⎛ 0 ⎞⎛ 0 0 0 0 0 ∗ ⎟ ⎟ ⎟ ⎟ ⎟ 0 ⎟ ⎟ ⎟ 0 ⎟ ⎟ T ΔH1 tP⎟ ⎠ 0 0 ⎞T ⎞T ⎞⎛ ⎞T ⎛ ⎞⎛ ⎞T PU −MT1 PU ⎟⎜ ⎟ ⎜ T ⎟⎜ ⎟ ⎜−C PU⎟⎜−CT PU⎟ ⎟⎜ 0 ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ T ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎟⎜ M ⎟ ⎜ ⎟ ⎟ ⎜ ⎟⎜ 2 ⎟ ⎟⎜ ⎟ ⎟⎜ T ⎟ ε1−1 ⎜ ⎜ 0 ⎟⎜ 0 ⎟ ⎟⎜ M ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ 3 ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎟⎜ 0 ⎟ ⎠⎝ ⎠ ⎝ ⎠⎝ ⎠ 0 0 ⎞ ⎟ 0 ⎟ ⎟ ⎟ T ⎟ M2 ⎟ ⎟ MT3 ⎟ ⎟ ⎟ 0 ⎟ ⎠ 0 ⎜ ⎜ ⎜ ⎟ ⎜ ⎟ ⎟ ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ ⎜ ⎜ ⎟ ⎜ ⎟ ⎟ ⎟ ⎜ ⎜ ⎜ ⎟ ⎜ ⎟ ⎟ ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ ⎜ ⎜ ⎟ ⎟ T ⎜ ⎟ ⎟ ⎜ ⎟F t⎜ ⎟ ⎜ ⎟Ft⎜ ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎜ ⎜ ⎜ ⎟ ⎜ ⎟ ⎟ ⎟ ⎜ ⎜ ⎜ ⎟ ⎜ ⎟ ⎟ ⎟ ⎜ MT ⎟ ⎜ MT ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎝ 5⎠ ⎝ 5⎠ ⎝ ⎠ ⎝ ⎠ PU PU 0 0 ⎛ 0 ΔHT0 tP ⎞T ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎟⎜ ⎟ ⎜ ⎟⎜ ⎟ −1 ⎜ ε2 ⎜ ⎟⎜ ⎟ ε2 ⎜ ⎟⎜ ⎟ . ⎜ 0 ⎟⎜ 0 ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ ⎜ ⎟⎜ ⎟ ⎟⎜ ⎟ ⎜ MT ⎟ ⎜ MT ⎟ ⎜ 0 ⎟⎜ 0 ⎟ ⎝ 5 ⎠⎝ 5 ⎠ ⎝ ⎠⎝ ⎠ PU PU 0 0 3.3 Journal of Applied Mathematics 7 Together with 3.2, we get ⎛ Ψ AT tPC ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎜ ⎜ ⎜∗ ⎝ Γ2 ∗ ∗ ∗ ∗ ⎞ 0 HT0 tP ⎟ −CT PBt −CT PDt 0 0 ⎟ ⎟ ⎟ −δI 0 0 0 ⎟ ⎟ ⎟ < 0, ∗ −r −1 S 0 0 ⎟ ⎟ ⎟ T ∗ ∗ Φ2 H1 tP⎟ ⎠ ∗ PBt ∗ PDt ∗ ∗ 3.4 −P where Ψ −PAt − AT tP Q R rGT SG. Utilizing Lemma 2.2 again, we obtain ⎛ ⎜ ⎜∗ ⎜ ⎜ Σ⎜ ⎜∗ ⎜ ⎜∗ ⎝ PBt Γ2 ∗ ∗ ∗ ⎛ ⎞ ⎞ ⎛ ⎞T HT0 t ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ −CT PBt −CT PDt 0 ⎟ ⎟ ⎜ 0 ⎟ ⎜ 0 ⎟ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ −δI 0 0⎟ ⎟ ⎜ 0 ⎟P⎜ 0 ⎟ < 0. ⎜ ⎜ ⎟ ⎟ ⎟ ⎜ ⎟ ⎜ ⎟ ∗ −r −1 S 0⎟ ⎠ ⎝ 0 ⎠ ⎝ 0 ⎠ HT1 t HT1 t ∗ ∗ Φ2 Ψ AT tPC ∗ PDt 0 HT0 t 3.5 Constructing a positive definite Lyapunov-Krasovskii functional as follows: V t, xt yT tPyt 0 t −rt T t t θ t t−τt xT sQxsds T t f xsSfxsdsdθ t−ht T t xT sRxsds xT s − hsU1 xs − hsds xT s − τsU2 xs − τsds, where yt xt − Cxt − ht, T > t is a constant. By Ito’s differential formula, we get dV t, xt ≤ T 2y tP −Atxt Btfxt − τt Dt t t−rt fxsds xT tQxt − 1 − τ̇txT t − τtQxt − τt xT tRxt − 1 − ḣt xT t − htRxt − ht rfT xtSfxt t − fT xsSfxsds t−rt − xT t − htU1 xt − ht − xT t − τtU2 xt − τt 3.6 8 Journal of Applied Mathematics H0 txt H1 txt − τtT PH0 txt H1 txt − τt dt 2yT tPH0 txt H1 txt − τtdwt 2xt − Cxt − htT ≤ × P −Atxt Btfxt − τt Dt t t−rt fxsds xT tQxt − 1 − η1 xT t − τtQxt − τt xT tRxt − 1 − η2 xT t − htRxt − ht rfT xtSfxt t − fT xsSfxsds t−rt − xT t − htU1 xt − ht − xT t − τtU2 xt − τt T H0 txt H1 txt − τt PH0 txt H1 txt − τt dt 2yT tPH0 txt H1 txt − τtdwt. 3.7 From 2.5, for a scalar δ > 0, we have −δ fT xt − τtfxt − τt − xT t − τtGT Gxt − τt ≥ 0. 3.8 Using Lemma 2.4, we have t T t−rt −1 fxsds t r S t−rt fxsds ≤ t t−rt fT xsSfxsds. 3.9 Together 3.8, 3.9 with dV t, xt, we obtain dV t, xt ≤ xT t −PAt − AT tP Q R rGT SG xt xT tAT tPCxt − ht xT t − htCT PAtxt xT tPBtfxt − τt fT xt − τtBT tPxt T x tPDt t t−rt fxsds t t−rt T fxsds DT tPxt xT t − ht −U1 − 1 − η2 R xt − ht − xT t − htCT PBtfxt − τt Journal of Applied Mathematics 9 − fT xt − τtBT tPCxt − ht − xT t − htCT PDt − t t−rt t t−rt fxsds T DT tPCxt − ht xT t − τt fxsds × −U2 − 1 − η1 Q xt − τt − t t−rt T fxsds t r S −1 t−rt fxsds H0 txt H1 txt − τtT PH0 txt H1 txt − τt dt 2yT tPH0 txt H1 txt − τtdwt. 3.10 That is, dV t, xt ≤ ξT tΣξtdt 2yT tPH0 txt H1 txt − τtdwt, where ξT t xT t, xT t − ht, fT xt − τt, matrix Σ is given in 3.5. Taking the mathematical expectation, we get !t t−rt 3.11 T fxsds , xT t − τt, and the # " % $ dV t, xt ≤ E ξT tΣξt ≤ λmax ΣExt2 . E dt 3.12 From 3.5, we know Σ < 0, that is, λmax Σ < 0. By Lyapunov-Krasovskii stability theorems, the system 2.1 is globally robustly asymptotically stable. The proof is completed. Remark 3.2. To the best of our knowledge, few authors have considered the stochastically asymptotic stability for uncertain neutral-type neural networks driven by Wiener process. We can find recent papers 18, 22–24. However, it is assumed in 18 that the system is a linear model and all delays are constants. In 22, it is assumed that the time-varying delays satisfying τ̇t ≤ ρτ < 1, ḣt ≤ ρh < 1, in this paper, we relax it to τ̇t ≤ ρτ < ∞, ḣt ≤ ρh < ∞. In 23, 24, the authors discussed the robust stability for uncertain stochastic neural networks of neutral-type with time-varying delays. However, the distributed delays were not taken into account in the models. Hence, our results in this paper have wider adaptive range. Remark 3.3. Suppose that C 0, Dt 0 i.e., without neutral-type and distributed delays, then the system 2.1 becomes the one investigated in 15. Remark 3.4. In 17, the authors studied the global stability for uncertain stochastic neural networks with time-varying delay by Lyapunov functional method and LMI technique. However, the neutral term and distributed delays were not taken into account in the models. Therefore, our developed results in this paper are more general than those reported in 17. 10 Journal of Applied Mathematics Remark 3.5. It should be noted that the condition is given as linear matrix inequalities LMIs, therefore, by using the MATLAB LMI Toolbox, it is straightforward to check the feasibility of the condition. 4. Numerical Example Consider the following uncertain neutral-type delayed neural networks: dxt − Cxt − ht − A UFtM1 xt B UFtM2 fxt − τt D UFtM3 t t−rt 4.1 fxsds dt UFtM4 xt UFtM5 xt − τtdwt, where n 2, fi xi sin xi , i 1, 2, η1 0.7, η2 0.5, 0 < rt ≤ r 3, FT tFt ≤ I. The constant matrices are A D 3 0 0 3 , 0.04 0.03 B , −0.02 0.05 0.2 M2 M3 M4 0 0.2 0.16 0.2 0 , C , 0.04 0.08 0 0.2 0.6 0 0.1 0.5 , U , M1 0 0.6 0.5 0.3 0 0.4 0 1 0 , M5 , G . 0.2 0 0.4 0 1 4.2 By using the MATLAB LMI Control Toolbox, we obtain the feasible solution as follows: δ 2.0876, ε1 5.0486, ε2 8.0446, P 6.8465 −0.7257 , −0.7257 6.6012 3.4984 −0.8143 S , −0.8143 1.9588 Q U1 9.9371 −0.2792 , −0.2792 10.4388 1.1111 0.3889 0.3889 1.1060 R U2 , −2.3936 5.4991 , 8.0104 −2.3936 2.4193 −0.6561 −0.6561 2.6270 . 4.3 That is the system 4.1 is globally robustly stochastically asymptotically stable in the mean square. 5. Conclusion In this paper, the stochastically asymptotic stability problem has been studied for a class of uncertain neutral-type delayed neural networks driven by Wiener process by utilizing the Journal of Applied Mathematics 11 Lyapunov-Krasovskii functional and linear matrix inequality LMI approach. A numerical example is given to illustrate the applicability of the result. Acknowledgment This paper was fully supported by the National Natural Science Foundation of China no. 10771199 and no. 10871117. References 1 A. Bouzerdoum and T. R. Pattison, “Neural network for quadratic optimization with bound constraints,” IEEE Transactions on Neural Networks, vol. 4, no. 2, pp. 293–304, 1993. 2 M. Forti and A. Tesi, “New conditions for global stability of neural networks with application to linear and quadratic programming problems,” IEEE Transactions on Circuits and Systems I, vol. 42, no. 7, pp. 354–366, 1995. 3 M. P. Kennedy and L. O. Chua, “Neural networks for nonlinear programming,” IEEE Transactions on Circuits and Systems, vol. 35, no. 5, pp. 554–562, 1988. 4 S. Xu, J. Lam, D. W. C. Ho, and Y. Zou, “Delay-dependent exponential stability for a class of neural networks with time delays,” Journal of Computational and Applied Mathematics, vol. 183, no. 1, pp. 16–28, 2005. 5 R. Zhang and L. Wang, “Global exponential robust stability of interval cellular neural networks with S-type distributed delays,” Mathematical and Computer Modelling, vol. 50, no. 3-4, pp. 380–385, 2009. 6 J. H. Park, O. M. Kwon, and S. M. Lee, “LMI optimization approach on stability for delayed neural networks of neutral-type,” Applied Mathematics and Computation, vol. 196, no. 1, pp. 236–244, 2008. 7 R. Samli and S. Arik, “New results for global stability of a class of neutral-type neural systems with time delays,” Applied Mathematics and Computation, vol. 210, no. 2, pp. 564–570, 2009. 8 R. Rakkiyappan and P. Balasubramaniam, “LMI conditions for global asymptotic stability results for neutral-type neural networks with distributed time delays,” Applied Mathematics and Computation, vol. 204, no. 1, pp. 317–324, 2008. 9 R. Rakkiyappan and P. Balasubramaniam, “New global exponential stability results for neutral type neural networks with distributed time delays,” Neurocomputing, vol. 71, no. 4–6, pp. 1039–1045, 2008. 10 L. Liu, Z. Han, and W. Li, “Global stability analysis of interval neural networks with discrete and distributed delays of neutral type,” Expert Systems with Applications, vol. 36, no. 3, pp. 7328–7331, 2009. 11 R. Samidurai, S. M. Anthoni, and K. Balachandran, “Global exponential stability of neutral-type impulsive neural networks with discrete and distributed delays,” Nonlinear Analysis: Hybrid Systems, vol. 4, no. 1, pp. 103–112, 2010. 12 R. Rakkiyappan, P. Balasubramaniam, and J. Cao, “Global exponential stability results for neutraltype impulsive neural networks,” Nonlinear Analysis: Real World Applications, vol. 11, no. 1, pp. 122– 130, 2010. 13 J. H. Park and O. M. Kwon, “Further results on state estimation for neural networks of neutral-type with time-varying delay,” Applied Mathematics and Computation, vol. 208, no. 1, pp. 69–75, 2009. 14 J. Qiu and J. Cao, “Delay-dependent robust stability of neutral-type neural networks with time delays,” Journal of Mathematical Control Science and Applications, vol. 1, pp. 179–188, 2007. 15 J. Zhang, P. Shi, and J. Qiu, “Novel robust stability criteria for uncertain stochastic Hopfield neural networks with time-varying delays,” Nonlinear Analysis: Real World Applications, vol. 8, no. 4, pp. 1349– 1357, 2007. 16 L. Wang, Z. Zhang, and Y. Wang, “Stochastic exponential stability of the delayed reaction-diffusion recurrent neural networks with Markovian jumping parameters,” Physics Letters A, vol. 372, no. 18, pp. 3201–3209, 2008. 17 Y. Wu, Y. Wu, and Y. Chen, “Mean square exponential stability of uncertain stochastic neural networks with time-varying delay,” Neurocomputing, vol. 72, no. 10–12, pp. 2379–2384, 2009. 18 M. H. Jiang, Y. Shen, and X. X. Liao, “Robust stability of uncertain neutral linear stochastic differential delay system,” Applied Mathematics and Mechanics, vol. 28, no. 6, pp. 741–748, 2007. 12 Journal of Applied Mathematics 19 H. Zhang, M. Dong, Y. Wang, and N. Sun, “Stochastic stability analysis of neutral-type impulsive neural networks with mixed time-varying delays and Markovian jumping,” Neurocomputing, vol. 73, no. 13–15, pp. 2689–2695, 2010. 20 Q. Zhu and J. Cao, “Stability analysis for stochastic neural networks of neutral type with both Markovian jump parameters and mixed time delays,” Neurocomputing, vol. 73, no. 13–15, pp. 2671– 2680, 2010. 21 H. Bao and J. Cao, “Stochastic global exponential stability for neutral-type impulsive neural networks with mixed time-delays and Markovian jumping parameters,” Communications in Nonlinear Science and Numerical Simulation, vol. 16, no. 9, pp. 3786–3791, 2011. 22 X. Li, “Global robust stability for stochastic interval neural networks with continuously distributed delays of neutral type,” Applied Mathematics and Computation, vol. 215, no. 12, pp. 4370–4384, 2010. 23 H. Chen, Y. Zhang, and P. Hu, “Novel delay-dependent robust stability criteria for neutral stochastic delayed neural networks,” Neurocomputing, vol. 73, no. 13–15, pp. 2554–2561, 2010. 24 G. Liu, S. X. Yang, Y. Chai, W. Feng, and W. Fu, “Robust stability criteria for uncertain stochastic neural networks of neutral-type with interval time-varying delays,” Neural Computing and Applications. In press. 25 X. Mao, Stochastic Differential Equations and Their Applications, Horwood Publishing Series in Mathematics & Applications, Horwood Publishing, Chichester, UK, 1997. 26 Y. Y. Wang, L. Xie, and C. E. de Souza, “Robust control of a class of uncertain nonlinear systems,” Systems & Control Letters, vol. 19, no. 2, pp. 139–149, 1992. 27 K. Gu, “An integral inequality in the stability problem of time-delay systems,” in Proceedings of the 39th IEEE Confernce on Decision and Control, pp. 2805–2810, Sydney, Australia, December 2000. Advances in Operations Research Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Advances in Decision Sciences Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Mathematical Problems in Engineering Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Algebra Hindawi Publishing Corporation http://www.hindawi.com Probability and Statistics Volume 2014 The Scientific World Journal Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Differential Equations Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Volume 2014 Submit your manuscripts at http://www.hindawi.com International Journal of Advances in Combinatorics Hindawi Publishing Corporation http://www.hindawi.com Mathematical Physics Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Complex Analysis Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Mathematics and Mathematical Sciences Journal of Hindawi Publishing Corporation http://www.hindawi.com Stochastic Analysis Abstract and Applied Analysis Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com International Journal of Mathematics Volume 2014 Volume 2014 Discrete Dynamics in Nature and Society Volume 2014 Volume 2014 Journal of Journal of Discrete Mathematics Journal of Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Applied Mathematics Journal of Function Spaces Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Optimization Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014