Abstract and Applied Analysis

Abstract and Applied Analysis / 2009 / Article

Research Article | Open Access

Volume 2009 |Article ID 725846 | 16 pages | https://doi.org/10.1155/2009/725846

Stochastic Passivity of Uncertain Neural Networks with Time-Varying Delays

Academic Editor: Elena Litsyn
Received22 Jul 2009
Accepted18 Oct 2009
Published04 Jan 2010

Abstract

The passivity problem is investigated for a class of stochastic uncertain neural networks with time-varying delay as well as generalized activation functions. By constructing appropriate Lyapunov-Krasovskii functionals, and employing Newton-Leibniz formulation, the free-weighting matrix method, and stochastic analysis technique, a delay-dependent criterion for checking the passivity of the addressed neural networks is established in terms of linear matrix inequalities (LMIs), which can be checked numerically using the effective LMI toolbox in MATLAB. An example with simulation is given to show the effectiveness and less conservatism of the proposed criterion. It is noteworthy that the traditional assumptions on the differentiability of the time-varying delays and the boundedness of its derivative are removed.

1. Introduction

During the last two decades, many artificial neural networks have been extensively investigated and successfully applied to various areas such as signal processing, pattern recognition, associative memory, and optimization problems [1]. In such applications, it is of prime importance to ensure that the designed neural networks are stable [2].

In hardware implementation, time delays are likely to be present due to the finite switching speed of amplifiers and communication time. It has also been shown that the processing of moving images requires the introduction of delay in the signal transmitted through the networks [3]. The time delays are usually variable with time, which will affect the stability of designed neural networks and may lead to some complex dynamic behavior such as oscillation, bifurcation, or chaos [4]. Therefore, the study of stability with consideration of time delays becomes extremely important to manufacture high quality neural networks [5]. Many important results on stability of delayed neural networks have been reported, see [110] and the references therein for some recent publications.

It is also well known that parameter uncertainties, which are inherent features of many physical systems, are great sources of instability and poor performance [11]. These uncertainties may arise due to the variations in system parameters, modelling errors, or some ignored factors [12]. It is not possible to perfectly characterize the evolution of an uncertain dynamical system as a deterministic set of state equations [13]. Recently, the problem on robust stability analysis of uncertain neural networks with delays has been extensively investigated, see [1114] and the references therein for some recent publications.

Just as pointed out in [15], in real nervous systems, synaptic transmission is a noisy process brought on by random fluctuations from the release of neurotransmitters and other probabilistic causes. In the implementation of artificial neural networks, noise is unavoidable and should be taken into consideration in modelling. Therefore, it is of significant importance to consider stochastic effects to the dynamical behavior of neural networks [16]. Some recent interest results on stability of stochastic neural networks can be found, see [1526] and the references therein for some recent publications.

On the other hand, the passivity theory is another effective tool to the stability analysis of nonlinear system [27]. The main idea of passivity theory is that the passive properties of system can keep the system internal stability [27]. Thus, the passivity theory has received a lot of attention from the control community since 1970s [2831]. Recently, the passivity theory for delayed neural networks was investigated, some criteria checking the passivity were provided for certain or uncertain neural networks, see [3238] and references therein. In [32], the passivity-based approach is used to derive stability conditions for dynamic neural networks with different time scales. In [3336], authors investigated the passivity of neural networks with time-varying delay. In [37, 38], stochastic neural networks with time-varying delays were considered, several sufficient conditions checking the passivity were obtained. It is worth pointing out that, the given criteria in [3337] have been based on the following assumptions: () the time-varying delays are continuously differentiable; () the derivative of time-varying delay is bounded and is smaller than one; () the activation functions are bounded and monotonically nondecreasing. However, time delays can occur in an irregular fashion, and sometimes the time-varying delays are not differentiable. In such a case, the methods developed in [3338] may be difficult to be applied, and it is therefore necessary to further investigate the passivity problem of neural networks with time-varying delays under milder assumptions. To the best of our knowledge, few authors have considered the passivity problem for stochastic uncertain neural networks with time-varying delays as well as generalized activation functions.

Motivated by the above discussions, the objective of this paper is to study the passivity of stochastic uncertain neural networks with time-varying delays as well as generalized activation functions by employing a combination of Lyapunov functional, the free-weighting matrix method and stochastic analysis technique. The obtained sufficient conditions require neither the differentiability of time-varying delays nor the monotony of the activation functions, and are expressed in terms of linear matrix inequalities (LMIs), which can be checked numerically using the effective LMI toolbox in MATLAB. An example is given to show the effectiveness and less conservatism of the proposed criterion.

2. Problem Formulation and Preliminaries

In this paper, we consider the following stochastic uncertain neural networks with time-varying delay:

for , where is the state vector of the network at time , corresponds to the number of neurons; is a positive diagonal matrix, , and are known constant matrices; , and are time-varying parametric uncertainties; is the diffusion coefficient matrix and is an -dimensional Brownian motion defined on a complete probability space () with a filtration satisfying the usual conditions (i.e., it is right continuous and contains all -null sets); denotes the neuron activation at time ; is a varying external input vector; is the time-varying delay, and is assumed to satisfy , where is constant.

The initial condition associated with model (2.1) is given by

Let denote the state trajectory of model (2.1) from the above initial condition and the corresponding trajectory with zero initial condition.

Throughout this paper, we make the following assumptions.

(H1) [33] The time-varying uncertainties , and are of the form

where , , , , and are known constant matrices of appropriate dimensions, , and are known time-varying matrices with Lebesgue measurable elements bounded by

(H2) [10] For any , and there exist constants and such that

for all .

(H3) [15] There exist two scalars , such that the following inequality:

holds for all .

Definition 2.1 (see [33]). System (2.1) is called globally passive in the sense of expectation if there exists a scalar such that for all and for all , where stands for the mathematical expectation operator with respect to the given probability measure .

To prove our results, the following lemmas that can be found in [39] are necessary.

Lemma 2.2 (see [39]). For given matrices , and with and a scalar , the following holds:

Lemma 2.3 (see [39]). For any constant matrix , , scalar , vector function such that the integrations concerned are well defined, then

Lemma 2.4 (see [39]). Given constant matrices , and , where , , then is equivalent to the following conditions:

3. Main Results

For presentation convenience, in the following, we denote

Theorem 3.1. Under assumptions (H1)–(H3), model (2.1) is passive in the sense of expectation if there exist two scalars , , three symmetric positive definite matrices (), two positive diagonal matrices and , and matrices () such that the following two LMIs hold: where in which , , , , , , .

Proof. Let , , then model (2.1) is rewritten as Consider the following Lyapunov-Krasovskii functional as By It differential rule, the stochastic derivative of along the trajectory of model (3.5) can be obtained as From the definition of , we have By assumption (H1) and Lemma 2.2, we get It follows from (3.8) and (3.9) that Integrating both sides of (3.5) from to , we have Hence, By Lemmas 2.2 and 2.3, and noting , we get Integrating both sides of (3.5) from to , we have Similarly, by using of the same way, and noting , we get From assumption (H2), we have which are equivalent to where denotes the unit column vector having 1 element on its th row and zeros elsewhere. Let then that is Similarly, one has It follows from (3.7), (3.10), (3.13), (3.15), (3.20) and (3.21) that By assumption (H3) and inequality (3.2), we get From the proof of [19], we have Taking the mathematical expectation on both sides of (3.22), and noting (3.24), we get where , and with , , , , .
It is easy to verify the equivalence of and by using Lemma 2.4. Thus, one can derive from (3.3) and (3.25) that
From (3.27) and the definition of , we can get From Definition 2.1, we know that the stochastic neural networks (2.1) are globally passive in the sense of expectation, and the proof of Theorem 3.1 is then completed.

Remark 3.2. Assumption (H2) was first proposed in [10]. The constants and () in assumption (H2) are allowed to be positive, negative or zero. Hence, Assumption (H2) is weaker than the assumption in [2737]. In addition, the conditions in [3237] that the time-varying delay is differentiable and the derivative is bounded or smaller than one have been removed in this paper.

Remark 3.3. In [36], authors considered the passivity of uncertain neural networks with both discrete and distributed time-varying delays. In [37], authors considered the passivity for stochastic neural networks with time-varying delays and random abrupt changes. It is worth pointing out that, the method in this paper can also analyze the passivity for models in [36, 37].

Remark 3.4. It is known that the obtained criteria for checking passivity of neural networks depend on the constructed Lyapunov functionals or Lyapunov-Krasovskii functionals in varying degrees. Constructing proper Lyapunov functionals or Lyapunov-Krasovskii functionals can reduce conservatism. Recently, the delay fractioning approach has been used to investigate global synchronization of delayed complex networks with stochastic disturbances, which has shown the potential of reducing conservatism [22]. Using the delay fractioning approach, we can also investigate the passivity of delayed neural networks. The corresponding results will appear in the near future.

Remark 3.5. When we do not consider the stochastic effect, model (2.1) turns into the following model: Furthermore, model (3.29) also comprises the following neural network model with neither stochastic effect nor uncertainty which have been considered in [33, 34]. For models (3.29) and (3.30), one can get the following results.

Corollary 3.6. Under assumptions (H1)-(H2), model (3.29) is passive if there exist a scalar , three symmetric positive definite matrices (), two positive diagonal matrices and , and matrices () such that the following LMI holds: where in which , , , , , , .

Corollary 3.7. Under assumption (H2), model (3.30) is passive if there exist a scalar , three symmetric positive definite matrices (), two positive diagonal matrices and , and matrices () such that the following LMI holds: where , , , , .

4. An Example

Consider a two-neuron neural network (3.30), where

Figure 1 depicts the states of the considered network with initial conditions , , .

It can be verified that assumption (H2) is satisfied, and , , . By the Matlab LMI Control Toolbox, we find a solution to the LMI in (3.33) as follows:

Therefore, by Corollary 3.7, we know that model (3.30) is passive. It should be pointed out that the conditions in [3336] cannot be applied to this example since that require the differentiability of the time-varying delay.

5. Conclusions

In this paper, the passivity has been investigated for a class of stochastic uncertain neural networks with time-varying delay as well as generalized activation functions. By employing a combination of Lyapunov-Krasovskii functionals, the free-weighting matrix method, Newton-Leibniz formulation, and stochastic analysis technique, a delay-independent criterion for checking the passivity of the addressed neural networks has been established in terms of linear matrix inequalities (LMIs), which can be checked numerically using the effective LMI toolbox in MATLAB. The obtained results generalize and improve the earlier publications and remove the traditional assumptions on the differentiability of the time-varying delay and the boundedness of its derivative. An example has been provided to demonstrate the effectiveness and less conservatism of the proposed criterion.

We would like to point out that it is possible to generalize our main results to more complex neural networks, such as neural networks with discrete and distributed delays [10, 26], and neural networks of neutral-type [7, 20], neural networks with Markovian jumping [24, 25]. The corresponding results will appear in the near future.

Acknowledgments

The authors would like to thank the reviewers and the editor for their valuable suggestions and comments which have led to a much improved paper. This work was supported by the National Natural Science Foundation of China under Grants 60974132 and 50608072, and in part by Natural Science Foundation Project of CQ CSTC2008BA6038 and 2008BB2351, and Scientific & Technological Research Projects of CQ KJ090406, and the Ministry of Education for New Century Excellent Talent Support Program.

References

  1. Ju H. Park and O. M. Kwon, “Delay-dependent stability criterion for bidirectional associative memory neural networks with interval time-varying delays,” Modern Physics Letters B, vol. 23, no. 1, pp. 35–46, 2009. View at: Publisher Site | Google Scholar
  2. K. Gopalsamy, “Stability of artificial neural networks with impulses,” Applied Mathematics and Computation, vol. 154, no. 3, pp. 783–813, 2004. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  3. S. Arik, “An analysis of exponential stability of delayed neural networks with time varying delays,” Neural Networks, vol. 17, no. 7, pp. 1027–1031, 2004. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  4. W. Lu and T. Chen, “Dynamical behaviors of Cohen-Grossberg neural networks with discontinuous activation functions,” Neural Networks, vol. 18, no. 3, pp. 231–242, 2005. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  5. T. Chen and L. Wang, “Global μ-stability of delayed neural networks with unbounded time-varying delays,” IEEE Transactions on Neural Networks, vol. 18, no. 6, pp. 1836–1840, 2007. View at: Publisher Site | Google Scholar
  6. J. Cao and X. Li, “Stability in delayed Cohen-Grossberg neural networks: LMI optimization approach,” Physica D, vol. 212, no. 1-2, pp. 54–65, 2005. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  7. Ju H. Park and O. M. Kwon, “Further results on state estimation for neural networks of neutral-type with time-varying delay,” Applied Mathematics and Computation, vol. 208, no. 1, pp. 69–75, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  8. Ju H. Park, “Global exponential stability of cellular neural networks with variable delays,” Applied Mathematics and Computation, vol. 183, no. 2, pp. 1214–1219, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  9. Ju H. Park, “On global stability criterion for neural networks with discrete and distributed delays,” Chaos, Solitons & Fractals, vol. 30, no. 4, pp. 897–902, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  10. Y. Liu, Z. Wang, and X. Liu, “Global exponential stability of generalized recurrent neural networks with discrete and distributed delays,” Neural Networks, vol. 19, no. 5, pp. 667–675, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  11. N. Ozcan and S. Arik, “A new sufficient condition for global robust stability of bidirectional associative memory neural networks with multiple time delays,” Nonlinear Analysis: Real World Applications, vol. 10, no. 5, pp. 3312–3320, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  12. V. Singh, “Novel global robust stability criterion for neural networks with delay,” Chaos, Solitons & Fractals, vol. 41, no. 1, pp. 348–353, 2009. View at: Publisher Site | Google Scholar | MathSciNet
  13. Ju H. Park, “Further note on global exponential stability of uncertain cellular neural networks with variable delays,” Applied Mathematics and Computation, vol. 188, no. 1, pp. 850–854, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  14. O. M. Kwon and Ju H. Park, “Exponential stability analysis for uncertain neural networks with interval time-varying delays,” Applied Mathematics and Computation, vol. 212, no. 2, pp. 530–541, 2009. View at: Publisher Site | Google Scholar | MathSciNet
  15. Z. Wang, Y. Liu, K. Fraser, and X. Liu, “Stochastic stability of uncertain Hopfield neural networks with discrete and distributed delays,” Physics Letters A, vol. 354, no. 4, pp. 288–297, 2006. View at: Publisher Site | Google Scholar
  16. X. Wang, Q. Guo, and D. Xu, “Exponential p-stability of impulsive stochastic Cohen-Grossberg neural networks with mixed delays,” Mathematics and Computers in Simulation, vol. 79, no. 5, pp. 1698–1710, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  17. P. Balasubramaniam and R. Rakkiyappan, “Global asymptotic stability of stochastic recurrent neural networks with multiple discrete delays and unbounded distributed delays,” Applied Mathematics and Computation, vol. 204, no. 2, pp. 680–686, 2008. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  18. C. Huang, Y. He, and H. Wang, “Mean square exponential stability of stochastic recurrent neural networks with time-varying delays,” Computers & Mathematics with Applications, vol. 56, no. 7, pp. 1773–1778, 2008. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  19. J. Yu, K. Zhang, and S. Fei, “Further results on mean square exponential stability of uncertain stochastic delayed neural networks,” Communications in Nonlinear Science and Numerical Simulation, vol. 14, no. 4, pp. 1582–1589, 2009. View at: Publisher Site | Google Scholar | MathSciNet
  20. Ju H. Park and O. M. Kwon, “Synchronization of neural networks of neutral type with stochastic perturbation,” Modern Physics Letters B, vol. 23, no. 14, pp. 1743–1751, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  21. Ju H. Park and O. M. Kwon, “Analysis on global stability of stochastic neural networks of neutral type,” Modern Physics Letters B, vol. 22, no. 32, pp. 3159–3170, 2008. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  22. Y. Wang, Z. Wang, and J. Liang, “A delay fractioning approach to global synchronization of delayed complex networks with stochastic disturbances,” Physics Letters A, vol. 372, no. 39, pp. 6066–6073, 2008. View at: Publisher Site | Google Scholar
  23. Z. Wang, H. Shu, J. Fang, and X. Liu, “Robust stability for stochastic Hopfield neural networks with time delays,” Nonlinear Analysis: Real World Applications, vol. 7, no. 5, pp. 1119–1128, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  24. Z. Wang, Y. Liu, L. Yu, and X. Liu, “Exponential stability of delayed recurrent neural networks with Markovian jumping parameters,” Physics Letters A, vol. 356, no. 4-5, pp. 346–352, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  25. Y. Liu, Z. Wang, J. Liang, and X. Liu, “Stability and synchronization of discrete-time Markovian jumping neural networks with mixed mode-dependent time delays,” IEEE Transactions on Neural Networks, vol. 20, no. 7, pp. 1102–1116, 2009. View at: Publisher Site | Google Scholar
  26. Z. Wang, Y. Liu, and X. Liu, “State estimation for jumping recurrent neural networks with discrete and distributed delays,” Neural Networks, vol. 22, no. 1, pp. 41–48, 2009. View at: Publisher Site | Google Scholar
  27. W. Yu, “Passivity analysis for dynamic multilayer neuro identifier,” IEEE Transactions on Circuits and Systems, vol. 50, no. 1, pp. 173–178, 2003. View at: Publisher Site | Google Scholar | MathSciNet
  28. C. Li, H. Zhang, and X. Liao, “Passivity and passification of fuzzy systems with time delays,” Computers & Mathematics with Applications, vol. 52, no. 6-7, pp. 1067–1078, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  29. H. Gao, T. Chen, and T. Chai, “Passivity and passification for networked control systems,” SIAM Journal on Control and Optimization, vol. 46, no. 4, pp. 1299–1322, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  30. J. Zhao and D. J. Hill, “A notion of passivity for switched systems with state-dependent switching,” Journal of Control Theory and Applications, vol. 4, no. 1, pp. 70–75, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  31. W. Gui, B. Liu, and Z. Tang, “A delay-dependent passivity criterion of linear neutral delay systems,” Journal of Control Theory and Applications, vol. 4, no. 2, pp. 201–206, 2006. View at: Publisher Site | Google Scholar | MathSciNet
  32. W. Yu and X. Li, “Passivity analysis of dynamic neural networks with different time-scales,” Neural Processing Letters, vol. 25, no. 2, pp. 143–155, 2007. View at: Publisher Site | Google Scholar
  33. C. Li and X. Liao, “Passivity analysis of neural networks with time delay,” IEEE Transactions on Circuits and Systems II, vol. 52, no. 8, pp. 471–475, 2005. View at: Publisher Site | Google Scholar
  34. Ju H. Park, “Further results on passivity analysis of delayed cellular neural networks,” Chaos, Solitons & Fractals, vol. 34, no. 5, pp. 1546–1551, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  35. X. Lou and B. Cui, “Passivity analysis of integro-differential neural networks with time-varying delays,” Neurocomputing, vol. 70, no. 4–6, pp. 1071–1078, 2007. View at: Publisher Site | Google Scholar
  36. B. Chen, H. Li, C. Lin, and Q. Zhou, “Passivity analysis for uncertain neural networks with discrete and distributed time-varying delays,” Physics Letters A, vol. 373, no. 14, pp. 1242–1248, 2009. View at: Publisher Site | Google Scholar | MathSciNet
  37. X. Lou and B. Cui, “On passivity analysis of stochastic delayed neural networks with random abrupt changes,” New Mathematics and Natural Computation, vol. 3, no. 3, pp. 321–330, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  38. Q. Song, J. Liang, and Z. Wang, “Passivity analysis of discrete-time stochastic neural networks with time-varying delays,” Neurocomputing, vol. 72, no. 7–9, pp. 1782–1788, 2009. View at: Publisher Site | Google Scholar
  39. Q. Song and J. Cao, “Global dissipativity analysis on uncertain neural networks with mixed time-varying delays,” Chaos, vol. 18, no. 4, Article ID 043126, 10 pages, 2008. View at: Publisher Site | Google Scholar | MathSciNet

Copyright © 2009 Jianting Zhou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

606 Views | 482 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19. Sign up here as a reviewer to help fast-track new submissions.