- About this Journal ·
- Abstracting and Indexing ·
- Advance Access ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Journal of Computational Methods in Physics

Volume 2013 (2013), Article ID 172906, 17 pages

http://dx.doi.org/10.1155/2013/172906

## Passivity Analysis of Markovian Jumping Neural Networks with Leakage Time-Varying Delays

^{1}Department of Mathematics, Kovai Kalaimagal College of Arts and Science, Coimbatore, Tamil Nadu 641 109, India^{2}Department of Mathematics, Avinashilingam Deemed University for Women, Coimbatore, Tamil Nadu 641 043, India

Received 30 March 2013; Accepted 17 June 2013

Academic Editor: Ali Cemal Benim

Copyright © 2013 N. Mala and A. R. Sudamani Ramaswamy. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

This paper is concerned with the passivity analysis of Markovian jumping neural networks with leakage time-varying delays. Based on a Lyapunov functional that accounts for the mixed time delays, a leakage delay-dependent passivity conditions are derived in terms of linear matrix inequalities (LMIs). The mixed delays includes leakage time-varying delays, discrete time-varying delays, and distributed time-varying delays. By employing a novel Lyapunov-Krasovskii functional having triple-integral terms, new passivity leakage delay-dependent criteria are established to guarantee the passivity performance. This performance not only depends on the upper bound of the time-varying leakage delay but also depends on the upper bound of the derivative of the time-varying leakage delay . While estimating the upper bound of derivative of the Lyapunov-Krasovskii functional, the discrete and distributed delays should be treated so as to appropriately develop less conservative results. Two numerical examples are given to show the validity and potential of the developed criteria.

#### 1. Introduction

In the past few decades, neural networks (NNs) have been a hot research topic because of their emerged application in static image processing, pattern recognition, fixed-point computation, associative memory, combinatorial optimization [1–5]. Because the interactions between neurons are generally asynchronous in biological and artificial neural networks, time delays are usually encountered. Since the existence of time delays is frequently one of the main sources of instability for neural networks, the stability analysis for delayed neural networks had been extensively studied and many papers have been published on various types of neural networks with time delays based on the LMI approach [6–14].

On the other hand, the main idea of passivity theory is that the passive properties of a system can keep the system internally stable. In addition, passivity theory is frequently used in control systems to prove the stability of systems. The problem of passivity performance analysis has also been extensively applied in many areas such as signal processing, fuzzy control, sliding mode control [15], and networked control [16]. The passivity idea is a promising approach to the analysis of the stability of NNs, because it can lead to more general stability results. It is important to investigate the passivity analysis for neural networks with time delays. More recently, dissipativity or passivity performances of NNs have received increasing attention and many research results have been reported in the literature, for example, [17–21].

In practice, the RNNs often exhibit the behavior of finite state representations (also called clusters, patterns, or modes) which are referred to as the information latching problems [22]. In this case, the network states may switch (or jump) between different RNN modes according to a Markovian chain, and this gives rise to the so-called Markovian jumping recurrent neural networks. It has been shown that the information latching phenomenon is recognized to exist universally in neural networks [23, 24], which can be dealt with extracting finite state representation from a trained network, that is, a neural network sometimes has finite modes that switch from one to another at different times. The results related to all kinds of Markovian jump neural networks with time delay can also be found in [25–27] and the references therein. It should be pointed out that all the above mentioned references assume that the considered transition probabilities in the Markov process or Markov chain are time invariant, that is, the considered Markov process or Markov chain is assumed to be homogeneous. It is noted that such kind of assumption is required in most existing results on Markovian jump systems [28, 29]. The detailed discussion about piecewise homogeneous and nonhomogeneous Markovian jumping parameters has been given in [30] and references therein.

On the other hand, a typical time delay called as leakage (or “forgetting”) delay may exist in the negative feedback terms of the neural network and it has a great impact on the dynamic behaviors of delayed neural networks and more details are given in [31–36]. In [34] the authors introduced leakage time-varying delay for dynamical systems with nonlinear perturbations and derived leakage delay-dependent stability conditions via constructing a new type of Lyapunov-Krasovskii functional and LMI approach. Recently, the passivity analysis for neural networks of neutral type with Markovian jumping parameters and time delay in the leakage term have been addressed in [37]. With reference to the results above, it has been studied that many results get to be found out for passivity analysis of Markovian jumping neural networks with leakage time-varying delays. Thus, the main purpose of this paper is to shorten such a gap by making the first attempt to deal with the passivity analysis problem for a type of continuous-time neural networks with time-varying transition probabilities and mixed time delays.

In this paper, the problem of passivity analysis of Markovian jump neural networks with leakage time-varying delay and discrete and distributed time-varying delays is considered. The Markov process in the under lying neural networks is assumed to be finite piecewise homogeneous, which is a special nonhomogeneous (time-varying) Markov chain. Motivated by [30] a novel Lyapunov-Krasovskii functional is constructed in which the positive definite matrices are dependent on the system mode and a triple-integral term is introduced for deriving the delay-dependent stability conditions. By employing a novel Lyapunov-Krasovskii functional having triple integral terms, new passivity leakage delay-dependent criteria are established to guarantee the passivity performance of the given systems. This performance not only depends on the upper bound of the time-varying leakage delay but also depends on the upper bound of the derivative of the time-varying leakage delay . When estimating an upper bound of the derivative of the Lyapunov-Krasovskii functional, we handle the terms related to the discrete and distributed delays appropriately so as to develop less conservative results. Two numerical examples are given to show the validity and potential of the development of the proposed passivity criteria.

*Notations.* Let denote the -dimensional Euclidean space and the superscript “” denotes the transpose of a matrix or vector. denote the identity matrix with compatible dimensions. For square matrices and , the notation denotes positive-definite (positive-semidefinite, negative, negative semidefinite) matrix. Let be a complete probability space with a natural filtration and stand for the correspondent expectation operator with respect to the given probability measure . Also, let and denote the family of continuously differentiable function from to with the uniform norm .

#### 2. Problem Description and Preliminaries

Fix a probability space , is the sample space, is the -algebra of subsets of the sample space, and is the probability measure on , and consider the following Markov jump neural networks with mixed time-delays:
where * * and are the state of the th neuron at time with leakage time varying delay and denotes the neuron activation function; is a diagonal matrix with positive entries; and , are, respectively, the connection weight matrix, the discretely delayed connection weight matrix, and the distributively delayed connection weight matrix; is the output of the neural network, and is the output; and denote the discrete delay and distributed delay, respectively, and the time varying delay satisfies
where , and are some real constants. By the simple transformation, model (1) has an equivalent form as follows:
Here, is a right continuous markov chain on the probability space taking values in a finite state space with transition rate matrix given by
in which , , and for is the transition rate from mode at time to mode at time and .

Similarly, the parameter is also a right continuous markov chain on the probability space taking values in a finite state space with transition rate matrix given by in which , , and for , are the transition rate from mode at time to mode at time and .

In this paper, we make the following assumption, definition, and lemmas for deriving the main result.

*Assumption 1. *Each activation function in (1) is continuous and bounded and satisfies
where , and and are known real scalars. It follows from (6) that the neural activation function satisfies

Lemma 2 (Jensen Inequality). *For any matrix , any scalars and with and a vector function such that the integrals concerned are well defined, the following inequality holds:
*

Lemma 3. *For any constant matrix and scalars such that the following inequalities hold
*

The main purpose of this paper is to establish a delay-dependent sufficient condition to ensure that neural networks (1) are passive.

*Definition 4. *The system (1) is said to be passive, if there exists a scalar such that for all and for all the solutions of (1), the following inequality
holds under zero initial conditions.

#### 3. Main Results

In this section, we derive a new delay-dependent criterion for passivity of the delayed Markovian jumping neural networks (1) using the Lyapunov-Krasovskii functional method combining with LMI approach. For presentation convenience, in the following, we denote Now, we establish the following passivity condition for the system (1).

Theorem 5. *The given Markovian jumping neural networks (1) is passive if there exist
**
positive symmetric matrices , , , , ; the positive definite matrices the diagonal matrices , , and a scalar such that for any the following LMI holds:
**
where
**
and the remaining coefficients are all zero.*

*Proof. *Denote and consider the following Lyapunov-Krasovskii functional for neural network (1):
where
Define infinitesimal generator (denoted by ) of the markov process acting on as follows:
It can be calculated that
From (15), it can be seen that
Based on the above equation, along the solution of the neural network (3), we obtain that for each
Moreover, based on Lemma 2, we can get the following inequalities:
By using Lemma 3, we can also get that
Similarly, we can use Lemmas 2 and 3 for other integrals. On the other hand, we have from (6) that for any ,
which is equivalent to
where denotes the unit column vector having 1 element on its th row and zeros elsewhere. Thus, for any appropriately dimensioned diagonal matrix , the following inequality holds:
Similarly, for any appropriately dimensioned diagonal matrices , and , the following inequalities also hold:
Using inequalities (20)–(23) in (19) and adding (26)–(27) in (19), we get
where with

Hence we can obtain from (10) that,
Now, to show the passivity of the delayed neural networks in (1), we set
where .

Using Dynkin’s formula, we have
Now, we can deduce that

Thus, if (33) holds, then since and holds under zero initial condition, from (31) it follows that for any , which implies that (13) is satisfied and therefore the delayed neural networks (1) are locally passive. Next we shall prove that as . Taking expectation on both sides of (28) and integrating from to we have
By using Dynkin’s formula, we have
Hence
Using Jenson’s inequality and (36), we have
Similarly, it follows from the definition of that
Hence, it can be obtained that
where
From (39) and (40), it can be deduced that the trivial solution of system (1) is locally passive. Then the solutions of system (1) is bounded on . considering (1), we know that is bounded on , which leads to the uniform continuity of the solution on . From (36), we note that the following inequality holds:
By Barbalats’ lemma [38], it holds that as and this completes the proof of the global passivity of the system (1).

*Remark 6. *When , the system (1) becomes
The system (42) can be written in its equivalent form as follows:
The time varying delay satisfies
where are some constants and the leakage delay is a constant.

Now, the passivity condition for the neural networks (43) is given in the following corollary and the result follows from Theorem 5.

Corollary 7. *Neural networks (43) are passive if there exist
**
positive symmetric matrices ,; the positive definite matrices ; the diagonal matrices ; and a scalar such that for any the following LMI holds:
**
where
*