Journal of Applied Mathematics

Journal of Applied Mathematics / 2014 / Article

Research Article | Open Access

Volume 2014 |Article ID 145061 | 11 pages |

Existence and Exponential Stability of Solutions for Stochastic Cellular Neural Networks with Piecewise Constant Argument

Academic Editor: Sabri Arik
Received27 Oct 2013
Accepted06 Jan 2014
Published30 Mar 2014


By using the concept of differential equations with piecewise constant argument of generalized type, a model of stochastic cellular neural networks with piecewise constant argument is developed. Sufficient conditions are obtained for the existence and uniqueness of the equilibrium point for the addressed neural networks. th moment exponential stability is investigated by means of Lyapunov functional, stochastic analysis, and inequality technique. The results in this paper improve and generalize some of the previous ones. An example with numerical simulations is given to illustrate our results.

1. Introduction

Since Chua and Yang introduced cellular neural networks (CNNs) [1, 2] in 1988, delayed cellular neural networks (DCNNs) were proposed by Chua and Roska [3] in 1990; the dynamics of CNNs and DCNNs have received great attention due to their wide applications in classification of patterns, associative memories, optimization problems, and so forth. As is well known, such applications depend on the existence of an equilibrium point and its stability. Up to now, many results on stability of delayed neural networks have been developed [325].

In real nervous systems, there are many stochastic perturbations that affect the stability of neural networks. The results in [26] suggested that one neural network could be stabilized or destabilized by certain stochastic inputs. It implies that the stability analysis of stochastic neural networks has primary significance in the design and applications of neural networks, such as [1326], but only few works have been done on the th moment exponentially stable for stochastic cellular neural networks [2225].

The theory of differential equations with piecewise constant argument was initiated in Cooke and Wiener [27] and Shah and Wiener [28]. These equations represent a hybrid of continuous and discrete dynamical systems and combine the properties of both the differential and difference equations. It is well known that the reduction of differential equations with piecewise constant argument to discrete equations has been the main and possibly a unique way of stability analysis for these equations [2732]. Particularly, one cannot investigate the problem of stability completely, as only elements of a countable set are allowed to be discussed for initial moments. By introducing arbitrary piecewise constant functions as arguments, the concept of differential equations with a piecewise constant argument has been generalized in [12, 3338], where an integral representation formula was proposed as another approach to meet the challenges discussed above.

To the best of our knowledge, the equations with piecewise constant arguments were not considered as models of neural networks, except possibly in [12, 3638]. In [12], the authors assume CNNs may “memorize” values of the phase variable at certain moments of time to utilize the values during middle process till the next moment. Thus, they arrive at differential equations with a piecewise constant argument. Obviously, the distances between the “memorized” moments may be very variative. Consequently, the concept of generalized type of piecewise constant argument is fruitful. But these systems are deterministic; the dynamical behavior of stochastic neural networks with piecewise constant arguments has never been tackled.

Motivated by the discussion above, our paper attempts to fill the gap by considering stochastic cellular neural networks with piecewise constant arguments. In this paper, criteria on the th moment exponential stability of equilibrium point can be derived by constructing a suitable Lyapunov functional; the results obtained in this paper generalize and improve some of the existing results in [12, 38].

The remainder of this paper is organized as follows. In Section 2, we introduce some notations and assumptions. In Section 3, a sufficient condition for the existence and uniqueness of the solution is obtained. In Section 4, we establish our main results on th moment exponentially stable. A numerical example is given in Section 5 to demonstrate the theoretical results of this paper. Finally, some conclusions are given in Section 6.

2. Preliminaries

In this paper, let , be the family of -valued -adapted process such that, for every , , be the family of -valued -adapted process such that, for every , . , denotes the -dimensional real space, . We fix real-valued sequences such that with as .

We study stochastic cellular neural networks with piecewise constant arguments described by the differential equations: If , , . corresponds to the number of units in a neural network, stands for the potential (or voltage) of cell at time , , are activation functions, denotes the rate with which cell resets its potential to the resting state when isolated from the other cells and inputs, and and denote the strengths of connectivity between cells and at times and , respectively. denotes the external bias on the th unit. Moreover, is -dimensional Brownian motion defined on a complete probability space with a natural filtration generated by , where we associate with the canonical space generated by , and denote by the associated -algebra generated by with the probability measure . Let be the family of -valued random variables with , and let be the th row vector of .

Throughout this paper, the following standard hypotheses are needed.(H1)Functions , are Lipschitz-continuous on with Lipschitz constants , , respectively. That is, for all , .(H2)There exists a positive number such that , .(H3)Assume that , , .(H4)There exist nonnegative constants , such that for all , .

In the following, for further study, we give the following definitions and lemmas.

denotes a vector norm defined by

Definition 1 (see [26]). The equilibrium point of system (1) is said to be the th moment exponentially stable if there exist and such that , where is a solution of system (1) with initial value .
In such a case, The right hand of (6) is commonly known as the th moment Lyapunov exponent of this solution.
When , it is usually said to be exponentially stable in mean square.

Lemma 2 (see [26], (Burkholder-Davis-Gundy inequality)). Let . Define, for , Then, for every , there exist universal positive constants (depending only on ), such that for all .
In particular, one may takeif , ,if , ,if ,.

Lemma 3 (see [39]). If () denote nonnegative real numbers, then where denotes an integer.
A particular form of (9) is, namely,

Lemma 4 (see [39]). Assuming that there exists constant , , if , then the following inequality holds:

3. Existence and Uniqueness of Solutions

In this section, we will study the existence and uniqueness of the equilibrium point of neural networks (1).

Theorem 5. Assume that (H1)–(H4) are fulfilled. Then, for every , there exists a unique solution , of (1) such that and ; is a solution of the following integral equation:

Proof. Existence. Fix ; then there exists , such that . Without loss of generality, we assume that . We will prove that, for every , there exists a solution of (1) such that , .
For each , set and define, by the Picard iterations, Obviously, . Moreover, for , it is easy to see by induction that . For simpleness, we let Then where For any , where .
The Gronwall inequality implies Since is arbitrary, for , we must have Therefore .
Now we claim that, for all , where and will be defined below.
Firstly, we compute where Lemma 4 is used in the first inequality and Lemma 2 and Hölder inequality are used in the second inequality. So (20) holds for .
Next, assume (20) holds for ; then where .
That is, (20) holds for . Hence, by induction, (20) holds for all .
One can see from (20) that, for every , is a Cauchy sequence in . Hence we have as in . Letting in (19) gives Therefore, .
It remains to show that is a solution of system (1) satisfying . Note that Hence, we can let in (13), and (12) is derived. Again, using the same argument, we can continue from to . Hence, the mathematical induction completes the proof.
Uniqueness. Let and be the two solutions, . By noting we can easily show that The Gronwall inequality then yields that This implies that for . The uniqueness has been proved for . For every , we can prove the uniqueness by mathematical induction on . As the assumptions of the existence-and-uniqueness theorem hold on every finite subinterval of , then (1) has a unique solution on the entire interval (see [26]). Hence, the theorem is proved.

From the proof of Theorem 5, we easily obtain the following theorem.

Theorem 6. Assume that (H1)–(H3) are fulfilled. Then, for every , there exists a unique solution of the following system such that , :

Remark 7. When the system (1) neglects stochastic perturbations, it becomes the system (28) which is studied in [12, 38]. In order to obtain the existence and uniqueness, more restrictions on coefficients are needed in [12, 38], which might greatly restrict the application domain of the neural networks. Clearly, our results in this paper contain the results given in [12, 38], and the conditions are less restrictive compared with [12, 38].

4. th Moment Exponential Stability

In this section, we will establish some sufficient conditions ensuring the th moment exponential stability of the equilibrium point of system (1).

Let denote the family of all nonnegative functions on which are continuous once differentiable in and twice differentiable in . If , define an operator associated with (1) as where

Let ; then (1) can be written by where

It is clear that the stability of the zero solution of (31) is equivalent to that of the equilibrium of (1). Therefore, we restrict our discussion to the stability of the zero solution of (31).

For simplicity of notation, we denote

In order to obtain our results, the following assumption and Lemmas are needed:(H5).

Lemma 8. Let be a solution of (31) and let (H1)–(H5) be satisfied. Then the following inequality holds for all , where

Proof. Fix ; there exists , such that . Then from Lemma 4, it follows that Now we compute and : The Hölder inequality yields Substituting (38)–(40) into (37) yields that where , .
On the other hand, Lemma 2 and (H3) yield Substituting (41)-(42) into (36), we have where , .
By Gronwall inequality
Furthermore, for , we have Thus, it follows from condition (H5) that Therefore, (34) holds for ; the extension of (34) for all is obvious. This proves the theorem.

Lemma 9 (Mao [26]). Assume that there is a function and positive constants , , and , such that for all . Then for all .

For convenience, we adopt the following notation:

Theorem 10. Assume that (H1)–(H5) hold and, furthermore, that the following inequality is satisfied: Then system (31) is the th moment exponentially stable.

Proof. We define a Lyapunov function Obviously, (47) is satisfied with . Now we prove (48) holds.
For , , the operator with respect to (12) is given by