Research Article | Open Access
Analysis of Exponential Stability for Neutral Stochastic Cohen-Grossberg Neural Networks with Mixed Delays
This paper is concerned with the mean-square exponential input-to-state stability problem for a class of stochastic Cohen-Grossberg neural networks. Different from prior works, neutral terms and mixed delays are discussed in our system. By employing the Lyapunov-Krasovskii functional method, Itô formula, Dynkin formula, and stochastic analysis theory, we obtain some novel sufficient conditions to ensure that the addressed system is mean-square exponentially input-to-state stable. Moreover, two numerical examples and their simulations are given to illustrate the correctness of the theoretical results.
Over the last few years, Cohen-Grossberg neural networks have been paid more and more attention because of their potential applications in various fields such as signal processing, image processing, pattern recognition, associative memory, programming problems, and combinatorial optimization (see [1–5]). Cohen-Grossberg neural networks model was first introduced by Cohen and Grossberg in 1983, which has become one of the most important neural network models (see [6, 7]). Compared with recurrent neural networks, Hopfield neural networks, and cellular neural networks, it is more challenging and interesting to build Cohen-Grossberg neural networks model. In particular, recurrent neural networks, Hopfield neural networks, and cellular neural networks can be regarded as exceptional cases of Cohen-Grossberg neural networks. Many researchers have further studied the effect of Cohen-Grossberg neural networks models and have obtained some reliable conclusions (see [8–10]). Therefore, it is greatly significant to investigate Cohen-Grossberg neural networks.
As we all know, the stability of Cohen-Grossberg neural networks plays an essential role in practical applications since it is a prerequisite to ensure that a real system works properly. Also, the existence of time delays cannot be ignored in real systems for various reasons, such as the finite switching speed of amplifiers, the congestions in the signal processing and transmission between neurons. The time delay may cause instability, oscillation, and divergence to stochastic neural network systems. Given the actual situation, a great number of stability results have been presented on stochastic delayed neural networks studied in the literature (see [11–17]). In addition, neutral-type neural network is a particular time delay system because its main feature that the derivative of system state is connected to delays. In the physical sense, the state trend of a neutral delay system is related to the past state and the current state. Many physical systems can be described as the neutral-type model in the real world. In [18, 19], the stability of neutral-type stochastic neural network has been studied. Meanwhile, the neutral-type time delay is quite different from distributed delays and constant delays. There is no doubt that the distributed delay generally has integral term and acts on probability space. The constant delay is constant and discrete, which means that the effect of the state is delayed by a fixed value on the system. It can be observed that most of the literature works mainly focus on three simple cases containing constant delays, distributed delays, and time-varying delays (see [20–22]). In fact, mixed delays have been considered to be more efficient in modeling neural network systems because these simple delays are often not feasible when neural network systems become more complex. To sum up, it is of great significance to study the stability of stochastic Cohen-Grossberg neural networks with mixed delays.
Thankfully, there are a great number of stability results about the stochastic Cohen-Grossberg neural networks with mixed delays (see [23–26]). For example, by applying the LaSalle invariant principle of stochastic differential delay equations and the stochastic analysis theory as well as the adaptive feedback technique, Zhang and Deng  introduced several sufficient conditions to ensure the adaptive synchronization of Cohen-Grossberg neural network with mixed time-varying delays and stochastic perturbation. Based on a novel Lyapunov-Krasovskii functional and some new approaches and techniques, Zhu and Cao  investigated the problem of exponential stability for a class of Markovian jump impulsive stochastic Cohen-Grossberg neural networks with mixed time delays and known or unknown parameters.
Furthermore, it is worth considering that the control input has a great influence on the stochastic Cohen-Grossberg neural networks. Sontag first came up with the input-to-state stability to check robust stability. In [27–29], the mean-square exponential input-to-state stability is more common than the traditional exponential stability. Meanwhile, the traditional exponential stability consists of exponential stability, global asymptotical stability, and almost surely stability. As far as we know, the traditional stability is that the states of stochastic neural networks converge to the equilibrium point as time goes to infinity. However, it does not always happen in our real life. For instance, some systems including the pendulum, finance markets, and stock markets only remain bounded but necessarily converge to the equilibrium point. To date, few researchers have discussed the mean-square exponential input-to-state stability problem of stochastic Cohen-Grossberg neural networks due to the difficulty of mathematics. Hence, this situation encourages our current research.
Inspired by the above discussion, we will investigate the problem of the mean-square exponential input-to-state stability analysis for a class of the stochastic neutral-type Cohen-Grossberg neural networks with mixed delays in this paper. Different from the previous stability works, the mean-square exponential input-to-state stability of stochastic neutral-type Cohen-Grossberg neural network has not been investigated. By using the Lyapunov-Krasovskii functional, stochastic analysis theory, Itô formula, and Dynkin formula, the mean-square exponential input-to-state stability is obtained. Moreover, several examples and their simulations are given to illustrate the correctness and effectiveness of obtained results.
The remainder of this paper is structured as follows. In Section 2, we introduce the mathematical model of neutral stochastic Cohen-Grossberg neural networks, assumptions, and definition of the mean-square exponential stability. The main theoretical result and its proofs are derived in Section 3. In Section 4, two numerical examples and their simulations are given to guarantee the correctness of the proposed results. Finally, the conclusion and further discussion are presented in Section 5.
2. Model Description and Preliminaries
In this paper, we mainly focus on the following stochastic neutral-type Cohen-Grossberg neural networks with mixed delays:With initial conditions ,, here is the state variable of the th neuron at time ; and denote the amplification functions and behaved functions of the th unit at time , respectively. The constants ,,, and are the connection weight coefficients of the th unit on the th unit at time . , , , and are the neuron activation functions of the th unit at time . is the external constant input of the th neuron at time , and belongs to , where represents the class of essentially bounded functions from to with =. The noise perturbation is a Borel measurable function, and is an -dimensional standard Brownian motion defined on the complete probability space with a natural filtration . represents the constant delay and denotes the time-varying delay; are distributed delays. are supposed to be differential and satisfy , . Let , where is a positive constant. Let denote the family of continue functions from to with the norm , where is the Euclidean norm in . denotes the family of all measurable, -valued random variables such that , where stands for the correspondent expectation operator with respect to the given probability measure .
Remark 1. In fact, let and and replace in system (1) with ; our model reduces to the model in . In other words, the model discussed in  is a special case of the system (1). Since the delays are limited to time-varying delays and , our model is more general and useful in real life.
In order to obtain our main results, we need the following assumptions and definition.
Assumption 2. The functions , , , , and satisfy Lipschitz condition; i.e., there exist positive constants , , , , and such that for all .
Assumption 3. There exist nonnegative constants , , , and , , , such that for all .
Assumption 4. or .
Assumption 5 (see ). are bounded functions, and there exist positive constants and such that , .
Assumption 6 (see ). There exist positive constants , such that .
Assumption 7. is a matrix, denotes the matrix norm which satisfies .
Under Assumptions 2–4, it is not difficult to check that the Lipschitz and linear growth conditions hold, and thus there exists a unique solution of system (1). Let denote the solution which satisfies on in . It is clear that under Assumption 4, system (1) admits a trivial solution corresponding to the initial data .
Definition 8 (Zhu and Cao ). The trivial solution of system (1) is said to be mean-square exponentially input-to-state stable if for every and , there exist scalars , and such that the following inequality holds:
3. Main Results
Theorem 9. Under Assumptions 2–7, the trivial solution of system (1) is mean-square exponentially input-to-state stable, if there exist positive constants , , , , such that the following conditions hold:
Proof. Let . To obtain the mean-square exponential input-to-state stability, we can construct the following Lyapunov-Krasovskii functional:whereBy employing Itô formula, we get the following stochastic differential equation:where , , , and is the weak infinitesimal operator which satisfies where is a matrix and ; under Assumptions 2 and 3, we can obtain By using the elementary inequality and under Assumptions 5 and 6, we have By Itô formula, we get