Research Article | Open Access
Tianyu Wang, Quanxin Zhu, Jingwei Cai, "Mean-Square Exponential Input-to-State Stability of Stochastic Fuzzy Recurrent Neural Networks with Multiproportional Delays and Distributed Delays", Mathematical Problems in Engineering, vol. 2018, Article ID 6289019, 11 pages, 2018. https://doi.org/10.1155/2018/6289019
Mean-Square Exponential Input-to-State Stability of Stochastic Fuzzy Recurrent Neural Networks with Multiproportional Delays and Distributed Delays
We are interested in a class of stochastic fuzzy recurrent neural networks with multiproportional delays and distributed delays. By constructing suitable Lyapunov-Krasovskii functionals and applying stochastic analysis theory, It’s formula and Dynkin’s formula, we derive novel sufficient conditions for mean-square exponential input-to-state stability of the suggested system. Some remarks and discussions are given to show that our results extend and improve some previous results in the literature. Finally, two examples and their simulations are provided to illustrate the effectiveness of the theoretical results.
Since the theory and application of cellular neural networks was proposed by L.O. Chua and L. Yang in 1998, neural networks have become a hot topic. They can be applied to the analysis of static images and signal processing, optimization, pattern recognition, and image processing. Usually, a neural network is an information processing system. Its characteristic is local connections between cells, and its output functions are piecewise linear. Clearly, it is easy to realize large-scale nonlinear analog signals in real time and parallel processing, which improves the running speed. As is well-known, the stability is an important theoretical problem in the field of dynamics systems (e.g., see [1–24]). Thus, it is interesting to investigate the stability of nonlinear neural networks.
On one hand, the switching speed of amplifier is limited and the errors occur in electronic components. As a consequence, delays happen to dynamics systems, and the delays often destroy the stability of dynamics systems, even cause the heavy oscillation (e.g., see [25–35]). So it is significant to study the stability of delayed neural networks. For example,  discussed the global stability analysis for a class of Cohen-Grossberg neural network models. A new comparison principle is firstly introduced to study the stability of stochastic delayed neural networks in . Global asymptotic stability analysis for integrodifferential systems modeling neural networks with delays was investigated in . In , Zhu et al. considered the robust stability of Markovian jump stochastic neural networks with delays in the leakage terms. For more related results we refer the authors to [25, 40–43] and references therein. It is worthy to point out that all of the works aforementioned were focused on the traditional types of delays such as constant delays, time-varying bounded delays, and bounded distribute delays. However, delays in real lives may be unbounded. In this case, a class of so-called proportional delays can be used to describe the model of human brain, where delays give information of history and the entire history affects the present. Thus, it is interesting to study the stability of neural networks with proportional delays.
On the other hand, all of the works mentioned above were focused on the traditional neural networks models, which did not consider fuzzy logic. But in the factual operations, we always encounter some inconveniences such as the complicity and the uncertainty or vagueness. In fact, vagueness always opposite to exactness. Therefore, vagueness can not be avoided in the human way of regarding the world. So fuzzy theory is regarded as the best suitable setting to take vagueness into consideration. It is reported that there have appeared many results on the stability analysis of fuzzy neural networks in the literatures. For example, Li and Zhu introduced a new way to study the stability of stochastic fuzzy delayed Cohen-Grossberg neural networks . They used Lyapunov functional, stochastic analysis technique and nonnegative semimartingale convergence theorem to solve the problem. In , Balasubramaniam and Ali studied the robust exponential stability of uncertain fuzzy Cohen-Grossberg neural networks with time-varying delays. However, to the best of our knowledge, until now, there have been no works on the stability of fuzzy neural networks with proportional delays.
Motivated by the above discussion, in this paper we investigate the problem of the input-to-state stability analysis for a class of the stochastic fuzzy delayed recurrent neural networks with multiproportional delays and distributed delays. Some novel sufficient conditions are derived to ensure the mean-square exponential input-to-state stability of the suggested system based on constructing suitable Lyapunov-Krasovskii functionals and stochastic analysis theory, It’s formula and Dynkin’s formula. Several remarks and discussions are presented to show that our results extend and improve some previous results in the literature. Finally, two examples and their simulations are given to show the effectiveness of the obtained results.
The rest of the paper is as follows. In Section 2, we introduce the model, some necessary assumptions, and preliminaries. In Section 3, we investigate the mean-square exponential stability of the considered model. In Section 4, we provide two examples to illustrate the effectiveness of the obtained results. Finally, we conclude the paper in Section 5.
2. Model Formulation and Preliminaries
Let denote the family of continuous functions from to with the uniform norm . Denote by the family of all measurable, valued stochastic variables satisfying , and valued stochastic variables satisfying , in which stands for the correspondent expectation operator with respect to the given probability measure denotes the class of essentially bounded functions from to with , denotes real number. denotes dimensions Euclidean space.
In this section, we consider the following class of stochastic fuzzy delayed recurrent neural networks with multiproportional delays and distributed delays:for all , where represents the state variable of the th neuron at time is the self-feedback connection weight strength. The constants , ,,, and are the connection weights of the th neuron to the th neuron at time or . , , and are the th neuron activation functions at time or . is the control input of the ith neuron at time , and . and denote the fuzzy AND and fuzzy OR operation, respectively. The noise perturbation is a Borel measurable function, and , are scalar standard Brownian motions defined on a complete probability space with a natural filtration . The constants are proportional delay factors and satisfy , where are time-varying continuous functions that satisfy as and . is the time-varying delay, which satisfies and .
Throughout this paper, we assume that the following conditions are satisfied.
Assumption 1. There exist positive constants and such thatfor all and .
Assumption 2. There exist nonnegative constants , , and such thatfor all and .
Obviously, under Assumptions 1–3 we see that there exists a unique solution of system (1)-(2). Let denote the solution from the initial data on in . It is clear that system (1)-(2) has a trivial solution or zero solution corresponding to the initial data . By applying the following variable transformations , then system (1)-(2) is equivalently transformed into the following stochastic recurrent neural networks with constant delays and time-varying coefficientswhere
3. Main Results
Proof. Since system (6)-(7) is equivalent to system (1)-(2), we only need to prove that the trivial solution of system (6)-(7) is mean-square exponentially input-to-state stable. To this end, we let , , for the sake of simplicity.Then by Itô’s formula, we have the following stochastic differential equation:where and is the weak infinitesimal operator such thatLetting , then we obtain Now we define a Markov time as follows:By using the Dynkin formula, we havewhich impliesLetting on both sides (18), it follows from the monotone convergence theorem, (9), (10), and (12) thatOn the other hand, from the definition of , we haveCombining (19) and (20), the following inequation holds:From (21) we see that the trivial solution of system (6)-(7) is mean-square exponentially input-to-state stable. The proof of Theorem 4 is completed.
Remark 6. If we ignore the effects of delays, then system (1)-(2) becomes a stochastic recurrent neural network without delays. The results obtained in this paper are also applicable to the case of stochastic recurrent neural networks without delays.
Remark 7. Compared with the result in , our model is more general than that in . In fact, multiproportional delays and distributed delays are considered in this paper and they yield much difficulty in the proof of our result, whereas only a simple constant delay was discussed in .
4. Illustrative Examples
In this section, we will use two examples to show the effectiveness of the obtained result.
Example 1 (2-dimension case). Consider the case of 2-dimension stochastic recurrent neural networks with multiproportional delayswhereandOther parameters of system (22)-(23) are given as follows: Take , and from the definition of , we obtain . It is easy to check that Assumptions 1–3 are satisfied. Moreover, a simple computation yields