Abstract
We firstly employ the fixed point theory to study the stability of cellular neural networks without delays and with time-varying delays. Some novel and concise sufficient conditions are given to ensure the existence and uniqueness of solution and the asymptotic stability of trivial equilibrium at the same time. Moreover, these conditions are easily checked and do not require the differentiability of delays.
1. Introduction
Cellular neural networks (CNNs) were firstly proposed by Chua and Yang in 1988 [1, 2] and have become a research focus for their numerous successful applications in various fields such as optimization, linear, and nonlinear programming, associative memory, pattern recognition, and computer vision. Owing to the finite switching speed of neurons and amplifiers in the implementation of neural networks, it turns out that the time delays are inevitable and therefore the model of delayed cellular neural networks (DCNNs) is of greater realistic significance. Research on the dynamic behaviors of CNNs and DCNNs has received much attention, and nowadays there have been a large number of achievements reported [3–5].
In fact, besides delay effects, stochastic and impulsive as well as diffusing effects are also likely to exist in the neural networks. As a result, they have formed complex CNNs including impulsive delayed reaction-diffusion CNNs, stochastic delayed reaction-diffusion CNNs, and so forth. One can refer to [6–11] for the relevant researches. Synthesizing the existing publications about complex CNNs, we find that Lyapunov method is the primary technique. However, we also notice that there exist lots of difficulties in the applications of corresponding results to practical problems and so it does seem that new methods are needed to address those difficulties.
Encouragingly, Burton and other authors have recently applied the fixed point theory to investigate the stability of deterministic systems and obtained more applicable results, for example, see the monograph [12] and the papers [13–24]. Furthermore, there has been found that the fixed point theory is also effective to the stability analysis of stochastic (delayed) differential equations, see [25–31]. Particularly, in [26–28], Luo used the fixed point theory to study the exponential stability of mild solutions of stochastic partial differential equations with bounded delays and with infinite delays. In [29, 30], Sakthivel and Luo used the fixed point theory to investigate the asymptotic stability in pth moment of mild solutions to nonlinear impulsive stochastic partial differential equations with bounded delays and with infinite delays. In [31], Luo used the fixed point theory to study the exponential stability of stochastic Volterra-Levin equations. With these motivations, we wonder if we can use the fixed point theory to study the stability of complex neural networks, thus obtaining more applicable results.
In the present paper, we aim to discuss the asymptotic stability of CNNs and DCNNs. Our method is based on the contraction mapping theory, which is different from the usual method of Lyapunov theory. Some new and easily checked algebraic criteria are presented ensuring the existence and uniqueness of solution and the asymptotic stability of trivial equilibrium at the same time. These sufficient conditions do not require even the differentiability of delays, let alone the monotone decreasing behavior of delays.
2. Preliminaries
Let denote the n-dimensional Euclidean space and represent the Euclidean norm. . . corresponds to the space of continuous mappings from the topological space to the topological space .
In this paper, we consider the cellular neural network described by and the following cellular neural network with time-varying delays as where and is the number of neurons in the neural network. corresponds to the state of the ith neuron at time . . denotes the activation function of the jth neuron at time and is the activation function of the jth neuron at time . The constant represents the connection weight of the jth neuron on the ith neuron at time . The constant represents the rate with which the ith neuron will reset its potential to the resting state when disconnected from the network and external inputs. The constant represents the connection strength of the jth neuron on the ith neuron at time , where corresponds to the transmission delay along the axon of the jth neuron and satisfies ( is a constant). , . and . Denote .
Throughout this paper, we always assume that for and therefore (2.1) and (2.3) admit a trivial equilibrium .
Denote by the solution of (2.1) with the initial condition (2.2) and denote by the solution of (2.3) with the initial condition (2.4).
Definition 2.1 (see [32]). The trivial equilibrium of (2.1) is said to be stable if for any , there exists such that for any initial condition satisfying ,
Definition 2.2 (see [32]). The trivial equilibrium of (2.1) is said to be asymptotically stable if it is stable and for any ,
Definition 2.3 (see [32]). The trivial equilibrium of (2.3) is said to be stable if for any , there exists such that for any initial condition satisfying ,
Definition 2.4 (see [32]). The trivial equilibrium of (2.3) is said to be asymptotically stable if it is stable and for any initial condition ,
The consideration of this paper is based on the following fixed point theorem.
Lemma 2.5 (see [33]). Let be a contraction operator on a complete metric space , then there exists a unique point for which .
3. Asymptotic Stability of Cellular Neural Networks
In this section, we will simultaneously consider the existence and uniqueness of solution to (2.1)-(2.2) and the asymptotic stability of trivial equilibrium of (2.1) by means of the contraction mapping principle. Before proceeding, we firstly introduce the following assumption:(A1) There exist nonnegative constants such that for , ,
Let , where () is the space consisting of continuous functions such that and as , here is the same as defined in Section 2. Also is a complete metric space when it is equipped with a metric defined by where and .
Theorem 3.1. Assume the condition (A1) holds. If the following inequalities hold where , then the trivial equilibrium of (2.1) is asymptotically stable.
Proof. Multiplying both sides of (2.1) with gives
which yields after integrating from 0 to as
Now, for any , we define the following operator acting on as
where
The following proof is based on the contraction mapping principle, which can be divided into two steps as follows.
Step 1. We need to prove . Recalling the construction of , we know that it is necessary to show the continuity of on and as well as for .
From (3.7), it is easy to see . Moreover, for a fixed time , we have
It is not difficult to see that as which implies is continuous on .
Next we shall prove for . Since , we get . Then for any , there exists a such that implies . Choose . It is then derived form (A1) that
As , we obtain as . So for . We therefore conclude that .
Step 2. We need to prove is contractive. For any and , we compute
As , is a contraction mapping.
Therefore, by the contraction mapping principle, we see there must exist a unique fixed point of in which means is the solution of (2.1)-(2.2) and as .
To obtain the asymptotic stability, we still need to prove that the trivial equilibrium of (2.1) is stable. For any , from the conditions of Theorem 3.1, we can find satisfying such that .
Let . According to what have been discussed above, we know that there must exist a unique solution to (2.1)-(2.2), and
where , .
Suppose there exists such that and as . It follows from (3.11) that .
As and , we obtain . Hence
This contradicts to the assumption of . Therefore, holds for all . This completes the proof.
4. Asymptotic Stability of Delayed Cellular Neural Networks
In this section, we will simultaneously consider the existence and uniqueness of solution to (2.3)-(2.4) and the asymptotic stability of trivial equilibrium of (2.3) by means of the contraction mapping principle. Before proceeding, we give the assumption as follows.(A2) There exist nonnegative constants such that for , ,
Let , where () is the space consisting of continuous functions such that on and as , here is the same as defined in Section 2. Also is a complete metric space when it is equipped with a metric defined by where and .
Theorem 4.1. Assume the conditions (A1)-(A2) hold. If the following inequalities hold where , then the trivial equilibrium of (2.3) is asymptotically stable.
Proof. Multiplying both sides of (2.3) with gives
which yields after integrating from 0 to as
Now for any , we define the following operator acting on
where
and on for .
Similar to the proof of Theorem 3.1, we shall apply the contraction mapping principle to prove Theorem 4.1. The subsequent proof can be divided into two steps.
Step 1. We need prove . To prove , it is necessary to show the continuity of on and for and . In light of (4.7), we have, for a fixed time ,
where
It is easy to see that . Thus, is continuous on . Noting and , we obtain is indeed continuous on .
Next, we will prove for . As we did in Section 3, we know and as . In what follows, we will show as . In fact, since , we have . Then for any , there exists a such that implies . Select . It is then derived from (A2) that
As , we obtain as , which leads to for and . We therefore conclude .
Step 2. We need to prove is contractive. For any and , we estimate
Hence,
As , is a contraction mapping and hence there exists a unique fixed point of in which means is the solution of (2.3)-(2.4) and as .
To obtain the asymptotic stability, we still need to prove that the trivial equilibrium of (2.3) is stable. For any , from the conditions of Theorem 4.1, we can find satisfying such that .
Let . According to what have been discussed above, we know that there exists a unique solution to (2.3)-(2.4), and
where
Suppose there exists such that and as . It follows from (4.13) that .
As , and
we obtain . Hence
This contradicts to the assumption of . Therefore, holds for all . This completes the proof.
Remark 4.2. In Theorems 3.1 and 4.1, we use the contraction mapping principle to study the existence and uniqueness of solution and the asymptotic stability of trivial equilibrium at the same time, while Lyapunov method fails to do this.
Remark 4.3. The provided sufficient conditions in Theorem 4.1 do not require even the differentiability of delays, let alone the monotone decreasing behavior of delays which is necessary in some relevant works.
5. Example
Consider the following two-dimensional cellular neural network with time-varying delays with the initial conditions , on , where , , is bounded by .
It is easily to know that for . Compute From Theorem 4.1, we conclude that the trivial equilibrium of this two-dimensional cellular neural network is asymptotically stable.
Acknowledgments
This work is supported by National Natural Science Foundation of China under Grant 60904028 and 71171116.