Research Article | Open Access

# Lyapunov Stability Analysis of Gradient Descent-Learning Algorithm in Network Training

**Academic Editor:**L. Simoni

#### Abstract

The Lyapunov stability theorem is applied to guarantee the convergence and stability of the learning algorithm for several networks. Gradient descent learning algorithm and its developed algorithms are one of the most useful learning algorithms in developing the networks. To guarantee the stability and convergence of the learning process, the upper bound of the learning rates should be investigated. Here, the Lyapunov stability theorem was developed and applied to several networks in order to guaranty the stability of the learning algorithm.

#### 1. Introduction

Science has evolved from an attempt to understand and predict the behavior of the universe and the systems within it. Much of this owes to the development of suitable models, which agree with the observations. These models are either in a symbolic form which the humans use or in mathematical form that are found from physical laws. Most systems are causal, which can be categorized as either static, where the output depends on the current inputs, or dynamic, where the output depends on not only the current inputs but also past inputs and outputs. Many systems also possess unobservable inputs, which cannot be measured, but affect the systemβs output, that is, time series systems. These inputs are known as disturbances and aggravate the modeling process.

To cope with the complexity of dynamic systems, there have been significant developments in the field of artificial neural network during last three decades which have been applied for identification and modeling [1β5]. One major issue that instigates for proposing these different types of networks is to predict the dynamic behavior of many complex systems existing in nature. ANN is a powerful method in approximating a nonlinear system and mapping between input and output data [1]. Recently, wavelet neural networks (WNNs) have been introduced [6β10]. Such types of networks employ wavelets as the activation function in a hidden layer. Because of the ability of the localized analysis of wavelets collectively in their frequency and time domains and the learning ability of ANN, the WNN prompts a superior system model for complex and seismic applications. The majority of the applications of wavelet function are limited to a small dimension [11] although WNN can handle large-dimension problems as well [6]. Due to the dynamic behavior of recurrent network, they are suitable in dealing with the modeling of dynamic systems as compared with the static behavior of feed-forward network [12β19]. It has already been shown that recurrent networks are less sensitive to noise with relatively smaller network size and simpler structure. Their long-term prediction property makes them more powerful in dealing with dynamic systems. Recurrent networks are less sensitive to noise because the recurrent network could recognize and generate periodic waves in spite of the existence of a large amount of noise. This means that the network is able to regenerate the original periodic waves in the process of learning the teachers' signals with noises [2]. For unknown dynamic systems, the recurrent network results in a smaller-sized network as compared with the feed-forward network [12, 20]. For the time-series modeling, it generates a simpler structure [15β23] and gives long-term predictions [22, 24]. The recurrent network for system modeling learns and memorizes information in terms of embedded weights [21].

Different methods have been introduced for learning the parameters onnetwork based of the gradient descent. All learning methods like backpropagation-through-time [16, 17] or real-time recurrent learning algorithm [18] can be applied in order to adjust parameters of the feed-forward or recurrent networks. In [19], the quasi-Newton method was applied to improve the rate of convergence. In [9, 23], using the Lyapunov stability theorem, a mathematical way was introduced for calculating the upper bound of the learning rate for recurrent and feed-forward wavelet neural network based on the network parameters. Here, the Lyapunov stability theorem is developed and applied to several networks, and the learning procedure of the proposed networks is considered.

#### 2. Methodology

##### 2.1. Gradient-Descent Algorithm

The Gradient-descent (GD) learning can be achieved by minimizing the performance index as follows: where , is the output of the known network, is the actual data, and is the number of dataset. The reason for using a normalized mean square error is that it provides a universal platform for modeling evaluation irrespective of the application and target value specification while selecting an input to the model.

In the batch-learning scheme employing the -data set, achange in any parameter is covered by the following equation: and the parametric update equation is

##### 2.2. Lyapunov Method in Analysis of Stability

Consider a dynamic system, which satisfies

The equilibrium point is stable (in the sense of Lyapunov) at if for any there exists a such that

*Lyapunov Stability Theorem*

Let be a nonnegative function with the derivative along the trajectories of the system. Then(i)The origin of the system is locally stable (in the sense of Lyapunov) if is locally positive definite and is locally in and for all ;(ii)The origin of the system is globally uniformly asymptotically stable if is positive definite and excrescent and is positive definite.

To approve stability analysis of the networks based on GD learning algorithm, we can define discreet function as
Change of Lyapunov function is
from
Then
Difference of error is
where is the learning parameter and is error between output of plant and present output of network

By using (2.10) and (2.1) and putting them in (2.3),
where .

Therefore
where .

From the Lyapunov stability theorem, the stability is guaranteed if is positive and is negative. From (2.6), is already positive. The condition of stability depends on being negative. Therefore, is considered for all models.

Because , then the convergence condition is limited to
The maximum learning rate changes in a fixed range. Since does not depend on the model, the value of guarantees that the convergence can be found by minimizing the term of . Therefore,
where .

#### 3. Experimental Results

In this section, the proposed stability analysis is applied for some networks. The selected networks are neurofuzzy (ANFIA) [25, 26], Wavelet neurofuzzy, and recurrent wavelet network.

##### 3.1. Example 1: Convergence Theorems of the TSK Neurofuzzy Model

TSKmodel has a linear or nonlinear relationship of inputs in the output space. The rules of TSK model are in the following way: A linear form of in (3.1) is as follows: By taking the Gaussian membership function and an equal number of fuzzy sets to the rules with respect to the inputs, the firing strength of rules (3.1) can be written as where and are the center and standard deviation of the Gaussian membership functions, respectively. By applying the T-norm (product operator) of the membership functions of the premise parts of the rule and the weighted average gravity method for de-fuzzification, the output of the TSK model can be defined as

Theorem 3.1. *The asymptotic learning convergence of TSK neurofuzzy is guaranteed if the learning rate for different learning parameters follows the upper bound as will be mentioned below:
*

*Proof. *In equation (2.15) for neurofuzzy models can be written as
Because for all and since local models have same variables, that is, , therefore, from (3.7), (3.5) easily can be derived

##### 3.2. Example 2: Convergence Theorems of Recurrent Wavelet Neuron Models

Each neuron model in the proposed recurrent neuron models is summation or multiplication of Sigmoid Activation Function (SAF) and Wavelet Activation Function (WAF) as shown in Figure 1. Morlet wavelet function is considered in the recurrent models. In the series of developing different recurrent networks and neuron models, the proposed neuronsβ model is used in a one-hidden-layer feed-forward neural network as shown in Figure 2.

The output of feed-forward network is given inββ the following equation: where is the output of S-W neurons, is the weights between hidden neuron and output neurons, and is the number of hidden neuron,

The functions and are output of SAF and WAF for th S-W neuron, in the hidden layer, respectively. The functions and are expressed as follow. is th input. and are weights to input signal for SAF and WAF, in each hidden neuron, respectively.

To prove convergence of the recurrent networks, these facts are needed:Fact 1: let . Then .Fact 2: let . Then .Fact 3: let be a sigmoid function. Then Fact 4: let be a Morlet wavelet function. Then .

*(a) Summation Sigmoid-Recurrent Wavelet*

Suppose and .

From the facts 3 and 4: For parameter in all models
Therefore .

Differential of output of the model for another learning parameter is
Therefore,
Therefore
Therefore, .

*(b) Multiplication Sigmoid-Recurrent Wavelet*

From facts 3 and 4 suppose and

For parameter in all networks:
Therefore,
Therefore,
Therefore,
Therefore,

##### 3.3. Example 3: Convergence Theorems of the Wavelet Nuro-Fuzzy (WNF) Model

The consequent part of each fuzzy rule corresponds to a sub-WNN consisting of wavelet with the specified dilation value, where, in the TSK fuzzy model, a linear function of inputs is used while . Figure 1 shows the proposed WNN model which uses a combination of sigmoid and wavelet activation functions as a hidden neuron (Figure 2 without recurrent part) in the consequent part of each fuzzy rule.

Theorem 3.2. *The asymptotic learning convergence is guaranteed if the learning rate for different learning parameters follows the upper bound as will be mentioned below:
*

where , , or and or are the parametersβ learning rates of the consequent and the premise parts of the fuzzy rules. and are weights to inputs, signal for sigmoid and wavelet activation functions of local WNNs, in each hidden neuron, respectively. and are the center and standard deviation of the Gaussian membership functions of rule number m in WNF model, respectively.

*Proof. *In equation (2.15) for WNF models can be written as
Because for all , therefore (3.13) to (3.15) are easily derived.

From (2.15) and (3.4) for parameters or ,there is
and therefore (3.19) arederived.

#### 4. Conclusion

In this paper, a developed Lyapunov stability theorem was applied to guarantee the convergence of the gradient-descent learning algorithm in network training. The experimental examples showed that the upper bound of the learning parameter could be easily considered using this theorem. So, an adaptive learning algorithm can guaranty the fast and stable learning procedure.

#### References

- K. S. Narendra and K. Parthasarathy, βIdentification and control of dynamical systems using neural networks,β
*IEEE Transactions on Neural Networks*, vol. 1, no. 1, pp. 4β27, 1990. View at: Publisher Site | Google Scholar - S. Z. Qin, H. T. Su, and T. J. McAvoy, βComparison of four neural net learning methods for dynamic system identification,β
*IEEE Transactions on Neural Networks*, vol. 3, no. 1, pp. 122β130, 1992. View at: Publisher Site | Google Scholar - T. Yabuta and T. Yamada, βLearning control using neural networks,β in
*Proceedings of the IEEE International Conference on Robotics and Automation, (ICRA '91)*, pp. 740β745, Sacramento, Calif, USA, April 1991. View at: Google Scholar - P. Frasconi, M. Gori, and G. Soda, βLocal feedback multilayered networks,β
*Neural Computation*, vol. 7, no. 1, pp. 120β130, 1992. View at: Google Scholar - J. C. Patra, R. N. Pal, B. N. Chatterji, and G. Panda, βIdentification of nonlinear dynamic systems using functional link artificial neural networks,β
*IEEE Transactions on Systems, Man, and Cybernetics, Part B*, vol. 29, no. 2, pp. 254β262, 1999. View at: Publisher Site | Google Scholar - Q. Zhang and A. Benveniste, βWavelet networks,β
*IEEE Transactions on Neural Networks*, vol. 3, no. 6, pp. 889β898, 1992. View at: Publisher Site | Google Scholar - J. Zhang, G. G. Walter, Y. Miao, and W. Lee, βWavelet neural networks for function learning,β
*IEEE Transactions on Signal Processing*, vol. 43, no. 6, pp. 1485β1497, 1995. View at: Publisher Site | Google Scholar - T. I. Boubez and R. L. Peskin, βWavelet neural networks and receptive field partitioning,β in
*Proceedings of the IEEE International Conference on Neural Networks*, pp. 1544β1549, San Francisco, Calif, USA, March 1993. View at: Google Scholar - A. Banakar and M. F. Azeem, βArtificial wavelet neural network and its application in neuro-fuzzy models,β
*Applied Soft Computing*, vol. 8, no. 4, pp. 1463β1485, 2008. View at: Publisher Site | Google Scholar - Q. Zhang, βUsing wavelet network in nonparametric estimation,β
*IEEE Transactions on Neural Networks*, vol. 8, no. 2, pp. 227β236, 1997. View at: Publisher Site | Google Scholar - A. Benveniste, B. Juditsky, B. Delyon, Q. Zhang, and P. Y. Glorennec, βWavelets in identification,β in
*Proceedings of the 10th IFAC Symposium on System Identification, (SYSID '94)*, Copenhagen, Denmark, July 1994. View at: Publisher Site | Google Scholar - X. D. Li, J. K. L. Ho, and T. W. S. Chow, βApproximation of dynamical time-variant systems by continuous-time recurrent neural networks,β
*IEEE Transactions on Circuits and Systems*, vol. 52, no. 10, pp. 656β660, 2005. View at: Google Scholar - B. Srinivasan, U. R. Prasad, and N. J. Rao, βBack propagation through adjoints for the identification of nonlinear dynamic systems using recurrent neural models,β
*IEEE Transactions on Neural Networks*, vol. 5, no. 2, pp. 213β228, 1994. View at: Publisher Site | Google Scholar - P. Frasconi and M. Gori, βComputational capabilities of local-feedback recurrent networks acting as finite-state machines,β
*IEEE Transactions on Neural Networks*, vol. 7, no. 6, pp. 1520β1525, 1996. View at: Google Scholar - R. A. Jacobs, M. I. Jordan, S. J. Nowlan, and G. E. Hinton, βAdaptive mixtures of local experts,β
*Neural Computation*, vol. 3, pp. 79β87, 1991. View at: Publisher Site | Google Scholar - D. E. Rumelhart, G. E. Hinton, and R. J. Williams, βLearning internal representations by error propagation,β in
*Parallel Distributed Processing I*, D. E. Rumelhart and J. L. McClelland, Eds., pp. 675β695, MIT Press, Cambridge, UK, 1986. View at: Google Scholar - P. Werbos, βGeneralization of backpropagation with application to a recurrent gas Markov model,β
*Neural Networks*, vol. 1, pp. 339β356, 1988. View at: Google Scholar - R. J. Williams and D. Zipser, βA learning algorithm for continually running fully recurrent networks,β
*Neural Networks*, vol. 1, pp. 270β280, 1989. View at: Google Scholar - R. J. Williams and D. Zipser, βMechanical system modeling using recurrent neural networks via quasi-Newton learning methods,β
*Applied Mathematical Modeling*, vol. 19, no. 7, pp. 421β428, 1995. View at: Publisher Site | Google Scholar - C.-F. Juang, βA TSK-type recurrent fuzzy network for dynamic systems processing by neural network and genetic algorithms,β
*IEEE Transactions on Fuzzy Systems*, vol. 10, no. 2, pp. 155β170, 2002. View at: Publisher Site | Google Scholar - C.-H. Lee and C.-C. Teng, βIdentification and control of dynamic systems using recurrent fuzzy neural networks,β
*IEEE Transactions on Fuzzy Systems*, vol. 8, no. 4, pp. 349β366, 2000. View at: Publisher Site | Google Scholar - P. A. Mastorocostas and J. B. Theocharis, βA recurrent fuzzy-neural model for dynamic system identification,β
*IEEE Transactions on Systems, Man, and Cybernetics, Part B*, vol. 32, no. 2, pp. 176β190, 2002. View at: Publisher Site | Google Scholar - S. J. Yoo, Y. H. Choi, and J. B. Park, βGeneralized predictive control based on self-recurrent wavelet neural network for stable path tracking of mobile robots: adaptive learning rates approach,β
*IEEE Transactions on Circuits and Systems*, vol. 53, no. 6, pp. 1381β1394, 2006. View at: Publisher Site | Google Scholar | MathSciNet - T. G. Barbounis, J. B. Theocharis, M. C. Alexiadis, and P. S. Dokopoulos, βLong-term wind speed and power forecasting using local recurrent neural network models,β
*IEEE Transactions on Energy Conversion*, vol. 21, no. 1, pp. 273β284, 2006. View at: Publisher Site | Google Scholar - T. Takagi and M. Sugeno, βFuzzy identification of systems and its applications to modeling and control,β
*IEEE Transactions on Systems, Man and Cybernetics*, vol. 15, no. 1, pp. 116β132, 1985. View at: Google Scholar | Zentralblatt MATH - J. S. R. Jang, βANFIS: adaptive-network-based fuzzy inference system,β
*IEEE Transactions on Systems, Man and Cybernetics*, vol. 23, no. 3, pp. 665β685, 1993. View at: Publisher Site | Google Scholar

#### Copyright

Copyright © 2011 Ahmad Banakar. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.