Research Article | Open Access

Volume 2011 |Article ID 145801 | https://doi.org/10.5402/2011/145801

Ahmad Banakar, "Lyapunov Stability Analysis of Gradient Descent-Learning Algorithm in Network Training", International Scholarly Research Notices, vol. 2011, Article ID 145801, 12 pages, 2011. https://doi.org/10.5402/2011/145801

# Lyapunov Stability Analysis of Gradient Descent-Learning Algorithm in Network Training

Academic Editor: L. Simoni
Received17 Mar 2011
Accepted13 May 2011
Published05 Jul 2011

#### Abstract

The Lyapunov stability theorem is applied to guarantee the convergence and stability of the learning algorithm for several networks. Gradient descent learning algorithm and its developed algorithms are one of the most useful learning algorithms in developing the networks. To guarantee the stability and convergence of the learning process, the upper bound of the learning rates should be investigated. Here, the Lyapunov stability theorem was developed and applied to several networks in order to guaranty the stability of the learning algorithm.

#### 1. Introduction

Science has evolved from an attempt to understand and predict the behavior of the universe and the systems within it. Much of this owes to the development of suitable models, which agree with the observations. These models are either in a symbolic form which the humans use or in mathematical form that are found from physical laws. Most systems are causal, which can be categorized as either static, where the output depends on the current inputs, or dynamic, where the output depends on not only the current inputs but also past inputs and outputs. Many systems also possess unobservable inputs, which cannot be measured, but affect the system’s output, that is, time series systems. These inputs are known as disturbances and aggravate the modeling process.

To cope with the complexity of dynamic systems, there have been significant developments in the field of artificial neural network during last three decades which have been applied for identification and modeling . One major issue that instigates for proposing these different types of networks is to predict the dynamic behavior of many complex systems existing in nature. ANN is a powerful method in approximating a nonlinear system and mapping between input and output data . Recently, wavelet neural networks (WNNs) have been introduced . Such types of networks employ wavelets as the activation function in a hidden layer. Because of the ability of the localized analysis of wavelets collectively in their frequency and time domains and the learning ability of ANN, the WNN prompts a superior system model for complex and seismic applications. The majority of the applications of wavelet function are limited to a small dimension  although WNN can handle large-dimension problems as well . Due to the dynamic behavior of recurrent network, they are suitable in dealing with the modeling of dynamic systems as compared with the static behavior of feed-forward network . It has already been shown that recurrent networks are less sensitive to noise with relatively smaller network size and simpler structure. Their long-term prediction property makes them more powerful in dealing with dynamic systems. Recurrent networks are less sensitive to noise because the recurrent network could recognize and generate periodic waves in spite of the existence of a large amount of noise. This means that the network is able to regenerate the original periodic waves in the process of learning the teachers' signals with noises . For unknown dynamic systems, the recurrent network results in a smaller-sized network as compared with the feed-forward network [12, 20]. For the time-series modeling, it generates a simpler structure  and gives long-term predictions [22, 24]. The recurrent network for system modeling learns and memorizes information in terms of embedded weights .

Different methods have been introduced for learning the parameters onnetwork based of the gradient descent. All learning methods like backpropagation-through-time [16, 17] or real-time recurrent learning algorithm  can be applied in order to adjust parameters of the feed-forward or recurrent networks. In , the quasi-Newton method was applied to improve the rate of convergence. In [9, 23], using the Lyapunov stability theorem, a mathematical way was introduced for calculating the upper bound of the learning rate for recurrent and feed-forward wavelet neural network based on the network parameters. Here, the Lyapunov stability theorem is developed and applied to several networks, and the learning procedure of the proposed networks is considered.

#### 2. Methodology

##### 2.1. Gradient-Descent Algorithm

The Gradient-descent (GD) learning can be achieved by minimizing the performance index as follows: where , is the output of the known network, is the actual data, and is the number of dataset. The reason for using a normalized mean square error is that it provides a universal platform for modeling evaluation irrespective of the application and target value specification while selecting an input to the model.

In the batch-learning scheme employing the -data set, achange in any parameter is covered by the following equation: and the parametric update equation is

##### 2.2. Lyapunov Method in Analysis of Stability

Consider a dynamic system, which satisfies

The equilibrium point is stable (in the sense of Lyapunov) at if for any there exists a such that

Lyapunov Stability Theorem
Let be a nonnegative function with the derivative along the trajectories of the system. Then(i)The origin of the system is locally stable (in the sense of Lyapunov) if is locally positive definite and is locally in and for all ;(ii)The origin of the system is globally uniformly asymptotically stable if is positive definite and excrescent and is positive definite.
To approve stability analysis of the networks based on GD learning algorithm, we can define discreet function as Change of Lyapunov function is from Then Difference of error is where is the learning parameter and is error between output of plant and present output of network
By using (2.10) and (2.1) and putting them in (2.3), where .
Therefore where .
From the Lyapunov stability theorem, the stability is guaranteed if is positive and is negative. From (2.6), is already positive. The condition of stability depends on being negative. Therefore, is considered for all models.
Because , then the convergence condition is limited to The maximum learning rate changes in a fixed range. Since does not depend on the model, the value of guarantees that the convergence can be found by minimizing the term of . Therefore, where .

#### 3. Experimental Results

In this section, the proposed stability analysis is applied for some networks. The selected networks are neurofuzzy (ANFIA) [25, 26], Wavelet neurofuzzy, and recurrent wavelet network.

##### 3.1. Example 1: Convergence Theorems of the TSK Neurofuzzy Model

TSKmodel has a linear or nonlinear relationship of inputs in the output space. The rules of TSK model are in the following way: A linear form of in (3.1) is as follows: By taking the Gaussian membership function and an equal number of fuzzy sets to the rules with respect to the inputs, the firing strength of rules (3.1) can be written as where and are the center and standard deviation of the Gaussian membership functions, respectively. By applying the T-norm (product operator) of the membership functions of the premise parts of the rule and the weighted average gravity method for de-fuzzification, the output of the TSK model can be defined as

Theorem 3.1. The asymptotic learning convergence of TSK neurofuzzy is guaranteed if the learning rate for different learning parameters follows the upper bound as will be mentioned below:

Proof. In equation (2.15) for neurofuzzy models can be written as Because for all and since local models have same variables, that is, , therefore, from (3.7), (3.5) easily can be derived

##### 3.2. Example 2: Convergence Theorems of Recurrent Wavelet Neuron Models

Each neuron model in the proposed recurrent neuron models is summation or multiplication of Sigmoid Activation Function (SAF) and Wavelet Activation Function (WAF) as shown in Figure 1. Morlet wavelet function is considered in the recurrent models. In the series of developing different recurrent networks and neuron models, the proposed neurons’ model is used in a one-hidden-layer feed-forward neural network as shown in Figure 2.

The output of feed-forward network is given in   the following equation: where is the output of S-W neurons, is the weights between hidden neuron and output neurons, and is the number of hidden neuron,

The functions and are output of SAF and WAF for th S-W neuron, in the hidden layer, respectively. The functions and are expressed as follow. is th input. and are weights to input signal for SAF and WAF, in each hidden neuron, respectively.

To prove convergence of the recurrent networks, these facts are needed:Fact 1: let . Then .Fact 2: let . Then .Fact 3: let be a sigmoid function. Then Fact 4: let be a Morlet wavelet function. Then .

(a) Summation Sigmoid-Recurrent Wavelet
Suppose and .
From the facts 3 and 4: For parameter in all models Therefore .
Differential of output of the model for another learning parameter is Therefore, Therefore Therefore, .

(b) Multiplication Sigmoid-Recurrent Wavelet
From facts 3 and 4 suppose and
For parameter in all networks: Therefore, Therefore, Therefore, Therefore,

##### 3.3. Example 3: Convergence Theorems of the Wavelet Nuro-Fuzzy (WNF) Model

The consequent part of each fuzzy rule corresponds to a sub-WNN consisting of wavelet with the specified dilation value, where, in the TSK fuzzy model, a linear function of inputs is used while . Figure 1 shows the proposed WNN model which uses a combination of sigmoid and wavelet activation functions as a hidden neuron (Figure 2 without recurrent part) in the consequent part of each fuzzy rule.

Theorem 3.2. The asymptotic learning convergence is guaranteed if the learning rate for different learning parameters follows the upper bound as will be mentioned below:

where , , or and or are the parameters’ learning rates of the consequent and the premise parts of the fuzzy rules. and are weights to inputs, signal for sigmoid and wavelet activation functions of local WNNs, in each hidden neuron, respectively. and are the center and standard deviation of the Gaussian membership functions of rule number m in WNF model, respectively.

Proof. In equation (2.15) for WNF models can be written as Because for all , therefore (3.13) to (3.15) are easily derived.
From (2.15) and (3.4) for parameters or ,there is and therefore (3.19) arederived.

#### 4. Conclusion

In this paper, a developed Lyapunov stability theorem was applied to guarantee the convergence of the gradient-descent learning algorithm in network training. The experimental examples showed that the upper bound of the learning parameter could be easily considered using this theorem. So, an adaptive learning algorithm can guaranty the fast and stable learning procedure.

1. K. S. Narendra and K. Parthasarathy, “Identification and control of dynamical systems using neural networks,” IEEE Transactions on Neural Networks, vol. 1, no. 1, pp. 4–27, 1990. View at: Publisher Site | Google Scholar
2. S. Z. Qin, H. T. Su, and T. J. McAvoy, “Comparison of four neural net learning methods for dynamic system identification,” IEEE Transactions on Neural Networks, vol. 3, no. 1, pp. 122–130, 1992. View at: Publisher Site | Google Scholar
3. T. Yabuta and T. Yamada, “Learning control using neural networks,” in Proceedings of the IEEE International Conference on Robotics and Automation, (ICRA '91), pp. 740–745, Sacramento, Calif, USA, April 1991. View at: Google Scholar
4. P. Frasconi, M. Gori, and G. Soda, “Local feedback multilayered networks,” Neural Computation, vol. 7, no. 1, pp. 120–130, 1992. View at: Google Scholar
5. J. C. Patra, R. N. Pal, B. N. Chatterji, and G. Panda, “Identification of nonlinear dynamic systems using functional link artificial neural networks,” IEEE Transactions on Systems, Man, and Cybernetics, Part B, vol. 29, no. 2, pp. 254–262, 1999. View at: Publisher Site | Google Scholar
6. Q. Zhang and A. Benveniste, “Wavelet networks,” IEEE Transactions on Neural Networks, vol. 3, no. 6, pp. 889–898, 1992. View at: Publisher Site | Google Scholar
7. J. Zhang, G. G. Walter, Y. Miao, and W. Lee, “Wavelet neural networks for function learning,” IEEE Transactions on Signal Processing, vol. 43, no. 6, pp. 1485–1497, 1995. View at: Publisher Site | Google Scholar
8. T. I. Boubez and R. L. Peskin, “Wavelet neural networks and receptive field partitioning,” in Proceedings of the IEEE International Conference on Neural Networks, pp. 1544–1549, San Francisco, Calif, USA, March 1993. View at: Google Scholar
9. A. Banakar and M. F. Azeem, “Artificial wavelet neural network and its application in neuro-fuzzy models,” Applied Soft Computing, vol. 8, no. 4, pp. 1463–1485, 2008. View at: Publisher Site | Google Scholar
10. Q. Zhang, “Using wavelet network in nonparametric estimation,” IEEE Transactions on Neural Networks, vol. 8, no. 2, pp. 227–236, 1997. View at: Publisher Site | Google Scholar
11. A. Benveniste, B. Juditsky, B. Delyon, Q. Zhang, and P. Y. Glorennec, “Wavelets in identification,” in Proceedings of the 10th IFAC Symposium on System Identification, (SYSID '94), Copenhagen, Denmark, July 1994. View at: Publisher Site | Google Scholar
12. X. D. Li, J. K. L. Ho, and T. W. S. Chow, “Approximation of dynamical time-variant systems by continuous-time recurrent neural networks,” IEEE Transactions on Circuits and Systems, vol. 52, no. 10, pp. 656–660, 2005. View at: Google Scholar
13. B. Srinivasan, U. R. Prasad, and N. J. Rao, “Back propagation through adjoints for the identification of nonlinear dynamic systems using recurrent neural models,” IEEE Transactions on Neural Networks, vol. 5, no. 2, pp. 213–228, 1994. View at: Publisher Site | Google Scholar
14. P. Frasconi and M. Gori, “Computational capabilities of local-feedback recurrent networks acting as finite-state machines,” IEEE Transactions on Neural Networks, vol. 7, no. 6, pp. 1520–1525, 1996. View at: Google Scholar
15. R. A. Jacobs, M. I. Jordan, S. J. Nowlan, and G. E. Hinton, “Adaptive mixtures of local experts,” Neural Computation, vol. 3, pp. 79–87, 1991. View at: Publisher Site | Google Scholar
16. D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning internal representations by error propagation,” in Parallel Distributed Processing I, D. E. Rumelhart and J. L. McClelland, Eds., pp. 675–695, MIT Press, Cambridge, UK, 1986. View at: Google Scholar
17. P. Werbos, “Generalization of backpropagation with application to a recurrent gas Markov model,” Neural Networks, vol. 1, pp. 339–356, 1988. View at: Google Scholar
18. R. J. Williams and D. Zipser, “A learning algorithm for continually running fully recurrent networks,” Neural Networks, vol. 1, pp. 270–280, 1989. View at: Google Scholar
19. R. J. Williams and D. Zipser, “Mechanical system modeling using recurrent neural networks via quasi-Newton learning methods,” Applied Mathematical Modeling, vol. 19, no. 7, pp. 421–428, 1995. View at: Publisher Site | Google Scholar
20. C.-F. Juang, “A TSK-type recurrent fuzzy network for dynamic systems processing by neural network and genetic algorithms,” IEEE Transactions on Fuzzy Systems, vol. 10, no. 2, pp. 155–170, 2002. View at: Publisher Site | Google Scholar
21. C.-H. Lee and C.-C. Teng, “Identification and control of dynamic systems using recurrent fuzzy neural networks,” IEEE Transactions on Fuzzy Systems, vol. 8, no. 4, pp. 349–366, 2000. View at: Publisher Site | Google Scholar
22. P. A. Mastorocostas and J. B. Theocharis, “A recurrent fuzzy-neural model for dynamic system identification,” IEEE Transactions on Systems, Man, and Cybernetics, Part B, vol. 32, no. 2, pp. 176–190, 2002. View at: Publisher Site | Google Scholar
23. S. J. Yoo, Y. H. Choi, and J. B. Park, “Generalized predictive control based on self-recurrent wavelet neural network for stable path tracking of mobile robots: adaptive learning rates approach,” IEEE Transactions on Circuits and Systems, vol. 53, no. 6, pp. 1381–1394, 2006. View at: Publisher Site | Google Scholar | MathSciNet
24. T. G. Barbounis, J. B. Theocharis, M. C. Alexiadis, and P. S. Dokopoulos, “Long-term wind speed and power forecasting using local recurrent neural network models,” IEEE Transactions on Energy Conversion, vol. 21, no. 1, pp. 273–284, 2006. View at: Publisher Site | Google Scholar
25. T. Takagi and M. Sugeno, “Fuzzy identification of systems and its applications to modeling and control,” IEEE Transactions on Systems, Man and Cybernetics, vol. 15, no. 1, pp. 116–132, 1985. View at: Google Scholar | Zentralblatt MATH
26. J. S. R. Jang, “ANFIS: adaptive-network-based fuzzy inference system,” IEEE Transactions on Systems, Man and Cybernetics, vol. 23, no. 3, pp. 665–685, 1993. View at: Publisher Site | Google Scholar