Abstract
We present a design method for iterative learning control system by using an output recurrent neural network (ORNN). Two ORNNs are employed to design the learning control structure. The first ORNN, which is called the output recurrent neural controller (ORNC), is used as an iterative learning controller to achieve the learning control objective. To guarantee the convergence of learning error, some information of plant sensitivity is required to design a suitable adaptive law for the ORNC. Hence, a second ORNN, which is called the output recurrent neural identifier (ORNI), is used as an identifier to provide the required information. All the weights of ORNC and ORNI will be tuned during the control iteration and identification process, respectively, in order to achieve a desired learning performance. The adaptive laws for the weights of ORNC and ORNI and the analysis of learning performances are determined via a Lyapunov like analysis. It is shown that the identification error will asymptotically converge to zero and repetitive output tracking error will asymptotically converge to zero except the initial resetting error.
1. Introduction
Iterative learning control (ILC) system has become one of the most effective control strategies in dealing with repeated tracking control of nonlinear plants. The ILC system improves the control performance by a self-tuning process in the traditional PID-type ILC algorithms for linear plants or affine nonlinear plants with nonlinearities satisfying global Lipschitz continuous condition [1–3]. Recently, the ILC strategies combined with other control methodologies such as observer-based iterative learning control [4], adaptive iterative learning control [5], robust iterative learning control [6], or adaptive robust iterative learning control [7], have been widely studied in order to extend the applications to more general class of nonlinear systems. However, more and more restrictions are required in theory to develop these learning controllers. Among these ILC algorithms, PID-type ILC algorithms are still attractive to engineers since they are simple and effective for real implementations and industry applications. A main problem of the PID-type ILC algorithms is that a sufficient condition required to guarantee learning stability and convergence will depend on plant’s input/output coupling function (matrix). In general, it is hard to design the learning gain if the nonlinear dynamic plant is highly nonlinear and unknown. In order to get the input/output coupling function (matrix), the ILC using a neural or fuzzy system to solve the learning gain implementation problem can be found in [8, 9]. A neural network or a fuzzy system was used to approximate the inverse of plant’s input/output coupling function (matrix). The inverse function (matrix) is claimed to be an optimal choice of the learning gain from a convergent condition point of view. As the nonlinear system is assumed to be unknown, some offline adaptive mechanisms are applied to update the network parameters in order to approximate the ideal optimal learning gain.
Actually, for control of unknown nonlinear systems, neural-network-based controller has become an important strategy in the past two decades. Multilayer neural networks, recurrent neural networks and dynamic neural network [10–16] were used for the design of adaptive controllers. On the other hand, fuzzy logic system, fuzzy neural network, recurrent fuzzy neural networks and dynamic fuzzy neural network were also a popular tool for the design of adaptive controllers [17–22]. These concepts have also been applied to the design of adaptive iterative learning control of nonlinear plants [23–25]. However, few ILC works were developed for general unknown nonlinear plants, especially nonaffine nonlinear plants. As the authors can understand, a real-time recurrent network (RTRN) was developed in [26] for real-time learning control of general unknown nonlinear plants. But unfortunately, their learning algorithm depends on the generalized inverse of weight matrix in the RTRN. If the generalized inverse of weight matrix does not exist, the learning control scheme is not implementable.
In this paper, we consider the design of an iterative learning controller for a class of unknown nonlinear dynamic plants. Motivated by our previous work in [27], an improved version of an identifier-based iterative learning controller is proposed by using an output recurrent neural network (ORNN). Two ORNNs are used to design an ORNN-based iterative learning control system. The proposed ORNN-based ILC system includes an ORNN controller (ORNC) and an ORNN identifier (ORNI). The ORNC is used as an iterative learning controller to achieve the repetitive tracking control objective. The weights of ORNC are tuned via adaptive laws determined by a Lyapunov-like analysis. In order to realize the adaptive laws and guarantee the convergence of learning error, some information of the unknown plant sensitivity is required for the design of adaptive laws. Hence, the ORNI is then applied as an identifier to provide the required information from plant sensitivity. In a similar way, the weights of ORNI are tuned via some adaptive laws determined by a Lyapunov-like analysis. Both of the proposed ORNC and ORNI update their network weights along the control iteration and identification process, respectively. This ORNN-based ILC system can be used to execute a repetitive control task of a general nonlinear plant. It is shown that the identification error will asymptotically converge to zero and repetitive output tracking error will asymptotically converge to zero except the initial resetting error.
This paper is organized as follows. The structure of ORNN is introduced in Section 2. In Section 3, we present the design of ORNC and ORNI for the ORNN-based ILC system. The adaptation laws are derived and the learning performance is guaranteed based on a Lyapunov-like analysis. To illustrate the effectiveness of the proposed ILC system, a numerical example is used in Section 4 for computer simulation. Finally a conclusion is made in Section 5.
In the subsequent discussions, the following notations will be used in all the sections.(i) denotes the absolute value of a function .(ii) denotes the usual Euclidean norm of a vector .(iii) denotes the norm of a matrix .
2. The Output Recurrent Neural Network
In this paper, two ORNNs are used to design an iterative learning control system. The structure of the ORNN is shown in Figure 1, which comprises an input layer, a hidden layer, and an output layer.
Let denotes the dimension of input vector of nonlinear function and denotes the number of neurons in the hidden layer, the ORNN which performs as an approximator of the nonlinear function is now described in a matrix form as follows: where and are output-hidden wight matrix and hidden-input weight matrix, respectively, is the external input vector, is the augmented neural input vector, and denotes the delay of ORNN output . The activation function vector is defined as where with being the th column vector, and , is a sigmoid function.
3. Design of Output-Recurrent-Neural-Network-Based Iterative Learning Control System
In this paper, we consider an unknown nonlinear dynamic plant which can perform a given task repeatedly over a finite time sequence as follows: where denotes the index of control iteration number and denotes the time index. The signals and are the system output and input, respectively. is the unknown continuous function, represents the respective output delay order. Given a specified desired trajectory , , the control objective is to design an output-recurrent-neural-network-based iterative learning control system such that when control iteration number is large enough, will converge to some small positive error tolerance bounds for all even if there exists an initial resetting error. Here the initial resetting error means that for all . To achieve the control objective, an iterative learning control system based on ORNN design is proposed in Figure 2. In this figure, denotes the delay in time domain and denotes the memory in control iteration domain.
Before we state the design steps of the proposed control structure, some assumptions on the unknown nonlinear system and desired trajectories are given as follows.(A1) The nonlinear dynamic plant is a relaxed system whose input and output are related by for all .(A2) There exists a bounded unknown upper bounding function such that , where the factor represents the sensitivity of the plant with respect to its input.(A3) The reference model is designed to generate the bounded desired trajectory , which is based on a specified bounded reference input with being a continuous function.
The design of the ORNN-based iterative learning control system is divided into two parts.
3.1. Part 1: Design of ORNC and Corresponding Adaptive Laws
Based on the assumptions on the nonlinear plant (5), we define a tracking error at th control iteration as follows: It is noted that there exist bounded constants such that the initial value of will satisfy The difference of between two successive iterations can be computed as [28] The ORNN is used to design an ORNC in order to achieve the iterative learning control objective. Let be the dimension of the external input vector and denote the number of neurons in hidden layer of the ORNC. The ORNC which performs as an iterative learning controller is described in a matrix form as follows: where and are output-hidden wight matrix and hidden-input weight matrix to be tuned via some suitable adaptive laws, respectively, and is the augmented neural input vector. For the sake of convenience, we define . Now substituting (9) into (8), we will have For simplicity, we define , , . After adding and subtracting to (10), we can find that Investigating the second term in the right hand side of (11) by using the mean-value theorem, we have where = with , has a value between and , .
Now if we substitute (12) into (11), we will have The adaptation algorithms for weights and of ORNC at (next) control iteration to guarantee the error convergence are given as follows: where is defined in assumption (A2). If we substitute adaptation laws (14) and (15) into (13), we can find that
Theorem 1. Consider the nonlinear plant (5) which satisfies assumptions (A1)–(A3). The proposed ORNC (9) and adaptation laws (14) and (15) will ensure the asymptotic convergence of tracking error as control iteration approaches infinity.
Proof. Let us choose a discrete-type Lyapunov function as then the change of Lyapunov function is Taking norms on (16), it yields for iteration . This further implies that , , for all for . Using Lyapunov stability of , and (7), the tracking error will satisfy This proves Theorem 1.
Remark 2. If the plant sensitivity is completely known so that and are available, then the control objective can be achieved by using the adaptation algorithms (14) and (15). However, the plant sensitivity is in general unknown or only partially known. In part 2, we will design an ORNN-based identifier (ORNI) to estimate the unknown plant sensitivity and then provide the sign function and upper bounding function of for adaptation algorithms of ORNC.
3.2. Part 2: Design of ORNI and Corresponding Adaptive Laws
After each control iteration, the ORNI subsequently begins to perform identification process. The trained ORNI will then provide the approximated plant sensitivity to the ORNC to start the next control iteration. We would like to emphasize that the ORNI only identifies the nonlinear plant after each control iteration. This concept is quite different from traditional control tasks [29] and very important to the proposed ORNN-based ILC structure.
The structure of ORNN is further applied to design an ORNI to identify the nonlinear plant after the th control iteration. The identification process is stated as follows. After each trial of controlling the nonlinear system, we collect the input output data and , as the training data for the identifier. When discussing the identification, we omit the control iteration index and introduce a new identification iteration index to represent the number of identification process. That is, the notation for the training data and the ORNI output are simplified as , and , respectively. For the ORNI, let be the dimension of external input vector and denote the number of neurons in hidden layer of the ORNI. The ORNI which performs as an iterative learning identifier for nonlinear plant (5) is now described in a matrix form as follows: where and are output-hidden wight matrix and hidden-input weight matrix to be tuned via some suitable adaptive laws, respectively, and is the augmented neural input vector. For the sake of convenience, we define .
Based on the assumptions on the nonlinear plant (5), we define an identification error at th identification process as follows: The difference of between two successive identification process can be computed as Now substituting (21) into (23), we will have For simplicity, we define , , . After adding and subtracting to (24), we can find Investigating the second term in the right hand side of (25) by using the mean-value theorem, we can derive where = with , has a value between and , .
Now if we substitute (26) into (25), we will have
The adaptation algorithms for weights and of ORNI at (next) identification process are given as follows: If we substitute adaptation laws (28) into (27), we have
Theorem 3. Consider the nonlinear dynamic plant (5) which satisfies assumptions (A1)–(A3). The proposed ORNI (21) and adaptation laws (28) will ensure that the asymptotic convergence of identification error is guaranteed as the numbers of identification approach infinity.
Proof. Let us choose a discrete-type Lyapunov function as then we can derive the change of Lyapunov function as Taking norms on (29), we have for iteration . This implies that , , for all for , and hence the identification error will satisfy , for all . This proves Theorem 3.
Remark 4. The ORNN is a promising tool for identification because it can approximate any “well-behaved” nonlinear function to any desired accuracy. This good function approximation is applied to estimate the unknown plant sensitivity in this paper. The plant sensitivity in (8) can be approximated as follows: Note that the index in the identifier output is removed once the identification process stops. Applying the chain rule to (21), it yields Also from (21), we have Since the inputs to ORNI are , and , we further have Thus, From (34), (35) and (37), we obtain where . If we define , and , then The sign function and upper bounding function of plant sensitivity after finishing the identification process at th control iteration can be obtained as follows: It is noted that we do not need the exact plant sensitivity for the design of adaptive law (14). Even though there may exist certain approximation error between and , we can still guarantee the convergence of learning error since only a upper bounding function is required. Also note that the value of ( or −1) can be easily determined from the identification result.
4. Simulation Example
In this section, we use the proposed ORNN-based ILC to iteratively control an unknown non-BIBO nonlinear dynamic plant [26, 29]. The difference equation of the nonlinear dynamic plant is given as where is the system output, is the control input. The reference model is chosen as where is a bounded reference input. The control objective is to force to track the desired trajectory as close as possible over a finite time interval except the initial point. The network weight adaptation for the ORNI and ORNC is designed according to (14), (15), and (28), respectively. In the ORNC, we set and , that is, only two hidden nodes in layer 2 are used to construct the ORNC. In a similar way, we let and , that is, only two hidden nodes in layer 2 are used to set up the ORNI. For simplicity, all the initial conditions of ORNC parameters are set to be 0 at the first control iteration. In addition, the initial ORNI parameters are set to be 0 at the first identification process which begins after the first control iteration. We assume that the plant initial condition satisfies where is a generator of random number with normal distribution, mean and variance . To study the effects of learning performances, we first show the maximum value of tracking error with respect to control iteration in Figure 3(a). It is noted that is omitted in calculating the maximum value of tracking error since it is not controllable. The identification error at 10th control iteration with respect to identification process is shown in Figure 3(b). According to the simulation results, it is clear that the asymptotic convergence proved in Theorems 1 and 3 is achieved. Since a reasonable tracking performance is almost observed at the 10th control iteration, the trajectories between the desired output and plant output at the 10th control iteration are shown to demonstrate the control performance in Figure 3(c). Figure 3(d) shows the comparison between the identification result of and the plant output . The nice identification result enables the ORNI to provide the required information for the design of ORNC. Finally, the bounded control input is plotted in Figure 3(e).
(a)
(b)
(c)
(d)
(e)
5. Conclusion
For controlling a repeatable nonaffine nonlinear dynamic plant, we propose an output-recurrent-neural-network-based iterative learning control system in this paper. The control structure consists of an ORNC used as an iterative learning controller and an ORNI used as an identifier. The ORNC is the main controller utilized to achieve the repetitive control task. The ORNI is an auxiliary component utilized to provide some useful information from plant sensitivity for the design of ORNC’s adaptive laws. All the network weights of ORNC and ORNI will be tuned during the control iteration and identification process so that no prior plant knowledge is required. The adaptive laws for the weights of ORNC and ORNI and the analysis of learning performances are determined via a Lyapunov-like analysis. We show that if the ORNI can provide the knowledge of plant sensitivity for ORNC, then output tracking error will asymptotically converge to zero except an initial resetting error. We also show that the objective of identification can be achieved by the ORNI if the number of identifications is large enough.
Acknowledgment
This work is supported by National Science Council, R.O.C., under Grant NSC99-2221-E-211-011-MY2.