Abstract

An iterative learning control algorithm with an adjustable interval is proposed for nonlinear systems to accelerate the convergence rate of iterative learning control. For -norm, the monotonic convergence of ILC was analyzed, and the corresponding convergence conditions were obtained. The results showed that the convergence rate was mainly determined by the controlled object, the control law gain, the correction factor, and the iteration interval size and that the control law gain was corrected in real time in the modified interval and the modified interval shortened as the number of iterations increased, further accelerating the convergence. The numerical simulation shows the effectiveness of the proposed method.

1. Introduction

Iterative learning control (ILC) [1], as a model-free control, has a simple structure and does not require specific model parameters. It can force the operation to meet the designed requirements within a finite interval after sufficient rounds of iterations. Because of the above attributes, ILC has become popular among many scholars in recent years.

Existing studies on ILC theory mainly focus on proving the convergence of the learning algorithm, learning speed, and learning law structure and the robustness, analysis method, initial value, and application of the learning process [28]. The convergence rate is an important indicator for evaluating the performance of an algorithm and also applies to iterative learning algorithms. Since the beginning of their research on ILC theory, Kawamura et al. [9] have proposed using the information obtained in previous learning to construct control inputs for the subsequent learning to improve the learning speed. Later, researchers found that initial controls that were often overlooked had a great impact on the convergence rate. Arif et al. [10] used databases to store previous information to construct new initial control inputs that greatly reduce the number of iterations required and improve the learning speed. Based on the knowledge in databases, Yang et al. [11] used a linear weighting algorithm and curve fitting method to construct new initial control inputs to reduce the number of iterations and improve the convergence rate of the learning law. Wang et al. [12] proposed a Radial Basis Function Network- (RBFN-) based iterative learning algorithm to construct initial control inputs based on experience to minimize the distance between the initial output and the expected trajectory. The disadvantages of the above-mentioned algorithms in the literature are increased complexity of the control structure and a lack of rigorous proof as to whether they can improve the convergence rate or not. Bien and Huk [13] used a high-order iterative control algorithm to construct the initial control input. However, the initial inputs constructed using this method contain redundant information. Therefore, this approach actually decreases the system’s learning speed to some extent.

Lin et al. [14] proposed the concept and analysis method of learning speed and performed a detailed analysis of the main factors that affect learning speed for the case of an open-closed P-type learning law. They found that convergence rate is related not only to the initial deviation, initial control variables, and learning gain of each run but also to the system’s dynamic equations, operation interval size, and dimensions. Piao et al. [15] studied the convergence rate of linear time invariant systems using D-type and P-type learning laws as examples. In their study, the sufficient conditions for the convergence of the system were obtained using time-weighted norms, and the quantitative relation between the number of iterations and the convergence condition was given; meanwhile, the Frobenius norm properties were used together with the gradient method to solve for the optimum gain matrix for iterative learning laws. Xu and Tan [16] adopted the Q-factor as the performance indicator of convergence rate and theoretically expanded the convergence rate comparison method compared with the method of using the convergence condition in [15, 16].

The focus of previous work has been to analyze the effects of the initial control and convergence condition on convergence rate. However, there have been few findings on control law design that accelerates the convergence rate. In the literature [14], using exponential gain to accelerate learning algorithms for P-type and D-type closed-loop learning has been proposed. A large number of simulation experiments have shown that the convergence rate of this design is sufficient for use in applications. In contrast to traditional iterative learning control, in which the learning is repeated in a fixed time interval, the control algorithm proposed in this paper has the following characteristics: a modified interval that decreases as the number of iterations increases is designed at the time axis, and the learning gain is also modified in real time; beyond the modified interval, the error is within the tolerance range, and no correction is needed. That is, the interval that requires learning is continuously shortened until the learning is completed. The overall computational load of this algorithm is reduced, which accelerates the learning. In this paper, a rigorous mathematical proof will be given, and a simulation experiment will be conducted using numerical examples to validate the effectiveness and correctness of the theory.

2. Mathematical Knowledge

The mathematical knowledge required to analyze the problem is given in this section.

Define a continuous vector function ; , where is a positive real number. Then, the -norm of the vector-valued function can be expressed aswhere is a function defined on the interval .

Lemma 1 (see [17]). If , , and are nonnegative continuous functions over the interval and there exists a nonnegative constant such thatthen

3. Problem Description and Analysis

Consider a class of nonlinear systemswhere ; is the number of iterations; , , and are, respectively, the state, control, and output of the system; and , , and are matrices with the corresponding dimension.

For the convenience of the proof, reasonable assumptions regarding the algorithm are given as follows.

Assumption 2. The unknown nonlinear function satisfies the Lipschitz condition; namely, there exists a constant such that

Assumption 3. There exists a unique ideal control such that the state and output of the system are equal to designed and .

In the proposed algorithm, the interval is divided equally into subintervals with length such that

Assume that system (4) uses a closed-loop P-type learning law:where is the proportional learning gain, is the exponential correction factor, and is the current correction starting time (note that in the ()th iteration, the error in interval is within the tolerance range, and the error in the remaining interval is not within the tolerance range).

Let be the demonstrative function; that is,Then, based on (9), (7) can be written as follows: The accelerated closed-loop P-type iterative learning control law (10) with gain correction that we constructed above can be described as follows: assume that the input of the first run is and its value is arbitrarily assigned in the entire interval of . Thus, the error of the first run is . The purpose of introducing into (13) is to obtain the learning boundary point for the next iteration; in the th iteration, no learning is needed before , and the corrections are only needed for interval . The error in interval can be calculated using the designed trajectory and the output . The construction of control law (7) shows that learning is only needed for intervals that do not match the designed trajectory and not for the interval that has met it. Thus, the size of the interval that requires learning gradually approaches zero as the number of iterations increases.

Theorem 4. The iterative learning system described by (4) and (7) indicates that if the conditions,(I),(II),are satisfied, then As , the output of the system converges to the designed trajectory in ; that is,

Proof. Assuming that, in the th iteration, the control time point stops at , it is clear from (11) that . Therefore, when , the error can be written asThen, when , the following can be derived from (4) and condition (II):Equation (18) can be reorganized asTake the -norm for both sides of (19) and multiply by (), and then take the maximum value in and combine with Lipschitz conditions to derive the following:whereFrom (4),-norm is taken on both sides of (22) and combined with Lipschitz conditions to derive The following is derived from Lemma 1:Both sides of the above equation are multiplied by (), and the maximum value in is taken:where .

In interval , the following can be derived by substituting (25) into (20): Select a large enough thatThe following can be obtained after multiple iterations:Therefore,

It is shown that adopting a scheme to shorten the learning interval of control law (7) can ensure its monotonic convergence in -norm, and when is large enough, the size of the interval that requires learning decreases and becomes infinitely close to zero by iterations. Eventually, the system output follows the designed trajectory over the entire interval.

Remark 5. From (18), (19), (20), (22), (23), (24), and (25), we can see that as time increases, the convergence becomes faster.

Remark 6. The learning process indicates that as the number of iterations increases, the length of the learning interval is shortened, which improves the learning efficiency.

4. Simulation Results and Analysis

Consider the following SISO nonlinear continuous time-varying system:The initial conditions are and , and it is required to follow the designed trajectory within the time interval .

The values in for the first iteration of the controller were generated randomly using the rand function. The learning gain in the control law (7) was set to , and the exponential correction parameter was set to . The convergence condition of control law (7) was calculated to be . The convergence condition of the traditional closed-loop P-type control law was . Obviously, , indicating that control law (7) had a faster convergence rate than the conventional closed-loop P-type control law. Figures 1(a), 1(b), 1(c), and 1(d) represent the course of the system output and the designed trajectory in the 1st, 3rd, 5th, and 7th iterations, respectively, in which the red solid line is the designed trajectory and the blue solid line is the tracking trajectory. The curves in Figure 1 show that the length of the interval requiring system correction gradually decreased with increasing number of iterations, and eventually the system output completely followed the designed trajectory over the entire interval.

For the convenience of comparative analysis, numerical simulation was conducted using the proposed algorithm and the conventional closed-loop P-type iterative learning algorithm. The tracking errors over the number of iterations are shown in Figure 2, in which the iterative tracking error takes the form .

Figure 2 shows that the modified P-type closed-loop control law converged to be within the error band (0.01) after 6 iterations, while the traditional control law took 11 iterations to converge to the above-mentioned error band. It is clear that the control algorithm proposed in this paper significantly improved the learning speed compared with the traditional learning law. It gave a much smaller error after the same number of iterations than the traditional learning law.

5. Conclusions

In this paper, a novel iterative learning algorithm was designed for nonlinear systems. The learning interval of this algorithm is adjustable, and the learning gain is modified in real time. As the number of iterations increases, the interval that actually needs correction shortens, accelerating the convergence rate. The convergence rate was discussed for the case of a closed-loop P-type iterative learning law, and its convergence conditions were obtained in -norm. The results indicated that the convergence condition of this algorithm was determined not only by the proportional learning gain but also by the learning interval and exponential factor.

Competing Interests

The authors declare that there is no conflict of interests regarding the publication of this article.

Acknowledgments

This research was supported by the National Natural Science Fund of China (Grant 51407143), the Specialized Research Fund for the Doctoral Program of Higher Education of China (Grant 20136102120049), and the Natural Science Foundation of Shaanxi Province, China (2015JM5227).