Abstract
This work deals with numerical methods of parameter optimization for asymptotically stable systems. We formulate a special mathematical programming problem that allows us to determine optimal parameters of a stabilizer. This problem involves solutions to a differential equation. We show how to chose the mesh in order to obtain discrete problem guaranteeing the necessary accuracy. The developed methodology is illustrated by an example concerning optimization of parameters for a satellite stabilization system.
1. Introduction
Consider differential equation that describes a system equipped with a stabilizer. Here, is a parameter. It is assumed that for all and the zero equilibrium position of system (1.1) is asymptotically stable whenever . The parameter should be chosen to optimize, in some sense, the behaviour of the trajectories. The choice of this parameter can be based on various criteria; obviously, it is impossible to construct a stabilizer optimal in all aspects. For example, for a linear controllable system, the pole assignment theorem guarantees the existence of a linear feedback yielding a linear differential equation with any given set of eigenvalues. One can choose a stabilizer with a very high damping speed. However, such a stabilizer is practically useless because of the so called pick-effect (see [1, 2]). Namely, there exists a large deviation of the solutions from the equilibrium position at the beginning of the stabilization process, whenever the module of the eigenvalues is big.
The aim of this paper is to develop an effective numerical tool oriented to optimization of stabilizer parameters according to different criteria that appear in the engineering practice.
Throughout this paper, we denote the set of real numbers by and the usual -dimensional space of vectors with components in by . The space of absolutely continuous functions defined in with values in is denoted by . We denote by the usual scalar product in and by the Euclidean norm. By , we denote the closed unit ball, that is, the set of vectors satisfying . The transpose of a matrix is denoted by . We use the matrix norm . If and are two subsets in and , we use the following notations: , .
2. Statement of the Problem
Denote by the solution to the Cauchy problem where is a parameter from a compact set . Let for all . Consider the functions Here, are compact sets, and are norms in , and . Consider the following mathematical programming problem: Many problems of stabilization systems' parameters optimization can be written in this form.
Minimization of the Final Deviation
The problem is to determine the optimal values of the system parameters that guarantee minimal deviation of the system state from the zero equilibrium position at the final moment of time. This problem can be formalized as follows:
For linear systems with , this problem is an approximation for the maximization of the degree of stability [3].
Minimization of the Maximal Deviation
This problem consists of determination of parameters that correspond to minimization of the maximum deviation of trajectories and satisfy certain restrictions at the final moment of time. This problem can be formalized as follows:
The above problems are of interest for stabilization theory; they both have form (2.3). Problem (2.3) has some special features, and its solution can be useful for parameter optimization of stabilization systems; however, its study can hardly be performed analytically for more or less complex systems. For this reason, we focus on the numerical aspects of this problem.
3. Numerical Methods
Let be small enough. We approximate problem (2.3) by the following problem where , , , , and is the Euler approximation for the solution . Problem (2.3) can be approximated by problems (3.1) with any given accuracy.
Assume that where matrix has eigenvalues with negative real part and the function satisfies and the Lipschitz condition with for all and in a neighbourhood of the zero equilibrium position. Consider functions defined by (2.2), assuming that the balls are contained in a sufficiently small neighbourhood of the origin. Consider . Let and be sets of indices such that the points , , and , form a -net in and , , respectively. Define the functions Problem (3.1) can be written as Denote by and the optimal parameters for problems (2.3) and (3.6), respectively.
Theorem 3.1. For any , there exists such that is an admissible solution to the following problem:
This theorem allows one to choose the parameters of discretization in order to obtain optimal stabilizer parameters with a necessary precision. A rigorous formulation of this claim is the following. Denote by the value of the problem Assume that problem (2.3) is calm in Clarke's sense (see [4]). Then, there exists a constant satisfying the inequality for all sufficiently small. It follows from Theorem 3.1 that where , is the solution of problem (3.1), and .
The exact formulas for leading to the proof of Theorem 3.1 are presented in the Appendix. The main tool used to obtain them is the following version of Filippov-Gronwall inequality [5].
Theorem 3.2. Let , where is a symmetric positive definite matrix. Consider the functions and , satisfying the following condition for almost all . Then, for all , whenever , where is the solution to the Cauchy problem , .
Note that the use of this theorem allows us to obtain more precise estimates for the number of points in the meshes needed to achieve a given discretization accuracy. The estimates based on the usual Gronwall inequality can be significantly improved for asymptotically stable systems if we take into account the behaviour of the trajectories for large values of time. Theorem 3.2 is a natural tool for this analysis. For example, according to the classical estimates, the number of points in the mesh in , needed to ensure a given precision, grows exponentially with the length of the time interval. Meanwhile, the estimates obtained from Theorem 3.2 for asymptotically stable systems (see Theorems A.2 and A.6) give a linear growth of the number of points in the mesh. This result is of practical importance. Optimization problem (3.6) is a hard nonsmooth problem. Our computational experience shows that the usage of the Nelder-Mead method is the most adequate approach to solve it. The numerical solution of this problem significantly depends on the structure of the involved functions. The problem of optimal choice of parameters is solved only once, at the stage of the control system's development, so one could afford to dedicate more resources to its solution. However, if the mesh is constructed using the classical precision estimates, the required computational effort can be extremely high, making it impossible to solve the problem in a reasonable time. Our estimates for the number of points of discretization allow us to construct an adequate mesh and to significantly reduce the CPU time.
4. Example: Optimal Parameters for Satellite-Stabilizer System
Consider motion of a connected two-body system in a circular orbit around the Earth. Body 1 is a satellite with the center of mass , and body 2 is a stabilizer with the center of mass . These two bodies are linked to each other at the point through a dissipative hinge mechanism (Figure 1). Let be the center of mass of the system.
We use three reference frames: is the orbital coordinate frame, its axis is directed along the radius vector of the point with respect to the center of the Earth, is directed along the velocity of the point , and is normal to the orbit plane. The axes of referential frames and are the central principal axes of inertia for bodies 1 and 2, respectively. Consider motion of the system in the orbit plane supposing that the bodies are connected in their centres of mass; that is, the points , , , and coincide. Let and be the angles between the axis and the axes and respectively. Denote by , the derivative of with respect to time. The equations of motion for this system can be written as [6] Here, , , and , , are the principal moments of inertia of the bodies, is the damping coefficient of the system, and is the constant angular velocity of the orbital motion of the system's center of mass. Introducing a new independent variable and the dimensionless parameters the equations of motion can be written as Here, the dot denotes the derivative with respect to . The parameters satisfy the following conditions: We study small oscillations of system (4.3) in the vicinity of the equilibrium position The equations of motion, linearized in the vicinity of the above stationary solution, take the form The characteristic equation for system (4.6) is Analysis of (4.7) allows one to obtain the necessary and sufficient conditions of asymptotic stability. The region of asymptotic stability is given by Taking into account the feasibility conditions for the system parameters, we arrive at the following set of admissible parameters for our optimization problem:
4.1. The Maximal Degree of Stability
Consider the set described by (4.9). Denote by a parameter that belongs to the set . Let be the roots of (4.7). The inclusion implies that , . The degree of stability is defined by Consider the following problem In [7, 8], it is proved that the maximal degree of stability is achieved when all the roots of the characteristic equations are real and equal. This situation becomes possible only when the conditions are satisfied. The above system has two sets of solutions
4.2. Numerical Optimization
Denote by the solution of linear system (4.6) with , defined in the interval . The parameters belong to asymptotic stability region defined in (4.9). Consider the following problem: Problem (4.15) can be reduced to an optimization problem without constrains using quadratic penalty functions; see [9]. If is big enough, this problem approximates problem (4.11), where the concept of degree of stability is used. The parameters given by (4.13) and (4.14) are close to optimal solutions of problem (4.15) only when . Put . The results of simulations show that the value of problem (4.15) is about , independently on the values of admissible parameters . For example, the following values are optimal parameters for problem (4.15). The corresponding minimal value is . Parameters (4.13) and (4.14) give the values and , respectively.
In practice, it is important to consider smaller time intervals . Solve problem (4.15) with . In this case, we see that the value of problem really depends on the choice of parameters. Let us estimate the global minimum in this problem. To this end, we solve problem (4.15) using all combinations of the following values: as initial guesses for numerical optimization. We obtain the following two sets of parameters with the best value of the problem: The estimate for the global minimum is Meanwhile, the value corresponding to parameters (4.13) and (4.14) is 0.4660. Thus, we see that the methodology based on resolution of problem (2.3) can be more adequate in the practice than that one using the concept of degree of stability.
Since we are studying the behaviour of a nonlinear system in a vicinity of its equilibrium position, it is also important to estimate the deviation of the linearized system trajectories from zero. The stabilizer constructed for the linearized system makes sense only if its trajectories belong to a small vicinity of the equilibrium position; otherwise, the influence of the nonlinearity can destabilize the system even in a very small neighbourhood of the equilibrium.
Consider the following problem: In this problem, the solutions of system (4.6) at the moment are constrained to be in a neighbourhood of the equilibrium position with the radius 0.005. The obtained optimal solutions minimize the maximum norm of the damping process of linear system (4.6) in the interval . After numerical optimization, we get the following optimal parameters: The corresponding value of problem (4.21) is . We can see that the couple of parameters in (4.22) and (4.23) are slightly different from (4.18) and (4.19). Moreover, Taking the optimal parameters of problem (4.11), we get Thus, the stabilizer with the parameters corresponding to the maximal degree of stability yields more significant deviation of the trajectories from the equilibrium position than the stabilizer with the parameters obtained solving problem (4.21).
Our aim is to find optimal parameters for system (4.3). To this end, we solve problem (4.15), with , but now, stands for the solution of system (4.3) with . We get the following two sets of optimal parameters: Consider optimal parameters (4.26) and (4.27), and denote by the solution of system (4.3) corresponding to these parameters and satisfying . The optimal value is The above value of is even smaller than , obtained for the linearized system.
To compare the solutions of all optimization problem, we consider the following function Here, stands for the solution to system (4.3). Table 1 gives the values of the function for parameters (4.13), (4.18), (4.22), and (4.26).
Observe that the values in the last column of Table 1, computed for parameters (4.26), always satisfy the condition . On the other hand, this condition is not satisfied for parameters (4.13), obtained maximizing the degree of stability. For the parameters obtained for linearization, the condition is satisfied only if is sufficiently small. This illustrates the advantages of applying the introduced methodology, based on numerical solution of optimization problem (2.3), directly to nonlinear systems.
5. Conclusions
The methods usually applied to optimize the parameters of a stabilization system are based on the idea of the maximum stability degree, that is, the minimization of the system's transition time. These methods, however, face the problem of the so-called peak effect when the deviation of the system trajectory from the equilibrium increases with the decrease of the time of response. The approach suggested in this paper consists of a numerical analysis of a stabilization system based on a more complete and flexible description of the system behaviour capable to take into account not only the stability degree but also the maximum deviation of the trajectory on a given time interval or at a given moment. For this optimization problem, we develop a numerical method and prove that it can guarantee a given accuracy for the problem solution. We obtained more precise estimates for the number of points in the meshes needed to achieve a given discretization accuracy than the estimates based on the usual Gronwall inequality. This method is applied to optimization of a stabilization system for a satellite with a gravitational stabilizer. The obtained results show that the above approach can offer solutions more adequate for practical implementation than those given by optimization of the stability degree.
Appendix
A. The Mathematical Basis
In this Appendix, we present a series of theorems, with schematic proofs, containing explicit estimates for the fineness of discretization needed to obtain the necessary precision of approximations and to prove Theorem 3.1.
A.1. Linear Systems
Consider a linear system where is a matrix. Assume that all its eigenvalues have negative real part. Let be a symmetric positive definite matrix satisfying the Lyapunov equation [10] Set The quadratic form is the Lyapunov function for system (A.1). Denote by and the minimal and maximal eigenvalues of , respectively. Let be a positive constant. Consider the Euler approximation for system (A.1), The following theorem provides an explicit estimate for the constant guaranteeing the equality .
Theorem A.1. Let . Consider . If then the following inequalities hold: where
It is easy to see that . The proof of this theorem uses the induction and the inequality
Assume that constant satisfies condition (A.5). Consider the polygonal Euler approximation to solution of system (A.1) Let . Set
Theorem A.2. Let be given. Assume that condition is satisfied. If then the inequality holds for all , whenever .
This theorem is a consequence of the inequality of the inclusion and of Theorem 3.2 with function defined by (A.9) and function defined by (A.11).
The following theorem is also a consequence of Theorem 3.2.
Theorem A.3. Let and be solutions to system (A.1). Assume that . If then the inequality , holds whenever .
Consider now , where is a closed interval. Let be the solution of system (A.1), with Let . Assume that parameters of function defined by (A.11) satisfy the following conditions, Assume that , is a finite set of points uniformly distributed on . If then we have Let be the length of the interval . Consider a finite set , , such that the difference is a constant. If then the set contains . Consider the Euler approximation of solution
Theorem A.4. Let . If then the following inequality holds:
The proof of this theorem uses the results of Theorems A.1, A.2, and A.3.
A.2. Nonlinear Systems
Assume that is a twice continuously differentiable function satisfying . Consider the function , where is a matrix. Assume that the eigenvalues of the matrix have negative real parts. Consider the system Since is twice continuously differentiable, there exists a constant such that function satisfies the following Lipschitz condition: for all and in a small neighbourhood of the equilibrium position . Consider the set defined by (A.3) and constants , as before. Define the Euler approximation for system (A.27), where is a positive constant.
Theorem A.5. Assume that is a constant satisfying Let . If then the following inequalities hold: where
The proof of this theorem uses the Lipschitz condition (A.28) in the form and the mathematical induction method. It is easy to see that .
Assume that the constant satisfies condition (A.31) and consider the function Put where is defined by (A.10). Denote by the solution to system (A.27) with the initial condition .
Theorem A.6. Assume that is a constant satisfying and let . If and then the inequality holds for all , whenever .
This theorem is a consequence of Theorem 3.2 with function defined by (A.35) and function defined by (A.11).
Consider now the Cauchy problems
Denote by and the respective solutions. The following theorem is also a consequence of Theorem 3.2.
Theorem A.7. Let be a constant satisfying Assume that the trajectories and belong to the set . If then we have for all , whenever .
Let . Assume that the parameters of function defined by (A.11) satisfy the following conditions: Consider the balls where the constants satisfy the following conditions: For each index , take a finite set of points, , uniformly distributed in the ball . If then we have Let , , be closed intervals with length . Consider a finite set of points , , in each interval . It is assumed that the difference is a constant. Let Define the sets If then we have . Let . Denote by the solution to the Cauchy problem Consider the Euler approximation of the solution with satisfying condition (A.47).
Theorem A.8. Let be given. If then the following inequalities: hold.
The proof of this theorem follows from Theorems A.5, A.6, and A.7.
Acknowledgments
The authors are grateful to the reviewers for their valuable suggestions. This research is supported by the Portuguese Foundation for Science and Technologies (FCT), the Portuguese Operational Programme for Competitiveness Factors (COMPETE), the Portuguese Strategic Reference Framework (QREN), and the European Regional Development Fund (FEDER).