- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Journal of Control Science and Engineering
Volume 2013 (2013), Article ID 697849, 14 pages
Stochastic Stabilization of Itô Stochastic Systems with Markov Jumping and Linear Fractional Uncertainty
1Institute of Intelligent Information Processing, Guizhou University, Guiyang 550025, China
2College of Computer Science and Information, Guizhou University, Guiyang 550025, China
Received 17 September 2012; Revised 19 December 2012; Accepted 6 January 2013
Academic Editor: Derong Liu
Copyright © 2013 Fei Long et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
For a class of Itô stochastic linear systems with the Markov jumping and linear fractional uncertainty, the stochastic stabilization problem is investigated via state feedback and dynamic output feedback, respectively. In order to guarantee the stochastic stability of such uncertain systems, state feedback and dynamic output control law are, respectively, designed by using multiple Lyapunov function technique and LMI approach. Finally, two numerical examples are presented to illustrate our results.
Modeled by a set of linear systems with the transitions among the linear systems governed by the Markov chain, the Markov jumping linear system (MJLS) can characterize and model different types of systems , for instance, fault-tolerant systems, target tracking systems, manufactory processes, networked control systems, multiagent systems, and so on. In the past years, many important results on MJLS have been addressed in the literature, such as, the stability analysis and control design which were discussed in [2–8]. Besides the aforementioned theoretical studies, MJLS also found applications in practical systems, such as networked direct current motor systems .
The interest of this topic lies in the fact that many practical systems, with random abrupt changes, sudden environmental changes, and so on, can be modeled by Itô stochastic linear systems with Markov jumping and configurable uncertainty . Therefore, it is very important to investigate robustly stochastic stabilization problem for the aforementioned Itô stochastic MJLS with linear fractional uncertainties. Recently, some results for Itô stochastic MJLS have been addressed in [11–18].
In this paper, motivated by the recent work , we consider robustly stochastic stabilization problem for Markov jumping Itô stochastic linear system subject to linear fractional uncertainty. An effective LMI-based approach for state feedback and dynamic output feedback controller is designed such that the closed-loop system is robust stable in the mean square sense, respectively.
The reminder of this paper is organized as follows: the problem statement is described in Section 2, while in Section 3 the stochastic stability for Itô stochastic MJLS with linear fractional uncertainty is addressed. In Section 4, the robustly stochastic stabilization of Itô stochastic MJLS with linear fractional uncertainty is investigated via state feedback and dynamical output feedback, respectively. Two numerical examples are presented in Section 5 to illustrate the obtained results. Finally, some conclusions are drawn in Section 6.
2. Problem Formulation
Consider the following unforced Itô stochastic MJLS with linear fractional uncertainties: where is a Markov process taking values in a finite space and is the state vector. is a standard Wiener process that is supposed to be independent of the Markov process . denotes the initial state of system at , and stands for the initial mode in the Markov process at . and , are system matrices with compatible dimensions that depend on and , and where , , , and are known real constant matrices with appropriate dimensions. is an unknown real matrix function with Lebesgue-measurable elements, satisfying with being the identity matrix.
The evolution of the Markov process is governed by the following probability transitions: where is the transition rate from mode to mode at time instant when , and is the little- notation defined by .
For the stochastic stability, we adopt the following definition. For more details, please refer to  and the references therein.
Definition 1. The Itô stochastic MJLS in (1) with all modes and all is said to achieve stochastic stability if there exists a finite positive constant such that the following inequality holds for any initial condition : where is the expectation conditioning on the initial value of .
3. Stochastic Stability
Before proceeding further, we recall the following lemma, which will be used in the proof of the stochastic stability of the Itô stochastic MJLS in (1).
Lemma 2 (see ). Let be given constant matrices with appropriate dimensions and norm-bounded time-varying uncertain matrix with . For any , one has
Lemma 3 (see ). Let the matrix G satisfy ; define the following set: Then, the set can be rewritten as
Theorem 4. The Itô stochastic uncertain MJLS in (1) is stochastically stable if there exists a set of symmetric and positive-definite matrices and a set of positive scalars such that the following sets of coupled inequalities hold for each and all admissible linear fractional uncertainties: where represents the blocks that are induced by symmetry and
Proof. Consider the following Lyapunov-like function:
where denotes the positive symmetric matrix. The infinitesimal generator can be considered as a derivative of the Lyapunov function along the trajectory of the Markov process in point at time . We need to derive the infinitesimal generator of first of all. According to the definition of , we have
where is a small positive number. Setting and applying the law of total probability and the probability transition of the Markov process yield
According to the Itô stochastic differential coefficient theorem , we have Note that and , then the infinitesimal generator becomes where .
Using Lemmas 2 and 3, we have where , .
Hence , where According to the Schur complement, , implies Therefore .
By the generalized Dynkin formula, we have The last inequality implies Furthermore, the condition in (8) indicates , so holds for any . Letting go to infinity, then, we know that
According to Definition 1, the Itô stochastic uncertain MJLS in (1) is stochastically stable. This ends the proof.
4. Stochastic Stabilization
4.1. State Feedback Stochastic Stabilization
In this subsection, we discuss how to design the robust state feedback control law for the following Itô stochastic uncertain MJLS: where , are system matrices with compatible dimensions that depend on and , and where , , , and are known real constant matrices with appropriate dimensions. is an unknown real matrix function with Lebesgue-measurable elements, satisfying with being the identity matrix. The robust state feedback control law to be designed is where is a gain with appropriate dimension to be determined for each mode . Notice that the gain of the above controller is model dependent, which requires the knowledge of the model at time to choose the appropriate gain among . Thus when the model switches from mode , which uses the gain , to mode, the controller gains must be switched instantaneously to the gain to guarantee the desired performances.
Theorem 5. Suppose that there exists a set of symmetric and positive-definite matrices , a set of matrices , and a set of positive scalars and such that the following sets of coupled inequalities hold for each and all admissible linear fractional uncertainties. Then the controller (24) with gain can stabilize stochastically the Itô stochastic uncertain MJLS in (22) where represents the blocks that are induced by symmetry and
Proof. With the use of the state feedback control law in (24) for system (22), the closed-loop system becomes
In order to guarantee the stochastic stability of system (27), using the stochastically stable condition in Theorem 4, we know that the following inequality should be satisfied for each : Let , and before and after multiplying the matrix inequality (28) by , then, we have
Applying expressions (2) and (23) of the linear fractional uncertainties, the above inequality can be described as
Again in view of Lemmas 2 and 3, the following two matrix inequalities are obvious: where Setting Then can be rewritten as follows: Let , and by means of the matrix inequalities (31) and (34), the closed-loop system (27) will be stochastically stable if the following matrix inequality holds: After using the Schur complement, we can get the LMIs (25), which end the proof of Theorem 5.
4.2. Dynamic Output Feedback Stochastic Stabilization
In Section 4.1, the design method of state feedback controller had been given under that the state vector and the mode are available at each time . However, some of the state variables are not measurable by their construction or the lack of appropriate sensors to give information. Alternatively we can use dynamic output feedback controller that uses the system’s measurement to design and to compute the control gains in this subsection.
In this subsection, we discuss how to design the dynamic output feedback control law for the following Itô stochastic uncertain MJLS: where , are system matrices with compatible dimensions that depend on and , and where , , , and are known real constant matrices with appropriate dimensions. is an unknown real matrix function with Lebesgue-measurable elements, satisfying with being the identity matrix. The dynamic output feedback control law is to be described by the following structure: where is the controller state and , , and are the controller gains to be determined.
Theorem 6. If there exists a set of matrices , , , and and a set of positive scalars , , and such that the following sets of inequalities (39)–(41) hold for each and all admissible linear fractional uncertainties, then the dynamic output feedback controller (38) with control gain (43) can stabilize stochastically the Itô stochastic uncertain MJLS in (36)
where represents the blocks that are induced by symmetry and
The dynamical output feedback controller gain is given by
Proof. Combining the dynamic system (36) with controller dynamics (38), we can obtain the following extended dynamics:
where and , are system matrices with compatible dimensions that depend on and , and
Applying Theorem 4, the dynamics (44) is stochastically stable if there exists a set of symmetric and positive-definite matrices such that the following inequalities hold for every :
Let , , and , and using Lemma 3, we have In view of (47) and Lemma 2, we have
Based on the matrix inequality (48), the condition (46) will be satisfied if the following holds:
It is obvious that inequality (49) is nonlinear in the design parameters , , , and , . In order to transform inequality (49) into an LMI, define , , , and as follows: where , are symmetric and positive-define matrices.
Before and after multiply inequality (49) by and , respectively, we have