- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Abstract and Applied Analysis

Volume 2014 (2014), Article ID 475808, 7 pages

http://dx.doi.org/10.1155/2014/475808

## A Simplified Predictive Control of Constrained Markov Jump System with Mixed Uncertainties

^{1}Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Institute of Automation, Jiangnan University, Wuxi 214122, China^{2}Department of Engineering, Faculty of Engineering and Science, University of Agder, 4898 Grimstad, Norway

Received 17 November 2013; Revised 27 January 2014; Accepted 14 February 2014; Published 26 March 2014

Academic Editor: Shuping He

Copyright © 2014 Yanyan Yin et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

A simplified model predictive control algorithm is designed for discrete-time Markov jump systems with mixed uncertainties. The mixed uncertainties include model polytope uncertainty and partly unknown transition probability. The simplified algorithm involves finite steps. Firstly, in the previous steps, a simplified mode-dependent predictive controller is presented to drive the state to the neighbor area around the origin. Then the trajectory of states is driven as expected to the origin by the final-step mode-independent predictive controller. The computational burden is dramatically cut down and thus it costs less time but has the acceptable dynamic performance. Furthermore, the polyhedron invariant set is utilized to enlarge the initial feasible area. The numerical example is provided to illustrate the efficiency of the developed results.

#### 1. Introduction

Hybrid systems are a class of dynamical systems denoted by an interaction between the continuous and discrete dynamics. In control community, the researchers tend to view hybrid systems as continuous state and discrete switching which focuses on the continuous state of dynamic system. Switched systems are a natural result from this point of view. Since switching systems can be applied to model the systems involving abrupt sudden changes which are widely found in the systems of economics and communications as well as manufacturing, more attention has been paid to them (see robust stabilization [1], finite-time analysis [2] and asynchronous switching [3]). When the system model is linear and the switching is driven by Markov process, it leads to Markov jump linear system (MJS). Specifically, MJS presents a stochastic Markov chain to describe the random changes of system parameters or structures, where the dynamic of MJS is switching among the models governed by a finite Markov chain. Due to this superiority, MJS has been widely investigated during the last twenty years. Attractive pioneer works have been obtained (see controller design [4], 2D MJS control [5], peak-to-peak filtering [6], and finite-time control [7, 8]). However, the cases of completely known transition probability (TP) considered in [4–8] are not always achievable since the TP is not easy to be fully accessible (see the delay or packet loss in networked control systems [9]). Thus it is necessary to investigate the partly unknown case [10–12].

On the other hand, the systems in practice are usually subject to input/output constraints. Thus, model predictive control (MPC) is then introduced to solve the problem of MJS with constraints since MPC can explicitly solve the constraints in control action. Successful MPC application in discrete-time MJS can be obtained in [13, 14]. Normally, MPC is reformulated as online quadratic program and results have been reported (see stability [15, 16] and enlarged terminal sets [17]). It should be noted that the online computation in the literature [15–17] leads to heavy computational burden. Thus, the researchers attempted to try a new alternative method to solve the problem. For this reason, explicit MPC [18] is presented. However, when the size of system increases, the time of searching explicit MPC law will also increase sharply.

Based on the above analysis, a simplified MPC design framework is introduced to reduce the burden of online computation for the constrained MJS with mixed uncertainties. The basic idea is that () steps of mode-dependent MPC are designed to steer the state to a final neighbour area which includes the origin. Then the final step of robust mode-independent MPC is devised to force the state towards the origin regardless of model uncertainty and transition probability uncertainty. This simplified MPC dramatically reduces the burden of computation with minor performance loss, which implies good balance between the calculation time and dynamical performance. Furthermore, the polyhedron invariant set is applied to further enlarge the initial feasible area.

The construction of the paper is as follows. Section 2 gives the basic dynamical of the system. Section 3 gives the finite-step simplified MPC algorithm and it is formulated as LMIs. Section 4 presents a numerical example to show the efficiency of the results. Section 5 concludes the paper.

*Notations*. The notations are as follows: denotes a -dimensional Euclidean space, stands for the transpose of a matrix, denotes the expectation of the stochastic process or vector, a positive-definite matrix is described as , means the unit matrix with appropriate dimension, and means the symmetric term in a symmetric matrix.

#### 2. Problem Statement and Preliminaries

The constrained discrete-time MJSs with mixed uncertainties are considered in this paper: where , , , respectively, denote the state vector, the input vector, and the controlled output vector. The discrete-time Markov stochastic process takes values in a finite set , where contains modes of system (1), , and represents the initial mode. The uncertain system model and belong to the model sets Inputs and outputs constraints are subject to The transition probability (TP) matrix is denoted by , , where is the transition probability from mode at time to mode at time . The elements in TP matrix satisfy and : The uncertain transition probability (TP) implies that some elements in are unknown; a four-mode transition probability (TP) matrix may be where “?” represents the inaccessible element in TP matrix. For convenience, we denote , for all mode at sampling time , if , and redescribe it as , for all , where represents the th exact element in the th row of , .

Some preliminaries are introduced before proceeding.

*Definition 1 (see [6]). *For any initial mode and state , discrete-time MJS (1) is said to be stochastically stable if

*Definition 2. *For MJS (1), an ellipsoid set associated with the state is said to be asymptotically mode-dependent stable, if the following holds, whenever , then for and when .

Next, we first derive the online optimal MPC algorithm for system (1). The aim is to minimize the function cost related to worst-case performance and then in Section 4 the corresponding simplified MPC algorithm will be derived. Finally the polyhedron invariant set is applied to further improve the initial feasible district.

#### 3. Simplified MPC Design

##### 3.1. Online Optimal MPC

Theorem 3. *Consider MJS (1) with model uncertainties (2) and partly unknown TP matrix (6), at sampling time , if there exist a set of matrices , such that the following holds:
**
s.t. **
Then, it decides an upper bound on , where , , , are positive definite weighting matrices.*

*Proof. *It is assumed that at the sampling time , a state-feedback law , is applied to minimize the worst cost function of ; it is easy to show that is an upper bound on . Let , , be a quadratic Lyapunov function. For any , the following constraint holds
Summing (12) from to on both sides and using the fact or , we obtain
which implies that is an upper bound on .

Theorem 4. *Consider MJS (1) with polytope model uncertainties (2) and partly unknown TP matrix (4), if there exist a set of positive definite matrices , , such that the following optimization problem (12) has an optimal solution:
**
s.t. **
then, the mode-dependent state-feedback which minimizes the upper bound on and simultaneously stabilizes the closed-loop system within an ellipsoid is calculated by , , where , , , , , , respectively, describe the th, th diagonal element of , , and , respectively, describe the th and th element of input and output constraints, , .*

* Proof. *Let ; in (13) can be solved by the following LMIs:
The input/output constraints are guaranteed by (16) and (17); the proof is similar to [19]; here we omit the proof. Equation (11) is equivalent to
Since , , , it leads to
One sufficient condition to ensure (22) is
Considering the Schur theory complement lemma, (16) and (17) can be derived.

Actually the feedback controller can make the closed-loop system stable in the ellipsoid . Assume that the optimal , at the moment are
Equations (18) and (19) lead to
is the optimal value at moment ; is a feasible one at moment . By the optimum definition,
then,
It is shown that decrease strictly as ,

From Definition 1, the system is stochastically stable. From (27), then
This implies that the ellipsoid is an asymptotically stable invariant one, which completes the proof.

Corollary 5. *Consider MJS (1) with model uncertainties (2) and TP matrix (4) at current moment ; supposing that there exists a set of positive definite matrices , , such that the following optimization problem has an optimal solution:
**
s.t. ** then the mode-independent state-feedback law can minimize the upper bound ** on the objective function ** and stabilize the closed-loop system in the ellipsoid ** and it is obtained by **, **, where **, **, **, **, ** and **, respectively, describe the **th and **th diagonal element of **, **, and ** and **, respectively, describe the **th and **th element of input and output constraints, **, **.*

##### 3.2. Simplified MPC Design

In this section, a simplified MPC for uncertain MJS (1) is developed based on the online algorithm in Theorem 4; Figure 1 shows the simplified MPC schematic diagram. Then the simplified mode-independent feedback controller is designed regardless of model uncertainty and TP uncertainty since much more constraints will be nonactive in the neighboring region of origin and this freedom of feasibility is applied to improve the procedure of controller design.

Theorem 6. *Consider uncertain MJS (1) associated with an initial state satisfying ; the simplified MPC Algorithm 7 robustly stabilizes the closed-loop system.*

*Proof. *For the -step implementation at , , the selection for in Algorithm 7 implies , which means the constructed ellipsoid is embedded in , that is, . For a settled , is decreasing monotonically associated with , which guarantees the unique search in the search table for the largest for . If belongs to and , , by applying Theorem 3, the control law will steer the state in to . Finally, the controller (applying Corollary 5) make the state to be in and converge to the origin. Furthermore, the LP programming algorithm is utilized to remove redundant constraints [20] and construct a sequence of polyhedral invariant set for MJS and thus enlarge the feasible domain.

*Algorithm 7 (simplified MPC applying polyhedral invariant set). *
Simplified MPC design is as follows.(1)Select , , which satisfy , .(2)For step 1 to , calculate the corresponding mode-dependent gains , , , , by applying Theorem 4 and store them in a search table.(3)For each , construct the corresponding polyhedral invariant set by the following algorithm: let , . Select row from and then check if is redundant through solving the Linear programming:
If , it implies that the constraint is nonredundant; then renew the nonredundant constraints as , .(4)Online implementation: search the state in the search table to fix the needed index , decide the smallest polyhedral invariant set , and finally implement .(5)Online implementation: continue to check if is satisfied; if it is true, then apply .

*Remark 8. *It should be noted that the more approximation of optimality can be obtained as increases; here can be chosen according to different prior requirements. Thus, we can adjust the numbers of design step in terms of different requirements.

#### 4. Illustrative Example

Consider the discrete-time MJS with four modes (): The detailed constraints are and , initial state is , and . The positive definite weighting matrices are and . The partly unknown TP matrix is randomly generated in Table 1.

Here we will show the 5-step example of the proposed Algorithm 7. Firstly, a state set is designed to compute the corresponding feedback gains . It is noted that the sequence of states guarantees that the constructed polyhedral invariant sets are embedded, that is, . In this example, the first four mode-dependent feedback laws , are obtained. When the state goes into the smallest polyhedral invariant set, the final-step (the fifth-step) gain is designed to steer the state to the origin regardless of model uncertainty and TP uncertainty.

For each chosen in Figure 2, the 5-step ellipsoid invariant sets (purple solid lines) and 5-step polyhedral invariant sets (blue and orange alternant dot dash lines) are illustrated using the numbers 1 to 5. The stabilizable region of polyhedral invariant set constructed by Algorithm 7 is dramatically larger than that of ellipsoid invariant set while the dynamic response of simplified algorithm is comparable with online algorithm.

The results are computed at the same platform (AMD 2.1 GHz, memory 3.0 GB and MATLAB R2010a); the average time and variances of 30 times’ running of the system are shown in Table 2. From the table, the burden of computation is significantly reduced by simplified algorithm.

#### 5. Conclusions

The problem of simplified predictive controller design for MJS with mixed uncertainties is investigated. The simplified algorithm drastically reduces the online computational burden with only a little loss of performance. A numerical example is provided to illustrate the validity of the results.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

This work was partially supported by the National Natural Science Foundation of China (61273087), the 111 Project (B12018), the Fundamental Research Funds for the Central Universities (JUSRP11459), the Program for Excellent Innovative Team of Jiangsu Higher Education Institutions, and the Fundamental Research Funds for the Central Universities (JUDCF12029).

#### References

- Z. Xiang, R. Wang, and Q. Chen, “Robust reliable stabilization of stochastic switched nonlinear systems under asynchronous switching,”
*Applied Mathematics and Computation*, vol. 217, no. 19, pp. 7725–7736, 2011. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - Z. Xiang, C. Qiao, and M. S. Mahmoud, “Finite-time analysis and ${H}_{\infty}$ control for switched stochastic systems,”
*Journal of the Franklin Institute*, vol. 349, no. 3, pp. 915–927, 2012. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - Z. Xiang, R. Wang, and Q. Chen, “Robust stabilization of uncertain stochastic switched nonlinear systems under asynchronous switching,”
*Proceedings of the IMechE, Part I: Journal of Systems and Control Engineering*, vol. 225, no. 1, pp. 8–20, 2011. - E. K. Boukas, “Static output feedback control for stochastic hybrid systems: LMI approach,”
*Automatica*, vol. 42, no. 1, pp. 183–188, 2006. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - H. Gao, J. Lam, S. Xu, and C. Wang, “Stabilization and ${H}_{\infty}$ control of two-dimensional Markovian jump systems,”
*IMA Journal of Mathematical Control and Information*, vol. 21, no. 4, pp. 377–392, 2004. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - S. He and F. Liu, “Robust peak-to-peak filtering for Markov jump systems,”
*Signal Processing*, vol. 90, no. 2, pp. 513–522, 2010. View at Publisher · View at Google Scholar · View at Scopus - S. He and F. Liu, “Finite-time ${H}_{\infty}$ control of nonlinear jump systems with time-delays via dynamic observer-based state feedback,”
*IEEE Transactions on Fuzzy Systems*, vol. 20, no. 4, pp. 605–614, 2012. - S. He and F. Liu, “Finite-time boundedness of uncertain time-delayed neural network with Markovian jumping parameters,”
*Neurocomputing*, vol. 103, no. 1, pp. 87–92, 2013. - Internet traffic report, 2008, http://www.internettracreport.com.
- L. Zhang and E.-K. Boukas, “${H}_{\infty}$ control for discrete-time Markovian jump linear systems with partly unknown transition probabilities,”
*International Journal of Robust and Nonlinear Control*, vol. 19, no. 8, pp. 868–883, 2009. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - J. Xiong, J. Lam, H. Gao, and D. W. C. Ho, “On robust stabilization of Markovian jump systems with uncertain switching probabilities,”
*Automatica*, vol. 41, no. 5, pp. 897–903, 2005. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - Y. Yin, P. Shi, and F. Liu, “Gain-scheduled robust fault detection on time-delay stochastic nonlinear systems,”
*IEEE Transactions on Industrial Electronics*, vol. 58, no. 10, pp. 4908–4916, 2011. View at Publisher · View at Google Scholar · View at Scopus - J. B. R. do Val and T. Basar, “Receding horizon control of Markov jump linear systems,” in
*Proceedings of the American Control Conference*, pp. 3195–3199, Albuquerque, NM, USA, June 1997. View at Scopus - B.-G. Park, J.-W. Lee, and W. H. Kwon, “Receding horizon control for linear discrete systems with jump parameters,” in
*Proceedings of the 36th IEEE Conference on Decision and Control*, pp. 3956–3957, San Diego, Calif, USA, December 1997. View at Scopus - D. Q. Mayne, J. B. Rawlings, C. V. Rao, and P. O. M. Scokaert, “Constrained model predictive control: stability and optimality,”
*Automatica*, vol. 36, no. 6, pp. 789–814, 2000. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - L. Liu, Z. Liu, and J. Zhang, “Nonlinear model predictive control with terminal invariant manifolds for stabilization of underactuated surface vessel,”
*Abstract and Applied Analysis*, vol. 47, no. 4, pp. 861–864, 2011. - J. A. De Doná, M. M. Seron, D. Q. Mayne, and G. C. Goodwin, “Enlarged terminal sets guaranteeing stability of receding horizon control,”
*Systems & Control Letters*, vol. 47, no. 1, pp. 57–63, 2002. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - R. S. C. Lambert, P. Rivotti, and E. N. Pistikopoulos, “A novel approximation technique for online and multi-parametric model predictive control,”
*Computer Aided Chemical Engineering*, vol. 29, pp. 739–742, 2011. View at Publisher · View at Google Scholar · View at Scopus - M. V. Kothare, V. Balakrishnan, and M. Morari, “Robust constrained model predictive control using linear matrix inequalities,”
*Automatica*, vol. 32, no. 10, pp. 1361–1379, 1996. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet - B. Pluymers, J. A. Rossiter, J. A. K. Suykens, and B. De Moor, “The efficient computation of polyhedral invariant sets for linear systems with polytopic uncertainty,” in
*Proceedings of the American Control Conference (ACC '05)*, pp. 804–809, June 2005. View at Scopus