Abstract
In this paper, a distributed predictive control with the model uncertainty which uses the data-driven strategy and robust theory (data-driven RDMPC) is proposed for the formation control of multiple mobile robots. The robust performance objective minimization is applied to replace the quadratic performance objective minimization to establish the optimization problem, where the model uncertainty is considered in the distributed system. The control policy is derived by applying the data-driven strategy, and the future predictive value is obtained by employing the linear law in the historical data. Lyapunov theory is referred to analyze the stability of the mobile robot formation system. The effectiveness of the proposed method is proved by a set of simulation experiments.
1. Introduction
In recent years, the robot formation control has become a key technology of multiple mobile robots’ research studies which can be used in the military field, production field, and service field [1–4]. In contrast with a single mobile robot, multiple mobile robots have the advantages of coordination and cooperation, which could improve the efficiency of completing tasks. The multiple mobile robots also enhance the fault tolerance and robustness of the system and strengthen the ability of robots’ environment recognition. Comparing with the leader-follower method [5–8], behavior-based method [9–11],and virtual structure method [12, 13], the model predictive control (MPC) has attracted attention because of its ability of improving the robustness of the system and having better dynamic control performance in solving the problem of consensus protocol in distributed formation control of multiple mobile robots. Also, MPC is convenient to establish the system model.
To further improve the performance of MPC, researchers investigate MPC combining with robust control [14, 15], adaptive control [16, 17], and sliding model control [18]. Among them, there are lots of works focusing on overcoming the uncertainty problem of formation control of multiple mobile robots. In the research of multiagent [19], a fault-tolerant controller was designed to compensate the influence of the system uncertainty. By proposing a static nonsmooth event-triggered protocol [20], the influence of the uncertainty is inhibited. Robust and adaptive control algorithms were used [21] for each agent to deal with the system uncertainty associated with leader input and follower agent dynamics. In the study of MAVs [22], adaptive control algorithms were employed locally for each vehicle to solve the uncertainty associated with the system and flying environment. Although these methods mentioned above have disposed of the uncertainty problem of the formation control with MPC, they are not combined with a distributed control. As for the MPC applied in the distributed control (distributed model predictive control (DMPC)), a cooperative DMPC algorithm was reported [23] where the relaxed constraints are used to find the optimization problem including the global optimization objective. This algorithm can be utilized to control the UAV tracking the ground target for real-time path planning [24] and multiple cooperative autonomous ships forming a multiship formation system [25]. To transport the plants and keep the formation at the same time [26], an objective function including trajectory tracking error and formation measurement error was adopted to meet the design requirements based on the DMPC. The DMPC is widely used in the research studies above, but the uncertainty is not considered. To the best of our knowledge, although lots of results have been available, the issue of simultaneous system uncertainty and DMPC has not been fully studied to date.
In practical application, the parameters in the model may not be accurate, and it is difficult to get an accurate model of system [27, 28]. Due to the uncertainty of the model, it is almost impossible to predict the future accurately, and the mismatch factors must be noticed. The model uncertainty of the system is ignored in the DMPC algorithm. In addition, the DMPC is too optimistic in the prediction process which would cause the prediction results deviating from the actual effect and lead to many restrictions on the decoupling of the distributed system [29, 30]. For this reason, it is necessary to study the DMPC with system uncertainty. A robust model predictive control approach [31] was suggested to achieve the path tracking of an omnidirectional robot with system uncertainty. The design of robust distributed MPC controller was casted [32] into solving a linear matrix inequality (LMI) optimization problem. A global performance was achieved by using an iterative algorithm. However, the iterative algorithm would cause the jitter fluctuation of the convergence curve and a low computational efficiency [33, 34]. A novel model-free adaptive SMC approach for unknown discrete nonlinear processes was designed [35] by bringing the merits of data-driven control strategy and SMC scheme together, which made the system with model uncertainty of strong robustness. A data-driven hierarchical control (DHC) structure was introduced [36] to ameliorate the performance of systems under the effect of the system uncertainty. The above references demonstrate both robust model predictive control and data-driven strategy which are effective methods to work out the problem of system uncertainty so that it is very important to study the data-driven robust distributed model predictive control with system uncertainty.
In this paper, a data-driven robust distributed model predictive control (data-driven RDMPC) with model uncertainty is proposed for multiple mobile robots. The main control method in this paper has these advantages. (1) The data-driven strategy is applied in this method using the historical data representing the future predictive value so that data processing is not limited to the model and the conservatism of system can be reduced. (2) The robust performance objective not the quadratic performance objective is chosen in this method to solve the optimization problem. As a result, the system can consider the model uncertainty, the fluctuation of the convergence curve can be decreased, and also the efficiency and accuracy of the algorithm can be promoted.
This paper is arranged as follows. The control problem is described in Section 2. In Section 3, a data-driven distributed model predictive controller is designed with considering the system uncertainty, and the stability of the system is analyzed. Section 4 presents a simulation example to illustrate the effectiveness of the method. Section 5 puts forward some conclusions.
2. Problem Formation
2.1. Notations and Graph Theory
Throughout the paper, the following notations will be used. The function means the predicted value of variables in time k. represents the variable x in current time and can be measured by sensors. And a quadric form function is given as where .
This paper examines the mobile robot formation control with the distributed control method. A finite index is set to mark the mobile robot, making each robot node as 1, 2,...,n. A directed and fixed graph G is . is a set of nodes, where is referred to the i-th mobile robot. is the edge set. indicates that the i-th robot can obtain the information from j-th robot. is the weighted adjacency matrix of G. If , we have while and .
2.2. Mathematical Model
The mobile robot i has on their path , i = 1, 2, …, n. The statement of the robot j communicating with the robot i is , and and are the position and orientation of robot i and j, respectively. The desired relative statement of robot i and j is where , , and represent the desired relative position and relative orientation. The kinematic models of i and j are established as follows:where and represent the control input of the forward and angular velocities of robot i and and represent the control input of the forward and angular velocities of robot j, and then the relative statement between them can be expressed as follows:
By taking the derivative of (3) and combining with (1) and (2), the following system is obtained:where is the desired angular speed error between robot i and robot j. System (4) between robot i and robot j can be linearized around as follows:where , , and are the statement and is the control input. Considering the communication between robot i and other robots, the linearized system of robot i is written as follows:where and , and the elements of are defined as ; therefore, , , and . It should be noted that is the sum of relative position errors between each mobile robot i and robot j and m is the number of robots communicating with robot i. After discretizing equation (6), the normative system of mobile robot i is as follows:where , T is the sampling period. Because of the uncertainty, the system will produce errors. Besides, the linearization process will also bring the errors to the system. To deal with these two errors concurrently, an uncertainty term related to control input and state variable is considered to describe the uncertainty of the system as follows:where and are known matrices. is an unknown but bounded time-independent matrix and satisfies . Then, considering the external disturbance of the mobile robot, the bounded disturbance term is added. The uncertain but bounded system model of mobile robot i is shown as follows:and the parameter of this uncertain system model is .
In the control process, the actuator usually bears limited control force. Therefore, it is very necessary to consider the saturation constraints of the input and the hard constraints of the state variables in the controller design [37]. Otherwise, the control will be inaccurate because of the control input and state variables exceed the limits, and this situation will inevitably reduce the tracking performance of the control system [38]. Therefore, saturation constraint (10) to limit the control input and hard constraint (11) to limit the state variable are added as follows:where and are the maximum of c-th element of and and is the c-th row of the assigned matrix .
3. Controller Design and Stability Analysis
3.1. Controller Design
It is worth emphasizing that each mobile robot has its own independent MPC controller. By communicating with each other, these robots receive the information which can form the coupling conditions to solve the optimal control problem online on their respective controllers. For each robot MPC controller, in the absence of an accurate model, the performance of solving the optimization problem which uses the method of quadratic performance objective minimization will be very poor, and the computational load will be large. So, the method applying the robust performance objective minimization to replace the quadratic performance objective minimization to establish the optimization problem is considered for the controller design. A feedback gain K is set to adjust and is continuously updated by the optimization problem. The optimization problem is shown as follows:
In order to propose an appropriate objective function, a trajectory parameter is needed to be included in the objective function of the normative system to construct the coupling relationship. A parameter is defined to describe the trajectory of mobile robot i, which is an iterative function related to control input, forward, and angular velocities:where and are the given parameter matrix. Defining as the desired value of the trajectory parameter difference between robot i and robot j, it is an important parameter to predict the trajectory of robot j through robot i and then the trajectory of robot j can be described as . The information which robots interact includes their position states and trajectory parameters. The position states and the trajectory parameters are used to calculate and form coupling conditions, respectively. The objective function of the normative system coupled with is as follows:where ,, , , , , , , , , , , and .
For the uncertain system (9), compared with , we focus on to control the system and get from the historical data without uncertain factors. should be limited to the feasible region of the optimization problem, so there is , and using the Schur complement, we can get the following:
If the ellipsoid Z is an invariant set, then the objective function can be divided into two parts as follows:
The optimization problem is transformed into solving a set of state feedback laws . In the case, the controller objective minimizes the upper bound of the worst-case objective function. Let and be the upper bound of objective function and , respectively.
Theorem 1. For the uncertain system (9), if there exists , then the system will be stable and Z is an invariant set when following conditions are satisfied:where , and represent the c-th row and c-th column of the matrix, and and represent the c-th element of the control input and state variable.
Proof. of Theorem 1. If is restricted in the ellipsoid Z and Z is an invariant set, then the objective function in infinite time domain can be regarded as two parts of (16), where is finite time domain. And the stability of the system can be analyzed by aiming at which is in infinite time domain. Consider a Lyapunov quadratic function , . For any time, and satisfy . According to the Lyapunov stability criterion, should meet the following inequality:Summing (21) from to and requiring or , the result is as follows:There is , then let . Substituting (9) and state feedback law into (21), we can get the following:where (23) can be regarded as two parts, and they are as follows:In the above formula, the feedback gain is defined as , and the solution of is given by the optimization problem. Substituting , feedback gain and into (23a) and (23b), and transforming (23a) and (23b) into matrix inequality form, the results are (19) and (20), respectively. Systems (19) and (20) guarantee the positive invariance of Z. In the invariant set Z, the saturation constraint (10) and hard constraint (11) are expressed as (19) and (20) in matrix form, respectively, and the control input is limited by , and the state variable is limited by .
Remark 1. In Theorem 1, equation (17) can get the constraint that the upper bound of is satisfied when the system is stable. After the upper bound of can also be expressed as some concrete constraints, the problem of minimizing the upper bound of the objective function in the worst case can be formulated and solved. And with the update of feedback gain, is always limited in the ellipsoid Z at any time in the optimization process.
Theorem 2. The upper bound of satisfies the following relation:where .
Proof. of Theorem 2. The objective function of is as follows:where can be measured by sensors and exists in historical data. is the function related to . can also be measured by sensors, and can interact with mobile robot j through wireless network or Bluetooth technology. Let , equation (24) is obtained after Schur complement.
Theorem 2 gives the condition of upper bound of ; then, the upper bound of the objective function in the infinite time in the worst case can be expressed as follows:The conditions about the upper bound of the objective function, feedback gain, and other parameters are given in the above theorems with the stable system, which can be used as the constraints of the optimization problem. Then, the optimization problem (12) can be described as LMI problem as follows:As mentioned in the paper above, we focus on not to control the system when solving the optimization problem. At this time, the difficulty is how to use the historical data from the previous time to represent the predicted value of the future time. So, consider the following linear law of historical data:where represents the coefficient matrix when the historical data are employed to indicate the future forecast value linearly. and denote storage units containing two sets of historical data of and , respectively. When , and . and are the remainders of the errors. Define as the storage unit including two sets of historical data of , but this unit has no direct connection with and , and the date of them exists independently. The system does not take the bounded external interference into consideration; after substituting (28) and (29) into the system model, the result is as follows:Substituting (30) into (15), we can getAfter expanding (31), set , and the result is shown as follows:Substituting (29) into (24), we can getInequalities (32) and (33) are obtained from historical data, and then the data-driven optimization problem (27) can be expressed as follows:
3.2. Stability Analysis
Lemma 1. When the system is inspired, the stored energy decreases with time and finally reaches a minimum, and then the system is gradually stable. If there is a positive definite scalar function with negative definite , the system is stable.
A positive definite Lyapunov function is defined in 3.1, where . If is negative definite, the system is asymptotically stable. The condition for to be negative definite is given in equation (21). By using the robust theory, the analysis of is transformed into the analysis of state feedback law , and the conditions (17) and (18) for state feedback gain are derived. According to Lemma 1, when (17) and (18) are held, the system will tend to stable.
4. Results
In this part, two simulations with a group of mobile robots are conducted to verify the proposed approach.
4.1. Simulation 1
In this simulation, the lane-change trajectory is considered. The desired trajectory is considered as follows:
Let The trajectory error of every two mobile robots can be expressed as and . The communication topology is shown in Figure 1. The parameters of robots are given as follows: , , , , , , , , , , , , , , and . Set the sampling period T = 0.1 s and coefficients ,, and . Each robot’s angle speed and line speed are and , respectively. Taking the uncertainty set as , the simulation is carried out for 50 s. The initial state values of each mobile robot are as follows:

The bounded external disturbance of each mobile robot is as follows:
The simulation is compared with the traditional DMPC. Figure 2 shows the state error of each mobile robot. Figure 3 illustrates the absolute value of state error of each mobile robot. Figure 4 displays and of each mobile robot. Figure 5 presents the different values between and of each mobile robot. Figure 6 demonstrates the trajectories of five mobile robots.

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

The result in Figure 2 shows that compared with the traditional DMPC, the optimization process of data-driven RDMPC uses the optimal solution instead of the iterative algorithm, which is more efficient and greatly reduces the fluctuation of the convergence curve. In Figure 3, the simulation results with the control target indicate that the simulation results of data-driven RDMPC are always smaller than those of the traditional DMPC in the process of gradual stabilization, which highlights the more accurate results of the data-driven RDMPC algorithm. From Figure 4, it is obvious that and almost coincide. To see errors, although the simulation results fluctuate, the different values representing of each mobile robot are always about as shown in Figure 5, and such a small error indicates that the parameter differences among mobile robots converge to the expected value between them. Figure 6 displays the trajectories of five robots, where the triangle, circle, and diamond represent the initial position, position at 30 s, and the position at 50 s of each mobile robot, respectively. It can be seen that five mobile robots start from their respective starting points and keep their formation shape in the process of advancing according to each ideal trajectory.
4.2. Simulation 2
In this simulation, we choose the sine function as the trajectory. The desired trajectory is considered as follows:
Let The trajectory error between every two mobile robots can be expressed as and . The communication topology is shown in Figure 1. The parameters and the bounded external disturbances of five mobile robots are the same as simulation 1. Taking the uncertainty set as , the simulation is carried out for 50 s. The initial state values of each mobile robot are follows:
Comparing with the traditional DMPC, Figure 7 reveals the state error of each mobile robot. Figure 8 shows the absolute value of state error of each mobile robot. The trajectory parameter difference and the desired trajectory parameter difference of each mobile robot are shown in Figure 9. Figure 10 indicates the different values between and of each mobile robot. Figure 11 displays the trajectories of five robots.

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

(a)

(b)

(c)

(d)

(e)

Through the simulation result, we can see that the fluctuation of the convergence curve using the data-driven RDMPC method is obviously been reduced than that using the traditional DMPC method in Figure 7. Taking the absolute value of , the result of data-driven RDMPC is more accurate than that of traditional DMPC as indicated in Figure 8. The simulation result of path parameter differences of each mobile robot shown in Figure 9 is very close to after analyzing the different values between them of each mobile robot in Figure 10. And the difference values are always about or , and we can think that tracks well. Figure 11 displays the trajectories of five robots, where the triangle, circle, and diamond represent the initial position, position at 30 s, and the position at 50 s of each mobile robot, respectively; these five mobile robots keep their formation shapes ideally.
5. Conclusions
The traditional distributed model predictive control has been widely used in the field of control; however, it is difficult to handle the model uncertainty of the system. This paper discusses a data-driven robust distributed model predictive control with the model uncertainty. The accuracy of the algorithm is improved by applying the robust performance objective minimization to replace the quadratic performance objective minimization when establishing the optimization problem with considering the model uncertainty of system. And on this basis, the fluctuation of convergence curve and the computational efficiency are abated due to the way of substituting for the iterative algorithm with optimal solution. The system conservatism is reduced after the data-driven strategy that the future predictive value is represented by employing the linear law of historical data which is put to use in deriving the control policy. The stability of the system is proved by Lyapunov theorem, and the effectiveness of this method is illustrated by a set of simulation examples.
Data Availability
The data used to support this study are available from the corresponding authors upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This research was funded by the Collaborative Innovation Center of Intelligent Green Manufacturing Technology and Equipment, grant number IGSD-2020-007; the Natural Science Foundation, grant number 5207053225; and the Department of Science and Technology of Hubei Province, grant number 2020BIB012.