Research Article | Open Access
Pablo S. Rivadeneira, Eduardo J. Adam, "Suboptimal Control Strategies for Finite-Time Nonlinear Processes with Input Constraints", Journal of Nonlinear Dynamics, vol. 2013, Article ID 182034, 11 pages, 2013. https://doi.org/10.1155/2013/182034
Suboptimal Control Strategies for Finite-Time Nonlinear Processes with Input Constraints
Novel techniques for the optimization and control of finite-time processes in real-time are pursued. These are developed in the framework of the Hamiltonian optimal control. Two methods are designed. The first one constructs the reference control trajectory as an approximation of the optimal control via the Riccati equations in an adaptive fashion based on the solutions of a set of partial differential equations called the and matrices. These allow calculating the Riccati gain for a range of the duration of the process and the final penalization . The second method introduces input constraints to the general optimization formulation. The notions of linear matrix inequalities allow us to recuperate the Riccati gain as in the first method, but using an infinite horizon optimization method. Finally, the performance of the proposed strategies is illustrated through numerical simulations applied to a batch reactor and a penicillin fed-batch reactor.
Batch processes have received much attention during the past two decades due to developing chemical and pharmaceutical products, new polymers, and recent biotechnological applications. Usually, the engineers state that a batch process has three operative stages: start-up, batch run, and shutdown. While these three stages are widely studied by the engineers for each particular batch process, it is important to note that in a wide number of cases, the batch run stage is far from an optimal operation, supported only with the experience of operators and engineers.
Control techniques have been made inroad in the industry in order to improve the performance of the process. In batch context, interesting ideas such as controllability, observability, and stability have been introduced despite the duration of batch processes. Numerous control techniques have been adapted to the peculiarities of these kinds of systems. One example is the model predictive control (MPC). This method is essentially numerical, usually implemented on-line. Most successful industrial applications of MPC reported so far are in refining and petrochemical plants (see for instance [1, 2]), where processes are run near optimal steady-states and model linearizations are reliable approximations. In spite of its many advantages, some undesired disadvantages (such as the high computational cost) have caused a major problem: only a few of the available commercial software packages are cautiously suggested for truly nonlinear batch processes as it is reported in .
Here, the optimal control paradigm is tackled, where the Hamiltonian formalism plays a major role. If the optimal control problem is regular, that is, a unique control minimizes the cost function of the problem, then there is an explicit form for the control law. For linear systems, it has been shown that this control strategy is in fact a proportional feedback law (see  for details). For nonlinear systems, the resulting control strategy is not a strict feedback control law depending on its state. Indeed, it is an open-loop strategy depending on a nonmeasurable co-state of Hamiltonian system associated with the process as it is illustrated in [4, 5].
This problem cannot generally be solved in real-time due to the missing initial condition for the costate. Several numerical approaches have been designed to cope with this problem, and many of them were based on the shooting methods (see ), recursive methods with generalized functions (see ), and techniques coming from artificial intelligence as the results presented by  and their references. The main disadvantage of these techniques, in the context of on-line applications, is the computational effort required. Therefore, these techniques remain as off-line procedures.
When input constraints are imposed to the problem, even in the linear control problem, the control law becomes nonlinear and phenomena like saturation appear. Many methods deal with this problem, for instance, control parametrization introduced by , neural networks (see  and their references), and other numerical approaches described by . However, few of them are designed for on-line applications. Another disadvantage presented in these methods can be the saturation of the actuators. For many batch processes, this situation is not desired.
MPC strategies and other techniques based on nonlinear programming schemes are proposed in the literature to solve both optimal problems in real-time (see for instance [11, 12]). However, the model process is converted to discrete time to pose the nonlinear programming problem and a large computation is required at each sampling time. In this paper, the continuous differential equations are maintained, analytical PDEs are used to calculate the control law, and lower computation requirement is needed in comparison with these methods. Recently, an analytical method to solve these problems was proposed by . This approach solves efficiently the optimal control problem but taking into account just the optimal objectives without considering the dynamic behavior of the solution. Here, the controllers obtained have a simple form (similar to PID controllers), and its dynamics can be adjusted to obtain smooth behaviors with a good performance. An interesting survey paper written by  can be seen for obtaining more details about the methods to solve both problems described before either in an off-line or in a real-time framework.
The contributions of this paper are as follows. (i) A method to generate control strategies that can be used to obtain reference trajectories (or batch recipes) by means of a linear approximation of the nonlinear optimal solution using auxiliary matrices , solutions of a set of partial differential equations (PDEs). The time-varying Riccati gain is calculated in parallel with the process based on , . (ii) A method to solve the constrained version of the optimization problem when restrictions in the manipulated variable are taken into account, exploiting the ideas of LMIs. A new LMI is proposed which can be designed to avoid saturation by tuning a parameter called . Although the control strategy is completely calculated and adapted for a finite-time horizon problem, the optimization problem is solved in an infinite horizon fashion. (iii) The controllers obtained in both cases have a simple form and can be implemented in real-time. Numerical procedures to implement the strategies are provided.
The rest of the paper has the following structure. In Section 2, a brief description of the Hamiltonian formalism is elaborated. Then the first method is proposed based on a set of PDEs and their relationships with the Riccati equations. Next, the restricted optimization problem is analyzed through the lines of linear matrix inequalities and optimal control. In Section 3, examples of batch processes are discussed. The strategies are reviewed by means of numerical simulations. Finally, in Section 4 the conclusions are presented.
2. Model Formulation and Solution Methodology
The dynamics of a fed-batch reactor usually is a nonlinear control system, which can be written in the affine form as follows: where is the vector of state variables with initial conditions , is the control vector, and and are smooth vector fields. The general dynamic optimization of a bioprocess needs to find the control vector , which minimizes a performance index as subject to the dynamics (1). The function is known as the Lagrangian, is a final penalization on the states, and is the duration of the batch run. Although and can be general functions, here only quadratics forms will be considered, namely, where and are symmetric and positive semidefinite matrices and is a positive-definite matrix. Under these conditions the Hamiltonian of such problem defined as is known to be regular; that is, there is a unique -optimal control that minimizes the Hamiltonian , and the derivative of with respect to vanishes at (). For control affine systems (1), the optimal control takes the following form: The control-Hamiltonian gives rise to the Hamiltonian canonical equations (HCEs) . Consider
2.1. Optimal Control for Linear Systems
In [4, 15] it is shown that the Hamiltonian canonical equations (HCEs) present several inconveniences when they are integrated. For instance, the adequate initial conditions of the costates are not known, and even if there are methods to obtain them, the HCEs are an unstable system. Besides, to construct the optimal trajectory with (6) for systems with many states can be difficult, and the optimal control problem becomes an unsolved problem. In contrast, to calculate the optimal control for a linear system is easier than the nonlinear case. The main idea is then to use the tools from linear systems to approximate the feed-forward control equation (6) to a linear feedback law as follows.(i)Only when the control action is calculated, the system considered will be a linear approximation of (1) around a small neighborhood of (, ), that is, where (, ) is a fixed point of the trajectory of the system (1) at time .(ii)The costates can be linearized in a small neighborhood of ; that is, . As a result, the optimal control law equation (6) becomes where is the solution to the differential Riccati equation (DRE): Notice that, in this paper, is the manipulated variable corresponding to nonlinear problem and is the approximated linear control law.
The main challenge here is either to solve the HCEs (8) corresponding to the linear problem (equivalently to find the missing conditions and ) or to calculate the solution of differential Riccati equation .
2.2. Method 1: Generating of Suboptimal Control Reference Trajectory
In order to find the missing boundary values and , the classical approach presented in [4, pages 370-371]) is revised. This method transforms the original boundary-value problem (8) into an initial-value one, by introducing the following auxiliary objects:(i)the Hamiltonian matrix , (which comes from (8) applied to the system (9)) (ii)the augmented Hamiltonian system (a linear -dimensional matrix ODE with a final condition) defined for the combination of two matrices through The solution to the augmented (linear) problem (13) being unique must verify and since in this linear case equation (8) also reads as follows: then the missing boundary conditions can be explicitly found, namely, (see  for a proof of the invertibility of and other details). Actually, the whole solution to the DRE can be recovered from the solution to problem (13), explicitly, as follows: As it will become clear in the following sections, it is desirable to have access to the missing boundary values for different values of the parameters without solving either the DRE or the HCEs. To solve the whole -family of LQR problems (with common values) consider each particular -problem: where the superscript refers to the individual problem of duration and final penalty matrix . Then (14) can be rewritten in the form Since the subjacent Hamiltonian system is linear, its solutions depend smoothly on parameters and initial conditions, and then the (partial) derivative of (22) with respect to (denoted by the subscript ) results in where the variables have been dropped for clarity here and whenever convenient and clear through the rest of the paper. At this point, it is convenient to start by adopting the simplest expressions for matrix to make sense of the partial derivatives with respect to the second parameter. Consequently, we will consider for simplicity that the form of the final penalty matrix will be scalar; that is, , . The partial derivatives of (22) with respect to the real variable are then Now, by partitioning into submatrices in the form then (24) reads as follows: which combined with (22) gives and by inserting these results in (23), the following relations are obtained: where These are first-order quasilinear PDEs for the matrices . Boundary conditions for a process of zero horizon are imposed in view of (21) and (17), that is, Now, the missing values, the final state and initial costate, for any -problem can be recuperated immediately as Concerning the existence and uniqueness of solutions to the quasilinear PDEs equation (28), there are only local results (see [16, page 51]), although the field of vector and matrix PDE integration is in active development (see for instance ). The solutions and illustrated before can be alternatively obtained from the procedure detailed by [18, 19] for the one- and -dimensional cases, respectively. The advantage of the procedure developed here is that it is not necessary to compute the PDEs for the final state and the initial coestate as it is required in [18, 19].
2.2.1. Relations amongst (, ) Solutions and Riccati Equations
For any -LQR problem the optimal initial costate has the form given by (10) where is the numerical solution of the corresponding DRE, that is, the final value matrix ODE in (11). Therefore, from (31) and (32), for each -problem the Riccati matrix should also verify allowing to solve the DRE as an initial-value problem (eventually on-line with the dynamics). But actually, as it is shown below (see (37)) the (, ) matrices for missing boundary conditions avoid solving the DRE for each particular -problem. They also avoid storing, necessarily as an approximation, the Riccati matrix for the values of chosen by the numerical integrator, possibly different from the time instants for which the control is constructed. Instead, the HCEs (8) can be integrated with initial conditions rendering the optimal trajectories for , which allow to generate the optimal control at each sampling time Even better, the feedback form for the control becomes directly available due to the linear dependence of (31) on initial conditions, namely since is also the optimal initial costate for the problem with parameters starting at . Then, as a side-product, an alternative formula for the Riccati matrix results:
2.2.2. Numerical Method for Nonlinear Systems
Now, a suboptimal control strategy for nonlinear systems can be developed based on the linearization (9) around a chosen pair . After that, the DRE is solved by the matrices , according to (37) for a range of and , and its solution is kept in memory.
Normally, this linearization is performed around an equilibrium point (or an steady-state point). And, this will work well near that point. However, the initial condition is far from the equilibrium and in fed-batch rectors equilibrium (or steady state) does not even exist as in continuous reactors. Here, the linearization will be calculated around the initial condition, so updating it in the running time is recommended.
In the following procedure, a basic adaptive scheme to update the controller gain as well as the matrices , at arbitrary time instants , is given:In: , , , , , ,,Out: , for all . Step 1: Take . At time , the first pair is computed from (9) taking and or as solution of . Step 2: Compute using (22) or as a solution of (28) by mean of the two matrices , and Step 3: run in parallel the process sending the control signal = during the interval . Step 4: at time instant , update the matrices , , , and of the control signal of Step 3 taking and . Update . Step 5: return to Step 3. This procedure continues until completing the whole process horizontally, that is, until to reach . Step 6: if , then the process is over.
Notice that describes a time-variant feedback control law, where the matrices , , , and are updated at each instant . In the scheme above, the time instants () are chosen arbitrarily. More sophisticated algorithms to detect the time of updating the linearization can be found in the specialized literature of adaptive control, but that is out of scope of this paper.
2.3. Method 2: Handling Input Constraints
The goal in this section is to develop a suboptimal feedback control for the nonlinear systems (1) but taking into account restrictions in the manipulated variable. This last objective cannot be included in Method 1 due to the lack of efficient tools for on-line optimization. However, in the last years, numerical approaches (as neural networks (see )) have been introduced as alternative tools to cope with this problem in the off-line context.
Here, an on-line procedure to generate suboptimal feedback control laws is developed. To incorporate restrictions, the manner to solve the optimization problem was changed, basically modifying its horizon to infinite (see ). So, the optimization problem to solve reads as follows: subject to dynamics (9) and
Note that the optimization horizon time was changed to infinity only to evaluate the cost function and to generate the control law, but the problem remains as a finite-time process. The finite-time horizon with input constraints case and on-line applications is still an open problem. Some first results in this context are illustrated in .
The solution to the optimal problem control described above is not easy to find, even when the system is linear. For these kinds of systems, the traditional Ricatti equations are no longer valid, and then a number of numerical approaches attempt to solve this problem like the MPC formulation  or the methods described in .
Linear matrix inequalities (LMIs) are an alternative method to treat the optimization problem posed above with a low computational effort, at least in the infinite horizon set-up and for linear time-invariant systems.
A linear matrix inequality or LMI is a matrix inequality of the form: where and are the optimization variables and the positive and symmetric matrices , , are given. The LMI (40) is a convex constraint on ; that is, the set is convex (for more details, it is suggested to see two interesting contributions [21, 22]).
In particular, the following control problems are handled as a single LMI-based optimization problem in this paper.
(i) If is considered as an infinite horizon, it is known that when the control weighting matrix in the cost functional is positive definite, the state weighting matrix is nonnegative definite and the pair is controllable; then the LQR problem is well posed and can be solved via the classic algebraic Riccati equation (ARE) (see  for more details). This algebraic equation can be converted to a constraint (inequality)  as follows: where is the solution of the last inequality and is a positive symmetric matrix. In this section, the Riccati gain used to generate the control action is renamed as because it is the solution of the ARE equation.
A common procedure to convert convex nonlinear (matrix norm, quadratic, etc.) inequalities to a LMI form is through the well-known Schur complements; that is, a LMI of the form where , , and depend on , is equivalent to the matrix inequality or equivalently one has
Note that to minimize the optimal control problem without constraints, it is equal to solve the ARE equation. And in  it is proved that to find the ARE’s solution is equivalent to minimize the LMI equation (45).
(ii) Physical limitations inherent in process equipment invariably impose hard constraints on the manipulated variable as in (39). These constraints can be seen as a LMI.
Considering that is a stabilizing control law, then one has the following.
Lemma 1. Assuming that a parameter exists, the following LMI is equivalent to
Proof. Consider that Since is a symmetric and square constant matrix, it is possible to decompose it into its singular values, and then the following inequality holds: where is a singular value; then Using (44), results equivalently to where is a real positive constant that satisfies a stabilization condition for any .
Note that to find , it is necessary to involve another procedure to design an invariant ellipsoid to guarantee either or , where is a positive definite matrix. But it requires a higher computational effort. However, a simple criterion can be used to approximate the value; that is, since normally the linear system involved is a dissipative system to open loop, then as it is suggested in .
One important advantage of the LMI is that the problem with multiple constrains can be expressed with multiple LMIs; that is, as a single LMI as follows: which makes us introduce the input constraints to the optimal problem as
The control action that minimizes this problem can be expressed as where is the solution of minimizing of the LMI (53).
The procedure to implement these results in nonlinear systems is as follows.In: , , , , , ,.Out: , for all . Step 1: Take . At time , the first pair is obtained either from (9) taking and or as solution of . Step 2: Compute the LMI (53) by mean of the two matrices , Step 3: run in parallel the process sending the control signal during the interval . Step 4: at time instant , update the matrices , , and of the control signal of Step 3 taking and . Take . Step 5: return to Step 3. This procedure continues until completing the whole optimization horizon, that is, until to reach . Step 6: if , then the optimization process is over.
The time instants when the matrices are updated are taken in the same fashion as Method 1. The numerical methods to solve LMI are well-known and they are already available in a standard software like MATLAB.
3. Numerical Examples
3.1. Nonlinear Batch Reactor
Consider a batch reactor with a nonlinear dynamic where an exothermic and irreversible second order chemical reaction takes place. It is assumed that the reactor has a cooling jacket whose temperature () can be directly manipulated. The goal is to regulate the reactor temperature () by means of inlet coolant temperature (). Furthermore, the manipulated variable has minimum and maximum constraints . The concentration is denoted by the state .
Initially, the batch reactor is on the nominal values, and the goal is to control the reaction temperature without restrictions in . The desired value for the temperature of the reactor is . For this objective, Method 1 is applied to the process. Two scenarios were simulated. In the first one, the linearization was not updated, and the control system works with the linearization around the initial condition. In the second one, it was updated at each . The optimal solution was computed by methods described in . The final penalization was chosen as . The results are illustrated in Figures 1, 2, and 3.
Dynamically, the optimal solution has a worse behavior than that of the suboptimal trajectories because its state and control have more variability than the state and control of the suboptimal trajectories. However, the final error between the final desired state and the reached state is lower for the optimal solution as well as its total cost. Note that the solution with is nearly arrived at the final desired state ().
The relative error between the optimal cost () and the suboptimal cost () was calculated as . When the linearization is updated with more frequency (for instance ), the error decreases as it is illustrated in Figure 3. For both updating schemes of linearization, the error is smaller than 3%. That means that Method 1 approximates appropriately the optimal cost of the nonlinear system.
When the input constraints are active, two control strategies are proposed. A traditional predictive controller (MPC) according to [12, 26] and a suboptimal controller following Method 2 are proposed here. The suboptimal controller can incorporate the possibility of satisfy restrictions in the manipulated variable handling saturation using the tuning of the parameter . Two -values are selected (). Furthermore, the linearization is updated at each and the time operation for the reactor is .
Notice that neither the MPC nor the suboptimal controller with can completely reject that the initial disturbance due to the saturation of the manipulated variable. This phenomenon is shown in Figure 4. When the reaction begins, both control schemes try to correct the increase of temperature in the reactor, but quickly the manipulated variable is saturated. However, when the saturation of the manipulated variable is avoided.
Note also that the saturation time with the suboptimal controller is less than that with the MPC, and this could be modified by setting the parameter . This is certainly an advantage in systems where it may result beneficially in an initial saturation.
Figure 5 shows the dynamic responses of the reactor temperature for the three cases here studied. Even though the performance is similar in all three cases, the functional objective was evaluated and a worse performance than the MPC was obtained. Anyway, the improvement in the functional is approximately 2% with the method proposed here.
3.2. Case Study: Penicillin Fed-Batch Reactor
In this model ,, , and are the biomass, penicillin concentration, substrate concentration (g/L), and the reactor volume (L), respectively. The control variable is the feed rate of substrate.
The goals for the control of the penicillin reactor were extracted from , and they are as follows.(i)To reach in hours ( is the time horizon) a final penicillin concentration of 8 g/L; that is, the desired output is .(ii)The batch process includes an input constraint, which must be satisfied during the operation process The values of the parameters of the cost function are , and .
3.2.1. Results of Method 1
The strategy starts by computing the linearization of the system given by (1); that is, and are initially calculated at , namely,
The two equations in (28) were calculated for this approximation and the Riccati gain is now available. Four time instants were selected: , , , and, . In these time instants the controller gain was updated.
In Figure 6, the evolution of the states for the finite horizon optimization problem and the control reference trajectory are depicted. The batch process has a typical response (according to ), but for this set-up there is an error between the desired state and the real state reached, , at the end of the operation time of the batch reactor. So, the parameter is increased and the results are shown in Figure 7. Clearly, the goals are now satisfied. Note that these strategies accept higher control values than . In other words, the constraint imposed in the manipulated variable is not satisfied yet.
3.2.2. Results of Method 2
In this subsection, the results of Method 2 are described. The optimization problem via LMI is solved by traditional numerical methods (see ) and implemented with standard software like MATLAB, as well as the first strategy.
The controller gain was updated four times according to the time instants from the first method solving the LMI (53) at each time.
The parameter in (46) was taken as  suggests and that ensures the validity of the assumption of Lemma 1. The strategy for the second method was simulated, and in Figure 8 the results are depicted. The reader can notice that the objectives fixed for the problem are satisfied; that is, the output is near the desired value and the manipulated variable satisfies the constraint avoiding saturation.
In this paper, complete control schemes have been developed and illustrated. The innovation in this paper is the novel way to obtain the control reference trajectory (or recipe), which is necessary in the context of batch reactors. These kinds of trajectories are constructed by using a set of new algebraic equations which are described above for the linear case. The other contribution is to handle constraints in the manipulated variable when the process is controlled and optimized. All tools developed to design come from modern optimal control theory based on the Hamiltonian formalism.
It is important to remark that the entire proposed schemes are designed to be implemented in real-time, with the controller parameters being updated on-line. This is because the resulting controller is based on the simplifications made of costate on the system, and, in consequence, a simple controller can be online-tuned. Thus, the two resulting strategies for tuning simple controllers are able to generate suboptimal trajectories such that, in the second case, different constraints can be satisfied.
Also, it is significant to comment that the first strategy was developed using finite horizon optimal control theory with the restrictions that support the definition of the objective function. Since the control action written in costate terms and assuming linearity in the costates, the resulting control law has time-varying gains. On the other hand, the second strategy is presented under infinite horizon optimal control framework and allows the inclusion of new constraints by means of LMI as for example maximum and minimum constraints in control variable. Maintaining the linearity in the co-states, it is possible to find a suboptimal control law, but with constant feedback gains. At this point, notice that the simplifications may lead to a suboptimal solution to the optimization problem. The counterpart of this is that the reduction in the complexity of the mathematical problem treatment leads to less computational effort.
Finally, the numerical results show a good performance in the controlled variables, with the suboptimal strategies, and, especially, in the second strategy, when the control is bounded. In general, all variables have a good performance inside the operation time fixed for the batch reactor achieving the requirements imposed upon the system.
- E. F. Camacho and C. Bordons, Model Predictive Control, Springer, London, UK, 2nd edition, 2004.
- B. W. Bequette, “Non-linear model predictive control: a personal retrospective,” Canadian Journal of Chemical Engineering, vol. 85, no. 4, pp. 408–415, 2007.
- S. J. Qin and T. A. Badgwell, “A survey of industrial model predictive control technology,” Control Engineering Practice, vol. 11, no. 7, pp. 733–764, 2003.
- E. D. Sontag, Mathematical Control Theory. Deterministic Finite Dimensional System, Springer, 1998.
- V. Costanza and P. S. Rivadeneira, “Minimal-power control of hydrogen evolution reactions,” Optimal Control Applications and Methods, vol. 31, no. 2, pp. 105–115, 2010.
- G. Fraser-Andrews, “A multiple-shooting technique for optimal control,” Journal of Optimization Theory and Applications, vol. 102, no. 2, pp. 299–313, 1999.
- C. J. Goh and K. L. Teo, “Control parametrization: a unified approach to optimal control problems with general constraints,” Automatica, vol. 24, no. 1, pp. 3–18, 1988.
- M. Abu-Khalaf and F. L. Lewis, “Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach,” Automatica, vol. 41, no. 5, pp. 779–791, 2005.
- T. Cheng, F. L. Lewis, and M. Abu-Khalaf, “Fixed-final-time-constrained optimal control of nonlinear systems using neural network HJB approach,” IEEE Transactions on Neural Networks, vol. 18, no. 6, pp. 1725–1737, 2007.
- H. H. Mehne and A. H. Borzabadi, “A numerical method for solving optimal control problems using state parametrization,” Numerical Algorithms, vol. 42, no. 2, pp. 165–169, 2006.
- C. Büskens and H. Maurer, “SQP-methods for solving optimal control problems with control and state constraints: adjoint variables, sensitivity analysis and real-time control,” Journal of Computational and Applied Mathematics, vol. 120, no. 1-2, pp. 85–108, 2000.
- H. A. Gonzalez, E. J. Adam, D. Odloak, and J. L. Marchetti, “Infinite horizon mpc applied to batch processes—part I,” in Proceedings of the 13th Reunión de Trabajo en Procesamiento de la Información y Control (RPIC'09), Rosario, Argentina, September 2009.
- M. Matinfar and M. Saeidy, “A new analytical method for solving a class of nonlinear optimal control problems,” Optimal Control Applications and Methods, 2013.
- A. Rao, “A survey of numerical methods for optimal control,” Advances in the Astronautical Sciences, vol. 135, pp. 497–528, 2009.
- V. Costanza, “Regular optimal control problems with quadratic final penalties,” Revista de la Unión Matemática Argentina, vol. 49, pp. 43–56, 2008.
- G. B. Folland, Introduction to Partial Differential Equations, Princeton University Press, Princeton, NJ, USA, 2nd edition, 1995.
- A. I. Zenchuk and P. M. Santini, “Dressing method based on the homogeneous Fredholm equation: quasilinear PDEs in multidimensions,” Journal of Physics A, vol. 40, no. 23, article 010, pp. 6147–6174, 2007.
- V. Costanza and P. S. Rivadeneira, “Finite-horizon dynamic optimization of nonlinear systems in real time,” Automatica, vol. 44, no. 9, pp. 2427–2434, 2008.
- V. Costanza, P. S. Rivadeneira, and R. D. Spies, “Equations for the missing boundary values in the hamiltonian formulation of optimal control problems,” Journal of Optimization Theory and Applications, vol. 149, no. 1, pp. 26–46, 2011.
- V. Costanza and P. S. Rivadeneira, “Optimal saturated feedback laws for lqr problems with bounded controls,” Computational and Applied Mathamatics, vol. 43, pp. 111–123, 2013.
- S. Boyd, L. El Ghaoui, E. Feron, and V. Balakrishnan, Linear Matrix Inequalities in System and Control Theory, Society for Industrial and Applied Mathematics, 1994.
- M. V. Kothare, V. Balakrishnan, and M. Morari, “Robust constrained model predictive control using linear matrix inequalities,” Automatica, vol. 32, no. 10, pp. 1361–1379, 1996.
- C. A. Cappelletti and E. J. Adam, “Diseño de un control de nivel de un sistema hidráulico con restricciones utilizando lmi,” in Proceedings of the 21st Congreso Argentino de Control Automático (AADECA'08), Buenos Aires, Argentina, September 2008.
- K. S. Lee and J. H. Lee, “Model predictive control for nonlinear batch processes with asymptotically perfect tracking,” Computers and Chemical Engineering, vol. 21, supplement 1, pp. S873–S879, 1997.
- E. J. Adam, “. Adaptive iterative learning control applied to nonlinear batch reactor,” in Proceedings of the 13th Reunión de Trabajo en Procesamiento de la Información y Control (RPIC'07), Río Gallegos, Argentina, October 2007.
- J. M. Maciejowski, Predictive Control with Constraints, Prentice Hall, 2002.
- J. E. Cuthrell and L. T. Biegler, “Simultaneous optimization and solution methods for batch reactor control profiles,” Computers and Chemical Engineering, vol. 13, no. 1-2, pp. 49–62, 1989.
- J. R. Banga, E. Balsa-Canto, C. G. Moles, and A. A. Alonso, “Dynamic optimization of bioprocesses: efficient and robust numerical strategies,” Journal of Biotechnology, vol. 117, no. 4, pp. 407–419, 2005.
Copyright © 2013 Pablo S. Rivadeneira and Eduardo J. Adam. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.