Journal of Applied Mathematics

Volume 2012, Article ID 530139, 12 pages

http://dx.doi.org/10.1155/2012/530139

## A Modified NM-PSO Method for Parameter Estimation Problems of Models

^{1}Department of Computer Science and Information Engineering, St. John’s University, No. 499, Section 4, Tam King Road, Tamsui District, New Taipei City, 25135, Taiwan^{2}Graduate Institute of Computer and Communication Engineering, National Taipei University of Technology, No. 1, Section 3, Chung-hsiao E. Road, Taipei 10608, Taiwan^{3}Department of Marketing and Logistics Management, St. John’s University, No. 499, Section 4, Tam King Road, Tamsui District, New Taipei City 25135, Taiwan^{4}Department of Electrical Engineering, St. John’s University, No. 499, Section 4, Tam King Road, Tamsui District, New Taipei City 25135, Taiwan

Received 9 June 2012; Accepted 9 September 2012

Academic Editor: Alain Miranville

Copyright © 2012 An Liu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

Ordinary differential equations usefully describe the behavior of a wide range of dynamic physical systems. The particle swarm optimization (PSO) method has been considered an effective tool for solving the engineering optimization problems for ordinary differential equations. This paper proposes a modified hybrid Nelder-Mead simplex search and particle swarm optimization (M-NM-PSO) method for solving parameter estimation problems. The M-NM-PSO method improves the efficiency of the PSO method and the conventional NM-PSO method by rapid convergence and better objective function value. Studies are made for three well-known cases, and the solutions of the M-NM-PSO method are compared with those by other methods published in the literature. The results demonstrate that the proposed M-NM-PSO method yields better estimation results than those obtained by the genetic algorithm, the modified genetic algorithm (real-coded GA (RCGA)), the conventional particle swarm optimization (PSO) method, and the conventional NM-PSO method.

#### 1. Introduction

The parameter estimation problems involve estimating the unknown parameters of the mathematical models based on a system of ordinary differential equations by using experiment data that are obtained under well-defined standard conditions. Traditional optimization methods such as the Nelder-Mead (NM) method [1, 2] and the Gauss-Newton method [3] can be applied to find reasonably good estimations of parameters of simple models. However, they are not robust enough for complex problems that involve a huge search space, and they tend to find local optimum points rather than the global optimum points. In addition, quasi-linearization methods and data-smoothing methods are also often used to solve parameter estimation problems [4].

To overcome the problem of finding the global optimum points, several heuristic optimization methods such as the genetic algorithm (GA) [5], the simulated annealing (SA) method, and the particle swarm optimization (PSO) method [6] for solving the parameter estimation problems have been proposed. Some modifications to the heuristic optimization methods have also been proposed in recent years. Khalik et al. proposed the real-coded GA (RCGA) method for parameter estimation to overcome the drawbacks of the binary-coded GA (BCGA) method [7, 8]. Ali et al. proposed the application of a modified differential evolution (MDE) method [9]. Schwaab et al. proved that less computational attempts are needed by the PSO method than the GA method and the SA method for solving parameter estimation problems [10]. Zahara and Liu [6] applied the PSO method and the conventional NM-PSO method to solve parameter estimation problems, demonstrated the advantages of the conventional NM-PSO method, and showed that it is an effective tool in solving unconstrained or constrained optimization problems.

The advantages of the heuristic methods are that they do not require information about the gradient of the objective function [11–13], that they are insensitive to the guessed solutions, and that they can find the global solutions by making extensive calculations of the objective function in the parameter space.

In this research, a modified hybrid Nelder-Mead simplex search and particle swarm optimization (M-NM-PSO) method is proposed to solve the parameter estimation problems. The proposed M-NM-PSO method is applied to three well-known cases, and the results obtained are compared to those obtained by the GA, RCGA, MDE, PSO, and conventional NM-PSO methods to demonstrate its superiority in terms of accuracy, rate of convergence, and feasibility.

The content of this paper is organized as follows. Section 2 describes briefly the parameter estimation problems. Section 3 presents the proposed M-NM-PSO algorithm. Section 4 discusses the numerical simulation cases and compares the results obtained by different methods, and the conclusions are summarized in Section 5.

#### 2. The Parameter Identification Problems

Assume that the mathematical model is defined either by a first-order differential equation or a second-order differential equation where is the parameter vector with unknown real parameters .

The experiment data are , , where are the independent time variables and are the experiment data or measured values of the corresponding dependent variables. Typically, we have . The problem considered herein is that of estimating the optimal parameter vector as accurately as possible from the given experiment data. This is a problem of minimizing the sum of square errors (SSE), which can be represented as where are obtained by the Runga-Kutta method and are the experiment data.

Three cases are analyzed in this study to validate the superiority of the proposed M-NM-PSO method. The first two cases are standard minimal SSE problems, and the third case is a more complex seven-dimensional isothermal continuously stirred tank reactor (CSTR) maximal SSE problem.

#### 3. The Proposed M-NM-PSO Method

##### 3.1. The Nelder-Mead (NM) Simplex Search Method

The NM simplex search method is proposed by Nelder and Mead [2], which is a good local search method designed for unconstrained optimization problem without using gradient information. This method rescales the simplex by four basic linear transformations: reflection, expansion, contraction, and shrinkage. Through these transformations, the simplex can successively improve itself towards the optimum point.

Take the problem of finding the simplex 3 solutions in a 2-dimensional search space for example. The basic NM procedures to minimize the two-variable function is illustrated in Figure 1. The NM simplex design begins with a starting point and initial step sizes to construct points and as shown in Figure 1. Suppose that is the highest (worst) of all function values and the point is to be replaced by the point . In this example, is the reflection of to the centroid point between and . Suppose . At this stage, two situations may arise.

*Case 1 (). *An extension point, point , is created. Point is replaced by point if , otherwise point is replaced by point .

*Case 2 (). *A contraction point, point , is created. if , otherwise .

Point is replaced by point if , otherwise a shrinkage operation is performed to reduce the size of the simplex by moving point and point towards point .

The advantage of the NM method is that it is intrinsically fast in finding an optimal solution, but the disadvantage is that the solution found may be a local optimal solution rather than a global one. We want to retain its advantages but not its shortcomings in our proposed M-MN-PSO algorithm.

##### 3.2. The Particle Swarm Optimization (PSO) Algorithm

Eberhart and Kennedy [14] were the first to propose the PSO algorithm. As shown in Figure 2, it begins by randomly initializing a flock of birds over the problem space where each bird is called a “particle”.

Each particle remembers the best solution which it has found and the best solution found by the entire swarm along the search trajectory. Their velocities and positions are updated by the following equations: where and are two acceleration constants called the cognitive parameter and the social parameter, respectively, and are typically set to 2.0. The function rand() generates uniformly a random value in the range . The parameter is an inertia weight. Eberhart and Shi [15] suggested that .

Equation (3.1) yields the new velocity of a particle which is determined by the particle’s previous velocity (), its best position (), and the global best position (). It is necessary to impose a maximum limit value on the velocity. If the computed new velocity exceeds this threshold, it is set to to prevent this particle from flying past the desired solutions in each iteration. Equation (3.2) specifies how each particle’s position is updated in the search space based on their movement over a time interval , which is usually set to 1.

The advantage of the PSO algorithm is that it tends to find the global solution rather than the local one, but improvements on its accuracy and speed of convergence are much desired.

Zahara proposed a hybrid NM-PSO method that combines the NM method and the PSO method, and applied this method to two study cases with excellent results [6]. This paper describes a modified version of the hybrid NM-PSO method, the M-NM-PSO method, with even better results.

##### 3.3. The Proposed M-NM-PSO Method

Two algorithms are integrated in the conventional hybrid NM-PSO optimization method: a conventional algorithm (the NM simplex search algorithm) and an evolutionary algorithm (the PSO algorithm). The efficiency of the NM simplex search algorithm is high because it converges rapidly, but it tends to converge to a local rather than a global optimal solution. On the other hand, the PSO algorithm is capable of finding a global optimal solution, but a large size of particle population and thus great amounts of memory storage and computation time are required during the optimization process.

Based on the above reasoning, the conventional hybrid NM-PSO method was proposed to overcome the shortcomings of the PSO algorithm and the NM algorithm, and to find the global optimal solution accurately and efficiently.

The conventional NM-PSO method was developed by Zahara and Liu [6]. In this method, optimal particles are reserved, and the NM operator is applied to the first particles and to update the particle. While the conventional NM-PSO method updates only the remaining particles, the proposed M-NM-PSO method updates all the N particles and thus converges towards the optimal solution more accurately and faster, and increases the possibility of finding a better solution. Figure 3 shows the schematic representation of the proposed M-NM-PSO method.

The procedures to implement the proposed M-NM-PSO method are as follows, and the pseudo codes of the proposed M-NM-PSO algorithm are shown in Pseudocode 1. Let the dimension of the problem to be solved is . First, particles are generated as a swarm. Next, the objective functions are arranged in the order from good to bad, and the particles are divided accordingly into three groups: the first particles, the particle, and the remaining particles. Then, the function values of the first particles and the particle are calculated using the NM simplex method to find the updated best particle. After the PSO method examines the positions of the best particles and readjusts the particles, the global optimal particle of the population is determined by the sorted fitness values. The above optimization procedures are repeated until the termination conditions are satisfied.

#### 4. Numerical Simulations and Comparisons

In this paper, the proposed M-NM-PSO method is applied to solve three well-known problems [4–9]. The results obtained are compared with those in the cited papers [4–9]. To demonstrate the superiority of the proposed M-NM-PSO method, only a population of 21 particles is used to find the solutions of the three cases. The method is implemented with Matlab, and the programs are run on a PC with a 3.2 GHz dual-core Intel processor and 4 GB memory capacity.

*Case 1 (an enzyme effusion problem). *The mathematical model of an enzyme effusion problem can be represented as

The experiment data are listed in Table 1, and initial conditions are used by the M-NM-PSO method to solve (4.1) and to estimate the values of the four parameters , , , and in the model.

The results obtained by the M-NM-PSO method using a population of just 21 particles along with those by Scitovski and Jukić [4], GA [5], RGA [7, 8], PSO [6], MDE [9], and conventional NM-PSO [6] are listed in Table 2.

The fact that the results of the M-NM-PSO method are reached after only 150 iterations with an SSE value of 3963.0 validates the superiority of the M-NM-PSO method. The results in Table 2 also show that the M-NM-PSO method yields better estimates and has smaller SSE than those of GA, RGA, PSO, and conventional NM-PSO. Figure 4 shows the plots of the estimated and measured data of , which demonstrate the excellent fitness of the estimated data to the measured data. Figure 5 shows the plots of the estimated data of .

*Case 2. *A mathematical model may be represented by a second-order ordinary differential equation (ODE)

The values of the parameters , , , and in (4.2) are to be estimated from the data in Table 3, and the results are listed in Table 4. Note that given the estimated values of the parameters , , , and , the SSE values of the RGA [7] should be 0.4315 rather than 0.3204, and 0.3969 rather than 0.2827, as listed in Table 4.

As expected, the M-NM-PSO method yields better estimated results and lower SSE values than those of the GA, RGA, PSO, and conventional NM-PSO methods. It yields the same results as the NM-PSO method but with less iteration. Figure 6 shows the plots of the estimated and measured data of in the given range, which demonstrates the excellent fitness of the estimated data to the measured data.

*Case 3 (an isothermal continuously stirred tank reactor (CSTR) problem). *A seven-dimensional isothermal CSTR problem [7, 16] is represented by the equations
where

This case is different from the previous two cases. The problem is to find the optimal parameter values , , , and that maximize the performance index (PI) or :

A first-order differential equation is used to find the optimal parameter vector .

The initial starting point **x**(0) is given by the vector , , and the bounds of the unknown parameter , , , are

This seven-dimensional system has a PI function with four control parameters and appears to be a difficult task for the proposed M-NM-PSO method. The results with fixed at 6.0 are shown in Table 5 for comparison. Compared with the RGA [7] method, the M-NM-PSO method uses a smaller particle population and yields better estimates and a larger PI. Compared with the conventional NM-PSO method, M-NM-PSO yields the same results in less iteration.

The M-NM-PSO method can still be applied using 21 particles even if the value of is not fixed, and the results are reached in 53 iterations: , , which are also shown in Table 5. As expected, the M-NM-PSO method not only yields better results but also requires less iteration than the other methods. The plots of the seven states () are shown in Figure 7.

#### 5. Conclusion

All of the results of the three cases indicate that the proposed M-NM-PSO method can be applied efficiently to solve the estimation problems of unknown parameters in mathematical models. The application of the proposed M-NM-PSO method is demonstrated by three study cases. The results indicate that the proposed M-NM-PSO method is indeed more accurate, reliable, and efficient in finding global optimal solutions than the other alternative algorithms or methods. Furthermore, the proposed M-NM-PSO method converges accurately as well as quickly, thus greatly improves the efficiency of solving the parameter estimation problems.

#### References

- J. E. Dennis Jr. and V. Torczon, “Direct search methods on parallel machines,”
*SIAM Journal on Optimization*, vol. 1, no. 4, pp. 448–474, 1991. View at Publisher · View at Google Scholar · View at Zentralblatt MATH - J. A. Nelder and R. Mead, “A simplex method for function minimization,”
*Computer Journal*, vol. 7, pp. 308–313, 1965. View at Google Scholar - N. Yildirim, F. Akcay, H. Okur, and D. Yildirim, “Parameter estimation of nonlinear models in biochemistry: a comparative study on optimization methods,”
*Applied Mathematics and Computation*, vol. 140, no. 1, pp. 29–36, 2003. View at Google Scholar - R. Scitovski and D. Jukić, “A method for solving the parameter identification problem for ordinary differential equations of the second order,”
*Applied Mathematics and Computation*, vol. 74, no. 2-3, pp. 273–291, 1996. View at Publisher · View at Google Scholar · View at Zentralblatt MATH - E. K. Nyarko and R. Scitovski, “Solving the parameter identification problem of mathematical models using genetic algorithms,”
*Applied Mathematics and Computation*, vol. 153, no. 3, pp. 651–658, 2004. View at Publisher · View at Google Scholar · View at Zentralblatt MATH - E. Zahara and A. Liu, “Solving parameter identification problem by hybrid particle swarm optimization,” in
*Proceedings of the International MultiConference of Engineers and Computer Scientists (IMECS'10)*, pp. 36–38, Hong Kong, China, March 2010. - M. A. Khalik, M. Sherif, S. Saraya, and F. Areed, “Parameter identification problem: real-coded GA approach,”
*Applied Mathematics and Computation*, vol. 187, no. 2, pp. 1495–1501, 2007. View at Google Scholar - N. Tutkun, “Parameter estimation in mathematical models using the real coded genetic algorithms,”
*Expert Systems with Applications*, vol. 36, no. 2, pp. 3342–3345, 2009. View at Google Scholar - M. Ali, M. Pant, A. Abraham, and V. Snasel, “Modified Differential Evolution Algorithm for Parameter Estimation in Mathematical Models,” in
*Proceedings of the IEEE International Conference on Systems Man and Cybernetics (SMC'10)*, pp. 2767–2772, Istanbul, Turkey, October 2010. - M. Schwaab, E. C. Biscaia Jr., J. L. Monteiro, and J. C. Pinto, “Nonlinear parameter estimation through particle swarm optimization,”
*Chemical Engineering Science*, vol. 63, no. 6, pp. 1542–1552, 2008. View at Google Scholar - I. Hassanzadeh and S. Mobayen, “Controller design for rotary inverted pendulum system using evolutionary algorithms,”
*Mathematical Problems in Engineering*, vol. 2011, Article ID 572424, 17 pages, 2011. View at Publisher · View at Google Scholar - T.-S. Zhan and C.-C. Kao, “Modified PSO method for robust control of 3RPS parallel manipulators,”
*Mathematical Problems in Engineering*, vol. 2010, Article ID 302430, 25 pages, 2010. View at Publisher · View at Google Scholar · View at Zentralblatt MATH - H. Zhu, Y. Wang, K. Wanga, and Y. Chen, “Particle Swarm Optimization (PSO) for the constrained portfolio optimization problem,”
*Expert Systems with Applications*, vol. 38, no. 8, pp. 10161–10169, 2011. View at Publisher · View at Google Scholar - R. C. Eberhart and J. Kennedy, “A new optimizer using particle swarm theory,” in
*Proceedings of the 6th International Symposium on Micro Machine and Human Science*, pp. 39–43, Nagoya, Japan, October 1995. - R. C. Eberhart and Y. Shi, “Tracking and optimizing dynamic systems with particle swarms,” in
*Proceedings of the 2001 Congress on Evolutionary Computation*, pp. 94–100, Seoul, Korea, May 2001. - S. Smith, “An evolutionary program for a class of continuous optimal control problems,” in
*Proceedings of the IEEE International Conferences on Evolutionary Computation*, pp. 418–422, Perth, Australia, December 1995.