Research Article | Open Access

Volume 2014 |Article ID 563860 | https://doi.org/10.1155/2014/563860

Wei-xiang Wang, You-lin Shang, Ying Zhang, "Global Minimization of Nonsmooth Constrained Global Optimization with Filled Function", Mathematical Problems in Engineering, vol. 2014, Article ID 563860, 5 pages, 2014. https://doi.org/10.1155/2014/563860

# Global Minimization of Nonsmooth Constrained Global Optimization with Filled Function

Accepted29 Aug 2014
Published25 Sep 2014

#### Abstract

A novel filled function is constructed to locate a global optimizer or an approximate global optimizer of smooth or nonsmooth constrained global minimization problems. The constructed filled function contains only one parameter which can be easily adjusted during the minimization. The theoretical properties of the filled function are discussed and a corresponding solution algorithm is proposed. The solution algorithm comprises two phases: local minimization and filling. The first phase minimizes the original problem and obtains one of its local optimizers, while the second phase minimizes the constructed filled function and identifies a better initial point for the first phase. Some preliminary numerical results are also reported.

#### 1. Introduction

Science and economics rely on the increasing demand for locating the global optimization optimizer, and therefore global optimization has become one of the most attractive research areas in optimization. However, the existence of multiple local minimizers that differ from the global solution confronts us with two difficult issues, that is, how to escape from a local minimizer to a smaller one and how to verify that the current minimizer is a global one. These two issues make most of global optimization problems unsolvable directly by classical local optimization algorithms. Up to now, various kinds of new theories and algorithms on global optimization have been presented . In general, global optimization methods can be divided into two categories: stochastic methods and deterministic methods. The stochastic methods are usually probability based approaches, such as genetic algorithm and simulated annealing method. These stochastic methods have their advantages, but their shortages are also obvious, such as being easily trapped in a local optimizer. Deterministic methods such as filled function method , tunneling method , and stretching function method  can, however, often skip from the current local minimizer to a better one.

The filled function approach, initially proposed for smooth optimization by Ge and Qin  and improved in , is one of the effective global optimization approaches. It furnishes us with an efficient way to use any local optimization procedure to solve global optimization problem. The essence of filled function method is to construct a filled function and then to minimize it to obtain a better initial point for the minimization of the original problem. The existing filled function methods are usually only suitable for unconstrained optimization problem. Moreover, it requires the objective function to be continuously differentiable and the number of local minimizers to be finite. But, in practice, optimization problems may be nonsmooth and often have many complicated constraints and the number of local minimizers may also be infinite. To deal with such situation, in this paper, we extend filled function methods to the case of nonsmooth constrained global optimization and propose a new filled function method. The proposed filled function method combines filled function method for unconstrained global optimization with the exterior penalty function method for constrained optimization.

The paper is organized as follows. In Section 2, a new filled function is proposed and its properties are discussed. In Section 3, a corresponding filled function algorithm is designed and numerical experiments are performed. Finally, in Section 4, some conclusive remarks are given.

In the rest of this paper, the generalized gradient of a nonsmooth function at the point is denoted by and the generalized directional derivative of in the direction at the point is denoted by . The interior, the boundary, and the closure of the set are denoted by , and , respectively.

#### 2. A New One-Parameter Filled Function and Its Properties

##### 2.1. Problem Formulation

Consider the following nonsmooth constrained global optimization problem : where , is a box set, and , are Lipschitz continuous with constants and , respectively, and is an index set. For simplicity, the set of local minimizers for problem is denoted by and the set of the global minimizers is denoted by .

To proceed, we assume that the number of minimizers of problem is infinite, but the number of different function values at the minimizers is finite.

Definition 1. A function is called a filled function of at a local minimizer if it satisfies the following:(1) is a strictly local maximizer of on ;(2)for any or one has , where ;(3)if is not empty, then there exists a point such that is a local minimizer of .

Definition 1 guarantees that when any local search procedure for unconstrained optimization is used to minimize the constructed filled function, the sequences of iterative point will not stop at any point at which the objective function value is larger than . If is not a global minimizer, then a point with could be identified in the process of the minimization of . Then, we can get a better local minimizer of by using as an initial point. By repeating these two steps, we could finally obtain a global minimizer or an approximate global minimizer of the original problem.

##### 2.2. A New Filled Function and Its Properties

We propose in this section a one-parameter filled function as follows: where is a parameter and is the current local minimizer of .

Theorems 24 show that when parameter is suitably large, the function is a filled function.

Theorem 2. Assume that ; then is a strictly local maximizer of .

Proof. Since , there is a neighborhood of with such that and , for all . We consider the following two cases.
Case  1. For all and , we have , and so
Case  2. For all , there exists at least one index such that , it follows that Thus
The above discussion indicates that is a strictly local maximizer of .

Theorem 3. Assuming that , then, for any or , it holds that when is big enough.

Proof. For any , or , we have , and so Thus where . Therefore, when is big enough, it holds that which implies that .

Theorem 4. Assume that but and . If is suitably large, then there exists a point such that is a local minimizer of .

Proof. By the conditions, there exists an such that .
Then, for any , Since, as , then when is suitably big, we have Assume that the function reaches its global minimizer over at . Since is an open set, then , and it holds that In the following, we will prove that which leads to the result.
The proof is by contradiction. Suppose that Then, as , which implies that when is suitably big, we have This is a contradiction.

#### 3. Algorithm and Numerical Examples

Based on the properties of the proposed filled function, we now give a corresponding filled function algorithm as follows.

##### 3.1. Filled Function Algorithm FFAM

Initialization Step(1)Set a disturbance constant .(2)Select an upper bound of denoted by and set .(3)Select directions , with integer , where is the number of variables.(4)Set .

Main Step(1)Start from an initial point ; minimize the problem by implementing a nonsmooth local search procedure and obtain the first local minimizer of .(2)Let .(3)Construct a filled function at : (4)If , then go to (7). Else, set as an initial point, minimize by implementing a nonsmooth local search algorithm, and obtain a local minimizer .(5)If , then set and go to (4). Else, go to (6).(6)If meets , then set and . Start from as a new initial point, minimize the problem by using a local search algorithm, and obtain another local minimizer of with . Set and go to (2). Else, go to (7).(7)Increase by setting .(8)If , then set and go to (3). Else, the algorithm stops and is taken as a global minimizer of the problem .

At the end of this section, we make a few remarks below.

(1) Algorithm FFAM is comprised of two stages: local minimization and filling. In stage 1, a local minimizer of is identified. In stage 2, filled function is constructed and then minimized. Stage 2 terminates when one point is located. Then, algorithm FFAM reenters into stage 1, with as an initial point to find a new minimizer of (if such one minimizer exists), and so on. The above process is repeated until some certain specified stopping criteria are met, and then the last local minimizer is regarded as a global minimizer.

(2) The motivation and mechanism behind the algorithm FFAM are given below.

In Step (3) of the Initialization Step, we can choose directions , as positive and negative unit coordinate vectors. For example, when , the directions can be chosen as , and .

In Step (1) and Step (6) of the Main Step, we can obtain a local optimizer of the problem by using any nonsmooth constrained local optimization procedure such as Bundle methods and Powell’s method. In Step (4) of the Main Step, we can minimize the proposed filled function by using Hybrid Hooke and Jeeves-Direct Method for Nonsmooth Optimization , Mesh Adaptive Direct Search Algorithms for Constrained Optimization , and so forth.

(3) The proposed filled function algorithm can also be applied to smooth constrained optimization. Any smooth local minimization procedure in the minimization phase can be used, such as conjugate gradient method and quasi-Newton method.

##### 3.2. Numerical Examples

We perform the numerical tests for three examples. All the numerical tests are programmed in Fortran 95. In nonsmooth case, we search for the local minimizers by using Hybrid Hooke and Jeeves-Direct Method for Nonsmooth Optimization  and the Mesh Adaptive Direct Search Algorithms for Constrained Optimization . In smooth case, we use penalty function method and conjugate gradient method to get the local minimizers.

The following are the three examples and their numerical results. And the symbols used in the tables are explained below:: the iteration number in finding the th local minimizer,: the th new initial point in finding the th local minimizer,: the th local minimizer,: the function value of the th local minimizer.

Problem 1. Consider Algorithm FFAM succeeds in finding an approximate global minimizer with . The computational results are given in Table 1.

 1 (−1, −1, −1) (−1.9802, −0.0132, −0.0006) −1.9410 2 (1.1931, 0.6332, −1.1932) (1.9889, −0.0001, −0.0111) −5.9477

Problem 2. Consider Algorithm FFAM succeeds in finding a global minimizer with . The computational results are listed in Table 2.

 1 (−1, −1) (−15.0000, 0.0000) 5.7164 2 (−1.0585, 0.5165) (0.0001, −0.2094) −0.3690 3 (0.0007, −0.0435) (0.0000, 0.0000) −2.7183

Problem 3. Consider Algorithm FFAM succeeds in finding a global minimizer with . This problem is taken from . We give this numerical example here to illustrate that algorithm FFAM is also suitable for smooth constrained global optimization. The computational results are given in Table 3.

 1 (0, 0) (0.6116, 3.4423) −4.0541 2 (2.1653, 2.2546) (2.3295, 3.1780) −5.5081

#### 4. Conclusion

In this paper, we present a new filled function for both nonsmooth and smooth constrained global optimization and investigate its properties. The filled function contains only one parameter which can be readily adjusted in the process of minimization. We also design a corresponding filled function algorithm. Moreover, in order to demonstrate the performance of the proposed filled function method, we make three numerical tests. The preliminary computational results show that the proposed filled function approach is promising.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

This work was supported by the NNSF of China (nos. 11471102 and 11001248), the SEDF under Grant no. 12YZ178, the Key Discipline “Applied Mathematics” of SSPU under no. A30XK1322100, and NNSF of Zhejiang (no. LY13A010006).

1. P. M. Pardalos and H. E. Romeijn, Eds., Handbook of Global Optimization: Volume 22: Heuristic Approaches, Kluwer Academic Publishers, Dordrecht, The Netherlands, 2002.
2. P. M. Pardalos and H. E. Romeijn, Eds., Handbook of Global Optimization, vol. 62 of Nonconvex Optimization and Its Applications, Kluwer Academic Publishers, Dordrecht, The Netherlands, 2002.
3. R. Horst and P. M. Pardalos, Eds., Handbook of Global Optimization, vol. 2 of Nonconvex Optimization and Its Applications, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1995.
4. R. P. Ge and Y. F. Qin, “A class of filled functions for finding global minimizers of a function of several variables,” Journal of Optimization Theory and Applications, vol. 54, no. 2, pp. 241–252, 1987.
5. W. Wang, Y. Yang, and L. Zhang, “Unification of filled function and tunnelling function in global optimization,” Acta Mathematicae Applicatae Sinica: English Series, vol. 23, no. 1, pp. 59–66, 2007. View at: Publisher Site | Google Scholar | MathSciNet
6. Z. Xu, H. Huang, P. M. Pardalos, and C. Xu, “Filled functions for unconstrained global optimization,” Journal of Global Optimization, vol. 20, no. 1, pp. 49–65, 2001. View at: Publisher Site | Google Scholar | MathSciNet
7. Y.-J. Yang and Y.-L. Shang, “A new filled function method for unconstrained global optimization,” Applied Mathematics and Computation, vol. 173, no. 1, pp. 501–512, 2006. View at: Publisher Site | Google Scholar | MathSciNet
8. A. V. Levy and A. Montalvo, “The tunneling algorithm for the global minimization of functions,” SIAM Journal on Scientific and Statistical Computing, vol. 6, no. 1, pp. 15–29, 1985. View at: Publisher Site | Google Scholar | MathSciNet
9. Y. Wang and J. Zhang, “A new constructing auxiliary function method for global optimization,” Mathematical and Computer Modelling, vol. 47, no. 11-12, pp. 1396–1410, 2008. View at: Publisher Site | Google Scholar | MathSciNet
10. C. J. Price, B. L. Robertson, and M. Reale, “A hybrid Hooke and Jeeves—direct method for non-smooth optimization,” Advanced Modeling and Optimization, vol. 11, no. 1, pp. 43–61, 2009. View at: Google Scholar | MathSciNet
11. C. Audet and J. Dennis, “Mesh adaptive direct search algorithms for constrained optimization,” SIAM Journal on Optimization, vol. 17, no. 1, pp. 188–217, 2006. View at: Publisher Site | Google Scholar | MathSciNet
12. C. A. Floudas and P. M. Pardalos, A Collection of Test Problems for Constrained Global Optimization Algorithms, Springer, Berlin , Germany, 1990. View at: MathSciNet