Research Article | Open Access

Volume 2018 |Article ID 5358191 | 9 pages | https://doi.org/10.1155/2018/5358191

# A Simple Alternating Direction Method for the Conic Trust Region Subproblem

Accepted27 Nov 2018
Published18 Dec 2018

#### Abstract

A simple alternating direction method is used to solve the conic trust region subproblem of unconstrained optimization. By use of the new method, the subproblem is solved by two steps in a descent direction and its orthogonal direction, the original conic trust domain subproblem into a one-dimensional subproblem and a low-dimensional quadratic model subproblem, both of which are very easy to solve. Then the global convergence of the method under some reasonable conditions is established. Numerical experiment shows that the new method seems simple and effective.

#### 1. Introduction

In this paper, we consider the unconstrained optimization problemwhere is continuously differentiable. The trust region method is a very effective method for the unconstrained optimization problem (1) (see ). Traditional trust region methods are based on a quadratic model and the corresponding quadratic program subproblem is, at the th iteration,where is the current iterate point, , is symmetric and an approximation to the Hessian of , refers to the Euclidean norm, and is the trust region radius at the th iteration. There are many methods that can be used to solve the subproblems (2)-(3). The simple, low cost, and effective methods are dogleg methods (see ). Now, we recall the simple dogleg algorithm for solving trust region subproblem with the quadratic model as in .

Algorithm 1.
tep 0. Input the data of the th iteration i.e., , and .
tep 1. Compute . If , then , and stop.
tep 2. Compute . If , then , and stop. Otherwise, go to tep 3.
tep 3. Computethen , where .

In 1980, Davidon first proposed the conic model (see ). It is an alternative model to substitute the quadratic model. For optimization problems; if the objective function has a strong nonquadratic or its curvature changes severely, then the conic model is better than the quadratic model both the effect of data fitting and the result of numerical calculations. In addition, the conic model can supply enough freedom to make best use of both information of gradients and function values in iterate points. In view of these good properties of conic model, it has attracted wide attention of many scholars . In , Ni proposed a new trust region subproblem and gave the optimality conditions for the trust region subproblems of a conic model. That is, at the th iteration, the trial step is computed by solving the following conic model trust region subproblemwherehorizon vector , is symmetric and positive semidefinite, and () is a sufficiently small positive number. We note that the conic model has a denominator and the shape of the trust region is irregular; therefore the conic trust region subproblems (5)-(7) are not easy to search for the descent point and difficult to solve. The trust region method often does not require the exact solution of trust region subproblem but only requires an approximate solution. The dogleg method for solving the trust region subproblem based on conic model is an approximate solution method; however its calculation is relatively complicated.

In this paper, we continue to study the subproblems (5)-(7). In order to find an easy way to solve, inspired by the alternating direction method of multipliers (ADMM), we consider obtaining the approximate solution by the two-step search in two orthogonal directions in the trust region . ADMM is an algorithm that solves convex optimization problems by breaking them into smaller pieces, each of which are then easier to handle. Because of its significant efficiency and easy implementation, it has recently found wide application in a number of areas (see ).

In the following, we use the alternating orthogonal direction search method to find the approximate solution of the subproblems (5)-(7). The rest of this paper is organized as follows. In the next section, the motivation and description of the simple alternating direction search method are presented. In Section 3, we give the quasi-Newton method based on the conic model for solving unconstrained optimization problems and prove its global convergence properties. The numerical results are provided in Section 4.

#### 2. Range of and the Approximate Solution of the Subproblem

In this section, we will modify the range of and give the motivation and description of the algorithm. We note that the conic model has one more parameter than . Therefore, can make full use of the existing function information to satisfy more interpolation conditions by using of the function values and the gradient values. All of these will improve the effectiveness of the algorithm. In general, chooses a descent direction, such as , , or (see ). For convenience, we omit the index of , , and in this section. Therefore, in this paper we assume thatand is positive (abbreviated as ).

LetFrom (8), we have and

Although in principle we are seeking the optimal solution of the subproblems (5)-(7), it is enough to find an approximate solution in the feasible region and guarantee a sufficient reduction in the model and the global convergence. Therefore, in order to simplify this algorithm we choose in (7) such that it satisfies

In the following, we consider the alternating direction search method to solve the subproblems (5)-(7) by making full use of the parameters . The new method is divided into two steps. First, we search in the direction of and then search in the direction which is perpendicular to .

Letwhere , and . Then, the solving process of subproblems (5)-(7) is divided into the following two stages.

In the first stage, set and then . Substituting it into (5)-(7), we havewhere

By the direct computation, we have that the derivative of iswhereFrom (17), we know that and then . Since , then has only one stationary point

By simple calculation, the following lemmas can be easily obtained.

Lemma 2. Suppose and . Then .

Proof. Combining with the definition of as in (10), we easily know thatFrom (17) and (20), we have

Lemma 3. Under the same conditions as Lemma 2, then is monotonically increasing in the trust region ; is monotonically decreasing for and .

Proof. From (19) and (20), we know thatThen combining with (18) we can obtain that the lemma obviously holds.

Let be the solution of the subproblems (14)-(15). Then, we can get the following theorems by analysis.

Theorem 4. Under the same conditions as Lemma 2, then the solution of the subproblems (14)-(15) is

Proof. (1) If then from (16) we know that . From Lemmas 2 and 3, we can obtain that if then and if then . Therefore, .
(2) If , then . From Lemmas 2 and 3, we can similarly get .
(3) If , then . From Lemmas 2 and 3, we know that Because of , then from (17) we have . Since , thenThen . The theorem is proved.

It is worth noting that if then from (17) we have . Therefore, for this case we set and exit the calculation of subproblem. Otherwise, we know that is inside the trust region . Then, we should carry out the calculation of the second stage below.

We set and substitute it into . Then the subproblems (5)-(7) becomewhereIn order to remove the equality constraint in (28), we use the null space technique. That is, for then there exist mutually orthogonal unit vectors orthogonal to the parameter vector . Set and , where . Then (27)-(28) can be simplified as the following subproblem:whereSet , , and . By Algorithm 1, we can obtain the solution of the subproblems (30)-(31). Then and . Thus, the subproblems (5)-(7) are solved approximately.

Now we could give the alternating direction search method for solving the conic trust region subproblems (5)-(7) as follows.

Algorithm 5. Given , and ,
tep 1. If , then set . Then solve the subproblems (5)-(7) by Algorithm 1 to get , and stop.
tep 2. Compute and by (9) and (10). let .
tep 3. Compute , and by (17) and (20).
tep 4. Compute by (24).
tep 5. If , then , and stop; otherwise, compute , , , and by (29) and (32).
tep 6. Set , and . Then solve the subproblems (30)-(31) by Algorithm 1 to get .
tep 7. Set and , and stop.

In order to discuss the lower bound of predicted reduction in each iteration, we define the following predicted reduction:

Now we should prove the following theorem to guarantee the global convergence of the algorithm proposed in the next section.

Theorem 6. Under the same conditions as Lemma 2, if is obtained from the above Algorithm 5, then there exists a positive constant such that

Proof. (1) If is obtained by tep 1 of Algorithm 5, then from Nocedal and Wright  we havewhere .
(2) If is obtained by tep 5 of Algorithm 5, then and , where as defined in (24). By computation, we haveFrom (24), we know that if then and . And then from (17) and (20), we can obtainandCombining with (37)-(39), we know thatwhereFor , then holds obviously.
(3) If is obtained by tep 7 of Algorithm 5, then , where . From (24), we know that . Combining with (33) and (34), we have Because is obtained by Algorithm 1, then from  we havewhere and , , and are defined by (29) and (32). Thus,where the second equality is from (20) and the last inequality is from (41).
Therefore, the theorem follows from (36), (40), and (44) with

#### 3. The Algorithm and Its Convergence

In this section, we propose a quasi-Newton method with a conic model for unconstrained minimization and prove its convergence under some reasonable conditions. In order to solve the problem (1), we approximate with a conic model of the formwhere , , , and are parameter vectors.

Now we give the simple alternating direction trust region algorithm based on conic model (46).

Algorithm 7.
tep 0. Choose parameters , , , and ; give a starting point , , , and an initial trust region radius ; set .
tep 1. Compute and . If , and then stop with as the approximate optimal solution; otherwise go to tep 2.
tep 2. Set , , , and . Then solve the subproblem (5)-(7) by Algorithm 5 to get one of the approximate solution .
tep 3. Compute the ratio of predicted reduction and the actual reduction wheretep 4. If , then set and go to tep 2. If , then set and choose the new trust region bound satisfyingtep 5. Generate and ; set , and go to tep 1.

The choice of parameter in the cone model method is crucial. In general, and are chosen to satisfy certain interpolation conditions, which means that the conic model function interpolates both the function values and the gradient values of the objective function at and . The choice of the parameters and can refer to  and , respectively. In this paper, we are not prepared to study the specific iterative formulas of and in depth and directly adopt the choice of in  and the choice of in .

From (49) and Theorem 6, we havewhere as defined by (45). The following theorem guarantees that the Algorithm 7 is globally convergent.

Theorem 8. Under the same conditions as Lemma 2, suppose that the level setand the sequence , , and is all uniformly bounded, is symmetric and positive definite, and is twice continuously differentiable in . Then for any , Algorithm 7 terminates in finite number of iterations, that is,

Proof. We give the proof by contradiction. Suppose that there is such thatFrom the hypothesis, we haveCombining with (51)-(55), we havewhere the first inequality follows fromand the second inequality is from andFrom tep 4 of Algorithm 7 and (56), we obtain that for all Since is bounded from below and , then we havewhich implies thatAnd thenOn the other hand, from (55) and (62) we can getThen from (55) to (63), we have where and . Combining with (56) and (65), we can get thatFrom (62) we have . Hence, there is a sufficiently large positive number such that andholds. From tep 4 of Algorithm 7, it follows that which is a contradiction to (62). The theorem is proved.

#### 4. Numerical Tests

In this section, Algorithm 7 is tested with some standard test problems from [16, 50]. The names of the 16 test problems are listed in Table 1. All the computations are carried out in Matlab R2016b on a microcomputer in double precision arithmetic. These tests use the same stopping criterion . The columns in the Tables have the following meanings: No. denotes the numbers of the test problems; is the dimension of the test problems; Iter is the number of iterations; is the number of function evaluations performed; is the number of gradient evaluations; is the final objective function value; is the Euclidean norm of the final gradient; CPU(s) denotes the total iteration time of the algorithm in seconds.

 No. Problem No. Problem 1 Cube 2 Penalty-I 3 Beale 4 Conic 5 Extended powell 6 Variably Dimensioned 7 Rosenbrock 8 Extended Trigonometric 9 Tridiagonal Exponential 10 Brent 11 Troesch 12 Cragg and Levy 13 Broyden Tridiagonal 14 Brown 15 Discrete Boundary Value 16 Extended Trigonometric

The parameters in these algorithms are

In order to analyze the effectiveness of our new algorithm, we compare Algorithm 7 with the alternating direction trust region method based on conic model (abbreviated as ADCTR) in . The numerical results of ADCTR and Algorithm 7 are listed in Table 2. We note that the optimal value of these test problems is . From Table 2, we can see that the performance of Algorithm 7 is feasible and effective. For the above 16 problems, Algorithm 7 is better than the ADCTR for 13 tests and is somewhat bad for 4 tests, and the two algorithms are same in efficiency for the other 1 tests. Therefore, it seems that Algorithm 7 is better than algorithm ADCTR in .

#### 5. Conclusions

The algorithm ADCTR and Algorithm 7 are similar; that is, the idea of alternating direction method is used to solve the conic trust region subproblem. However, Algorithm 7 in this paper takes into account the special property that the parameter vector is generally taken as the descending direction. Thus, under the assumption of , the calculation of Algorithm 7 is simpler to calculate and has shorter CPU time, better calculation effect, and also global convergence.

However, there are still many aspects worthy of further study, for example, weakening the positive definite condition of , using algorithms to solve large-scale problems, calculation of convergence rate, and so on.

#### Data Availability

All data generated or analysed during this study are included within the article.

#### Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

#### Acknowledgments

This work was supported by National Natural Science Foundation of China (11071117, 11771210) and the Natural Science Foundation of Jiangsu Province (BK20141409, BK20150420).

1. M. J. Powell and Y. Yuan, “A trust region algorithm for equality constrained optimization,” Mathematical Programming, vol. 49, no. 2, pp. 189–211, 1990. View at: Publisher Site | Google Scholar | MathSciNet
2. A. Vardi, “A trust region algorithm for equality constrained minimization: convergence properties and implementation,” SIAM Journal on Numerical Analysis, vol. 22, no. 3, pp. 575–591, 1985. View at: Publisher Site | Google Scholar | MathSciNet
3. P. T. Boggs, R. H. Byrd, and R. B. Schnabel, “A stable and efficient algorithm for nonlinear orthogonal distance regression,” SIAM Journal on Scientific Computing, vol. 8, no. 6, pp. 1052–1078, 1987. View at: Publisher Site | Google Scholar | MathSciNet
4. P. L. Toint, “Global convergence of a class of trust-region methods for nonconvex minimization in Hilbert space,” IMA Journal of Numerical Analysis (IMAJNA), vol. 8, no. 2, pp. 231–252, 1988. View at: Publisher Site | Google Scholar | MathSciNet
5. J. Z. Zhang and D. T. Zhu, “Projected quasi-Newton algorithm with trust region for constrained optimization,” Journal of Optimization Theory and Applications, vol. 67, no. 2, pp. 369–393, 1990. View at: Publisher Site | Google Scholar | MathSciNet
6. M. El-Alem, “A robust trust-region algorithm with a nonmonotonic penalty parameter scheme for constrained optimization,” SIAM Journal on Optimization, vol. 5, no. 2, pp. 348–378, 1995. View at: Publisher Site | Google Scholar | MathSciNet
7. M. J. D. Powell, “A hybrid method for nonlinear equations,” Numerical Methods for Nonlinear Algebraic Equations, pp. 87–114, 1970. View at: Google Scholar | MathSciNet
8. J. Dennis and H. H. Mei, “Two new unconstrained optimization algorithms which use function and gradient values,” Journal of Optimization Theory and Applications, vol. 28, no. 4, pp. 453–482, 1979. View at: Publisher Site | Google Scholar | MathSciNet
9. L. Zhang and Z. Q. Tang, “The hybrid dogleg method to solve subproblems of trust region,” Journal of Nanjing Normal University, vol. 24, no. 1, pp. 28–32, 2001. View at: Google Scholar | MathSciNet
10. J. Z. Zhang, X. J. Xu, and D. T. Zhu, “A nonmonotonic dogleg method for unconstrained optimization,” SIAM Journal on Scientific and Statistical Computing, vol. 8, no. 6, pp. 1052–1078, 1987. View at: Publisher Site | Google Scholar
11. Y. Zhao and C. Xu, “A new trust region dogleg method for unconstrained optimization,” Applied Mathematics-A Journal of Chinese Universities Series B, vol. 15, no. 1, pp. 83–92, 2000. View at: Publisher Site | Google Scholar | MathSciNet
12. H. L. Zhu, Q. Ni, and C. Y. Dang, “A new alternating direction trust region method based on conic model for solving unconstrained optimization,” https://arxiv.org/abs/1812.01935. View at: Google Scholar | MathSciNet
13. W. C. Davidon, “Conic approximations and collinear scalings for optimizers,” SIAM Journal on Numerical Analysis, vol. 17, no. 2, pp. 268–281, 1980. View at: Publisher Site | Google Scholar | MathSciNet
14. R. Schnabel, “Conic methods for unconstrained minimization and tensor methods for nonlinear equations,” Mathematical Programming the State of the Art, vol. 21, no. 1, pp. 417–438, 1982. View at: Google Scholar
15. D. C. Sorensen, “Newton's method with a model trust region modification,” SIAM Journal on Numerical Analysis, vol. 19, no. 2, pp. 409–426, 1982. View at: Publisher Site | Google Scholar | MathSciNet
16. M. F. Zhu, Y. Xue, and F. S. Zhang, “A quasi-Newton type trust region method based on the conic model,” Numerical Mathematics, vol. 17, no. 1, pp. 36–47, 1995 (Chinese). View at: Google Scholar | MathSciNet
17. X. P. Lu and Q. Ni, “A quasi-Newton trust region method with a new conic model for the unconstrained optimization,” Applied Mathematics and Computation, vol. 204, no. 1, pp. 373–384, 2008. View at: Publisher Site | Google Scholar | MathSciNet
18. C. X. Xu and X. Y. Yang, “Convergence of conic quasi-Newton trust region methods for unconstrained minimization,” Mathematical Application, vol. 11, no. 2, pp. 71–76, 1998. View at: Google Scholar | MathSciNet
19. Y.-X. Yuan, “A review of trust region algorithms for optimization,” ICIAM, vol. 99, no. 1, pp. 271–282, 2000. View at: Google Scholar | MathSciNet
20. D. M. Gay, “Computing optimal locally constrained steps,” SIAM Journal on Scientific and Statistical Computing, vol. 2, no. 2, pp. 186–197, 1981. View at: Publisher Site | Google Scholar | MathSciNet
21. J.-M. Peng and Y.-X. Yuan, “Optimality conditions for the minimization of a quadratic with two quadratic constraints,” SIAM Journal on Optimization, vol. 7, no. 3, pp. 579–594, 1997. View at: Publisher Site | Google Scholar | MathSciNet
22. W. Sun and Y. X. Yuan, “A conic trust-region method for nonlinearly constrained optimization,” Annals of Operations Research, vol. 103, pp. 175–191, 2001. View at: Publisher Site | Google Scholar | MathSciNet
23. S. Di and W. Sun, “A trust region method for conic model to solve unconstraind optimizaions,” Optimization Methods and Software, vol. 6, no. 4, pp. 237–263, 2007. View at: Publisher Site | Google Scholar
24. H. Gourgeon and J. Nocedal, “A conic algorithm for optimization,” SIAM Journal on Scientific Computing, vol. 6, no. 2, pp. 253–267, 1985. View at: Publisher Site | Google Scholar | MathSciNet
25. X. T. Zhu, M. Xi, W. Y. Sun, and J. B. S. Raimundo, “A new nonmonotone BB-TR method based on simple conic model for large scale unconstrained optimization,” Numerical Mathematics A Journal of Chinese Universities, vol. 38, no. 2, pp. 172–192, 2016. View at: Google Scholar | MathSciNet
26. X. Zhang, J. Wen, and Q. Ni, “Subspace trust-region algorithm with conic model for unconstrained optimization,” Numerical Algebra, Control and Optimization, vol. 3, no. 2, pp. 223–234, 2013. View at: Publisher Site | Google Scholar | MathSciNet
27. L. Zhao and W. Sun, “A conic affine scaling method for nonlinear optimization with bound constraints,” Asia-Pacific Journal of Operational Research, vol. 30, no. 3, pp. 1–30, 2013. View at: Google Scholar
28. Q. Ni, “Optimality conditions for trust-region subproblems involving a conic model,” SIAM Journal on Optimization, vol. 15, no. 3, pp. 826–837, 2005. View at: Publisher Site | Google Scholar | MathSciNet
29. D. Gabay and B. Mercier, “A dual algorithm for the solution of nonlinear variational problems via finite element approximation,” Computers & Mathematics with Applications, vol. 2, no. 1, pp. 17–40, 1976. View at: Publisher Site | Google Scholar
30. D. M. Gay, “Algorithm 611: Subroutines for Unconstrained Minimization Using a Model/Trust-Region Approach,” ACM Transactions on Mathematical Software, vol. 9, no. 4, pp. 503–524, 1983. View at: Publisher Site | Google Scholar | MathSciNet
31. G. Chen and M. Teboulle, “A proximal-based decomposition method for convex minimization problems,” Mathematical Programming, vol. 64, no. 1–3, pp. 81–101, 1994. View at: Publisher Site | Google Scholar | MathSciNet
32. J. Eckstein and M. Fukushima, “Some reformulations and applications of the alternating direction method of multipliers,” in Large Scale Optimization, pp. 115–134, Springer, 1994. View at: Google Scholar
33. B. S. He, L.-Z. Liao, D. R. Han, and H. Yang, “A new inexact alternating directions method for monotone variational inequalities,” Mathematical Programming, vol. 92, no. 1, pp. 103–118, 2002. View at: Publisher Site | Google Scholar | MathSciNet
34. S. Kontogiorgis and R. R. Meyer, “A variable-penalty alternating directions method for convex optimization,” Mathematical Programming, vol. 83, no. 1, pp. 29–53, 1998. View at: Publisher Site | Google Scholar | MathSciNet
35. K. Zhang, J. Li, Y. Song, and X. Wang, “An alternating direction method of multipliers for elliptic equation constrained optimization problem,” Science China Mathematics, vol. 60, no. 2, pp. 361–378, 2017. View at: Publisher Site | Google Scholar | MathSciNet
36. M. H. Xu, “Proximal alternating directions method for structured variational inequalities,” Journal of Optimization Theory and Applications, vol. 134, no. 1, pp. 107–117, 2007. View at: Publisher Site | Google Scholar | MathSciNet
37. Lu Li, Xingyu Wang, and Guoqiang Wang, “Alternating Direction Method of Multipliers for Separable Convex Optimization of Real Functions in Complex Variables,” Mathematical Problems in Engineering, vol. 2015, Article ID 104531, 14 pages, 2015. View at: Publisher Site | Google Scholar | MathSciNet
38. C. H. Chen, B. S. He, Y. Ye et al., “The direct extension of ADMM for multi-block convex minimization problems is not necessarily convergent,” Mathematical Programming, vol. 155, no. 1-2, pp. 57–79, 2016. View at: Google Scholar
39. M. L. Goncalves, J. G. Melo, and R. D. Monteiro, “Improved pointwise iteration-complexity of a regularized ADMM and of a regularized non-Euclidean HPE framework,” SIAM Journal on Optimization, 2016. View at: Google Scholar
40. J. Bai, J. Li, F. Xu, and H. Zhang, “Generalized symmetric {ADMM} for separable convex optimization,” Computational optimization and applications, vol. 70, no. 1, pp. 129–170, 2018. View at: Publisher Site | Google Scholar | MathSciNet
41. B. He and X. Yuan, “A class of ADMM-based algorithms for three-block separable convex programming,” Computational Optimization and Applications, vol. 70, no. 3, pp. 1–36, 2018. View at: Google Scholar
42. M. Hong and Z.-Q. Luo, “On the linear convergence of the alternating direction method of multipliers,” Mathematical Programming, vol. 162, no. 1-2, Ser. A, pp. 165–199, 2017. View at: Publisher Site | Google Scholar | MathSciNet
43. Y. T. Sun and J. L. Zhao, “An alternating directions method for structured split feasibility problems,” Journal on Numerical Methods and Computer Applications, vol. 39, no. 1, pp. 20–27, 2018 (Chinese). View at: Google Scholar
44. H. B. Zhu, Y. Gao, and Y. Hou, “Real-time pricing for demand response in smart grid based on alternating direction method of multipliers,” Mathematical Problems in Engineering, vol. 2018, Article ID 8760575, 10 pages, 2018. View at: Publisher Site | Google Scholar | MathSciNet
45. X.-R. Lv, Y. Li, and Y.-C. He, “Efficient impulsive noise mitigation for OFDM systems using the alternating direction method of multipliers,” Mathematical Problems in Engineering, vol. 2018, Article ID 4968682, 11 pages, 2018. View at: Publisher Site | Google Scholar | MathSciNet
46. J. Nocedal and S. J. Wright, Numerical Optimization, Science Press, Beijing, China, 2006.
47. M. J. D. Powell, “Algorithms for nonlinear constraints that use Lagrange functions,” Mathematical Programming, vol. 14, no. 1, pp. 224–248, 1978. View at: Publisher Site | Google Scholar | MathSciNet
48. M. Al-Baali, “Damped techniques for enforcing convergence of quasi-Newton methods,” Taylor and Francis, Inc., vol. 29, no. 5, pp. 919–936, 2014. View at: Google Scholar
49. Q. Ni, Optimization Method and Program Design , Science Press, Beijing, China, 2009 (Chinese).
50. J. J. Moré, B. S. Garbow, and K. E. Hillstrom, “Testing unconstrained optimization software,” ACM Transactions on Mathematical Software, vol. 7, no. 1, pp. 17–41, 1981. View at: Publisher Site | Google Scholar | MathSciNet

#### More related articles

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19. Sign up here as a reviewer to help fast-track new submissions.