The Scientific World Journal

Volume 2014, Article ID 307823, 6 pages

http://dx.doi.org/10.1155/2014/307823

## Levenberg-Marquardt Method for the Eigenvalue Complementarity Problem

^{1}School of Management, University of Shanghai for Science and Technology, Shanghai 200093, China^{2}College of Mathematics, Qingdao University, Qingdao 266071, China

Received 22 June 2014; Revised 28 August 2014; Accepted 29 August 2014; Published 30 October 2014

Academic Editor: Pu-yan Nie

Copyright © 2014 Yuan-yuan Chen and Yan Gao. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The eigenvalue complementarity problem (EiCP) is a kind of very useful model, which is widely used in the study of many problems in mechanics, engineering, and economics. The EiCP was shown to be equivalent to a special nonlinear complementarity problem or a mathematical programming problem with complementarity constraints. The existing methods for solving the EiCP are all nonsmooth methods, including nonsmooth or semismooth Newton type methods. In this paper, we reformulate the EiCP as a system of continuously differentiable equations and give the Levenberg-Marquardt method to solve them. Under mild assumptions, the method is proved globally convergent. Finally, some numerical results and the extensions of the method are also given. The numerical experiments highlight the efficiency of the method.

#### 1. Introduction

Eigenvalue complementarity problem (EiCP) is proposed in the study of the problems in mechanics, engineering, and economics. The EiCP is also called cone-constrained eigenvalue problem in [1–4]. The EiCP is to find a solution including a scalar and a nonzero vector satisfying a complementarity constraint on a closed convex cone. The EiCP can be reformulated to be a special complementarity problem or a mathematical programming optimization problem with complementarity constraints and can use nonsmooth or semismooth Newton type method to solve it, such as [5–7]. The Levenberg-Marquardt method is one of the widely used methods in solving optimization problems (see, for instance, [8–15]). Use a trust region strategy to replace the line search, the Levenberg-Marquardt method is widely considered to be the progenitor of the trust region method approach for general unconstrained or constrained optimization problems. The use of a trust region avoids the weaknesses of Gauss Newton method, that is, its behavior when the Jacobian is rank deficient or nearly so rank deficient. On the other hand, we reformulate the EiCP as a system of continuously differentiable equations that is one of the most interesting themes. The advantage of the reformulation is that we solve the equations with continuously differentiable functions for which there are rich powerful solution methods and theory analysis, including the powerful Levenberg-Marquardt method. So, in this paper, we give the Levenberg-Marquardt method to solve the EiCP. The EiCP, which we will consider, is the following problem. Given the matrix and the matrix , which are positive definite matrix, then we consider to find a scalar and a vector , such that

This paper is organized as follows. In Section 2, we give some background definitions and known properties. And we also give the Levenberg-Marquardt method for the EiCP. The global convergence analysis and some discussions of the Levenberg-Marquardt method is also given. In Sections 3 and 4, we give some numerical results and some extensions of the method.

Throughout the paper, denotes a real matrix of order and denotes a real matrix of order . and .

#### 2. Preliminaries

In this section, firstly, we reformulate (1) as a system of continuously differentiable equations and give some preliminaries used in the following. Then, we propose the Levenberg-Marquardt method for the EiCP.

As we all know, (1) can be rewritten as where is a continuously differentiable function and the matrix . Using the nonlinear complementarity problem (NCP) function , which satisfied the following basic property By the property, (2) can be recasted as the following system of equations: Then, solves the EiCP (1) if and only if solves (4). Define a merit function for (4) as We use a special NCP function named Fischer-Burmeister function, defined as We know that the favorable property of is that is a continuously differentiable function on the whole space, although is not a continuously differentiable function in general (see, for example, [16]). Thus, we give the system of continuously differentiable equations for (1) as the following equations: where , which is a continuously differentiable function. In what follows, we will give the Levenberg-Marquardt method. Denote The least-squares formulation of (7) is the following unconstrained optimization problem:

Now, we give the Levenberg-Marquardt method for solving (1). The global convergence result of the method is also given.

*Method 1 (the Levenberg-Marquardt method for the EiCP). *Given , , , , . , , , , and .

*Step 1. *If , then stop. Otherwise, compute by

*Step 2. *If , let ; otherwise, is computed by (10). Then, find the smallest nonnegative integer such that
Set .

*Step 3. *Let and , and go to Step 1.

Now, we give the global convergence of Method 1.

Theorem 1. *Suppose that , generated by Method 1. Then, each accumulation point of the sequence is a stationary point of .*

*Proof. *Suppose that , is a subsequence of , and . When there are infinitely many such that , by Proposition 1.16 in [17], we get the assertion. In the following, we assume that if is a convergent subsequence of , then is always computed by (10). We assume that for every convergent subsequence for which
we have
In the following, we also assume that . Suppose that is not a stationary point of . By (10), we have
therefore, we have
Obviously, the denominator in the above inequality is nonzero; otherwise, we have . Then, the algorithm has stopped. On the other hand, we know that there exists a constant such that
moreover,
By and the fact that the gradient is bounded on the convergent sequence , we get (13). We next prove (14). If (14) is not satisfied, there exists a subsequence of :
This implies that . From (18), we know that
which contradicts with (12). Thus, (14) holds. So, according to the definition given in [17], the sequence is uniformly gradient related to . We complete the proof.

*Remark 2. *In Method 1, we can also use some other line search, such as the nonmonotone line search. The line search is to find the smallest nonnegative integer such that
where , and is a integer.

*Remark 3. *In Method 1, we can also use the following equation to compute in Step 1. We can find an approximate solution of the equation
where is the residuals and satisfies
where for every .

#### 3. Numerical Results

We give some numerical experiments for the method. And we compare Method 1 with the scaling and projection algorithm (denoted by SPA in [18]). The numerical results indicate that Method 1 works quite well in practice. We consider the eigenvalue complementarity problems, which are all taken form [3, 18]. All codes for the method are finished in MATLAB. The parameters used in the method are chosen as , , , and .

*Example 4. *We consider
where

By [3], we know that Example 4 has three eigenvalues. Now, we consider random initial points to compute Example 4 by Method 1. Numerical results for Example 4 by Method 1 and SPA method are presented in Table 1.

Table 1 shows that Method 1 are able to detect all the solutions for the small size matrix. But the SPA method can only detect 2 solutions from [3].

*Example 5. *We consider
where

From [3], we also know that Example 5 have 9 eigenvalues. Now, we consider random initial points to compute the above Example 5 by Method 1. Numerical results for Example 5 by Method 1 and the SPA method are presented in Table 2.

Table 2 shows that Method 1 is able to detect all the solutions. But, the SPA method can only detect 4 Pareto eigenvalues from the analysis of [3].

*Example 6. *We consider
where

From [3], we also know that Example 6 have 23 eigenvalues. Now, we consider random initial points to compute Example 6 by Method 1. Numerical results for Example 6 by Method 1 and the SPA method are given in Table 3.

Table 3 shows that Method 1 is able to detect all 23 solutions. But, the SPA method can only detect 2 Pareto eigenvalues from [3]. The numerical results indicate that Method 1 works quite well for the big size EiCP in practice.

*Discussion*. In this section, we study the numerical behaviors of Method 1 for solving the Pareto eigenvalue problem. The EiCP problem is very useful in studying the optimization problems arising in many areas of the applied mathematics and mechanics. By using the F-B function, we reformulate the EiCP as a system of continuously differentiable equations. Then, we use the Levenberg-Marquardt method to solve it. From the above numerical results, we know that Method 1 is very effective for small size and big size EiCP problems.

#### 4. Extensions

##### 4.1. Bieigenvalue Complementarity Problems (BECP)

We also can use Method 1 to solve the bieigenvalue complementarity problems (denoted by BECP). The BECP is to find and such that where , , , and .

Let , , , , and . We can write the above bieigenvalue complementarity problems as , , , , , , , and . By using F-B function, similar to rewriting the EiCP, we can rewrite the above bieigenvalue complementarity problems as the following equations: Let Then, we can give the system of continuously differentiable equations for the bieigenvalue complementarity problems as the following continuously differentiable equations: In the following numerical test, the parameters used in the method are also chosen as , , , and .

*Example 7. *We consider the BECP, where , ,

We use Method 1 to solve the Example 7. Results for Example 7 with random initial point are given in Table 4.

##### 4.2. The Paretian Version

We consider using Method 1 to compute the following Paretian version: where standing for the pencil associated to a finite collection of real matrices of order and

*Example 8. *We consider the quadratic pencil model, where
We use Method 1 to compute Example 8. We present the results of Example 8 with random initial point in Table 5.

#### 5. Conclusion

In this paper, we reformulate the EiCP as a system of continuously differentiable equations and use the Levenberg-Marquardt method to solve them. The numerical experiments show that our method is a promising method for solving the EiCP. The numerical experiments of the extensions confirm the efficiency of our method.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

This work is supported by National Science Foundation of China (11171221, 11101231), Shanghai Leading Academic Discipline (XTKX 2012), and Innovation Program of Shanghai Municipal Education Commission (14YZ094).

#### References

- J. J. Júdice, M. Raydan, S. S. Rosa, and S. A. Santos, “On the solution of the symmetric eigenvalue complementarity problem by the spectral projected gradient algorithm,”
*Numerical Algorithms*, vol. 47, no. 4, pp. 391–407, 2008. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - J. J. Júdice, H. D. Sherali, and I. M. Ribeiro, “The eigenvalue complementarity problem,”
*Computational Optimization and Applications*, vol. 37, no. 2, pp. 139–156, 2007. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - C. F. Ma, “The semismooth and smoothing Newton methods for solving Pareto eigenvalue problem,”
*Applied Mathematical Modelling*, vol. 36, no. 1, pp. 279–287, 2012. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - S. Adly and A. Seeger, “A nonsmooth algorithm for cone-constrained eigenvalue problems,”
*Computational Optimization and Applications*, vol. 49, no. 2, pp. 299–318, 2011. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - S. Adly and H. Rammal, “A new method for solving Pareto eigenvalue complementarity problems,”
*Computational Optimization and Applications*, vol. 55, no. 3, pp. 703–731, 2013. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - J. J. Júdice, H. D. Sherali, I. M. Ribeiro, and S. S. Rosa, “On the asymmetric eigenvalue complementarity problem,”
*Optimization Methods and Software*, vol. 24, no. 4-5, pp. 549–568, 2009. View at Publisher · View at Google Scholar · View at Scopus - A. Pinto da Costa and A. Seeger, “Cone-constrained eigenvalue problems: theory and algorithms,”
*Computational Optimization and Applications*, vol. 45, no. 1, pp. 25–57, 2010. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - C. Wang, Q. Liu, and C. Ma, “Smoothing SQP algorithm for semismooth equations with box constraints,”
*Computational Optimization and Applications*, vol. 55, no. 2, pp. 399–425, 2013. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - F. Facchinei and C. Kanzow, “A nonsmooth inexact Newton method for the solution of large-scale nonlinear complementarity problems,”
*Mathematical Programming*, vol. 76, no. 3, pp. 493–512, 1997. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - J. Y. Fan and Y. Y. Yuan, “On the quadratic convergence of the Levenberg-Marquardt method without nonsingularity assumption,”
*Computing. Archives for Scientific Computing*, vol. 74, no. 1, pp. 23–39, 2005. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - S.-Q. Du and Y. Gao, “The Levenberg-Marquardt-type methods for a kind of vertical complementarity problem,”
*Journal of Applied Mathematics*, vol. 2011, Article ID 161853, 12 pages, 2011. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - J. Y. Fan, “A Shamanskii-like Levenberg-Marquardt method for nonlinear equations,”
*Computational Optimization and Applications*, vol. 56, no. 1, pp. 63–80, 2013. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - R. Behling and A. Fischer, “A unified local convergence analysis of inexact constrained Levenberg-Marquardt methods,”
*Optimization Letters*, vol. 6, no. 5, pp. 927–940, 2012. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - K. Ueda and N. Yamashita, “Global complexity bound analysis of the Levenberg-Marquardt method for nonsmooth equations and its application to the nonlinear complementarity problem,”
*Journal of Optimization Theory and Applications*, vol. 152, no. 2, pp. 450–467, 2012. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - F. Facchinei, A. Fischer, and M. Herrich, “A family of Newton methods for nonsmooth constrained systems with nonisolated solutions,”
*Mathematical Methods of Operations Research*, vol. 77, no. 3, pp. 433–443, 2013. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - H. Jiang, M. Fukushima, L. Qi, and D. Sun, “A trust region method for solving generalized complementarity problems,”
*SIAM Journal on Optimization*, vol. 8, no. 1, pp. 140–157, 1998. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - D. P. Bertsekas,
*Constrained Optimization and Lagrange Multiplier Methods*, Academic Press, New York, NY, USA, 1982. View at MathSciNet - A. P. da Costa and A. Seeger, “Numerical resolution of cone-constrained eigenvalue problems,”
*Computational & Applied Mathematics*, vol. 28, no. 1, pp. 37–61, 2009. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus