/ / Article
Special Issue

## Applications of Fixed Point and Approximate Algorithms

View this Special Issue

Research Article | Open Access

Volume 2012 |Article ID 412413 | https://doi.org/10.1155/2012/412413

Muhammad Aslam Noor, Khalida Inayat Noor, Eisa Al-Said, "On New Proximal Point Methods for Solving the Variational Inequalities", Journal of Applied Mathematics, vol. 2012, Article ID 412413, 7 pages, 2012. https://doi.org/10.1155/2012/412413

# On New Proximal Point Methods for Solving the Variational Inequalities

Academic Editor: Yonghong Yao
Received03 Oct 2011
Accepted19 Oct 2011
Published28 Nov 2011

#### Abstract

It is well known that the variational inequalities are equivalent to the fixed point problem. We use this alternative equivalent formulation to suggest and analyze some new proximal point methods for solving the variational inequalities. These new methods include the explicit, the implicit, and the extragradient methods as special cases. The convergence analysis of the new methods is considered under some suitable conditions. Results proved in this paper may stimulate further research in this direction.

#### 1. Introduction

Variational inequalities, the origin of which can be traced back to Stampacchia , are being used to study a wide class of diverse unrelated problems arising in various branches of pure and applied sciences in a unified framework. It is well known that the variational inequalities are equivalent to the fixed point problem. This alternative equivalent formulation has played an important and fundamental role in the existence, numerical methods, and other aspects of the variational inequalities. This equivalent formulation has been used to suggest the projection iterative method, the implicit iterative method, and the extragradient method, which is due to Korpelevich , for solving the variational inequalities. It has been shown  that the implicit iterative method and the extragradient method are equivalent. We remark that the implicit iterative method and the explicit iterative method are two different and distinct methods. We use this alternative equivalent formulation to suggest and analyze some new proximal point methods, which include the implicit and explicit methods as special cases. This is the main motivation of this paper. We also consider its convergence criteria under suitable conditions. We hope that the ideas and techniques of this paper may stimulate further research in this area of pure and applied sciences.

#### 2. Preliminaries

Let be a real Hilbert space, whose inner product and norm are denoted by and , respectively. Let be a nonempty, closed, and convex set in .

For a given nonlinear operator , we consider the problem of finding such that which is called the variational inequality, introduced and studied by Stampacchia .

For the applications, formulations, numerical methods, and other aspects of the equilibrium variational inequalities, see  and the references therein.

We now recall some well-known results and concepts.

Lemma 2.1. Let be a nonempty, closed, and convex set in . Then, for a given , satisfies the inequality if and only if where is the projection of onto the closed and convex set .

It is well known that the projection operator is nonexpansive, that is, This property plays a very important part in the studies of the variational inequalities and related optimization.

Using Lemma 2.1, one can easily show that the variational inequality (2.1) is equivalent to finding such that where is a constant.

Definition 2.2. An operator is said to be strongly monotone if and only if there exists a constant such that and Lipschitz continuous if there exists a constant such that

#### 3. Main Results

In this section, we use the fixed point formulation (2.5) to suggest a new unified implicit method for solving the variational inequality (2.1), and this is the main motivation of this paper. Using the equivalent fixed point formulation, one can suggest the following iterative method for solving the variational inequality (2.1).

Algorithm 3.1. For a given , find the approximate solution by the iterative scheme

Algorithm 3.1 is known as the projection iterative method. For the convergence analysis of Algorithm 3.1, see Noor .

For a given , we can rewrite (2.5) as This fixed point formulation is used to suggest the following new proximal point iterative method for solving the variational inequality (2.1).

Algorithm 3.2. For a given , find the approximate solution by the iterative scheme

Note that Algorithm 3.2 is an implicit-type iterative method. It is clear that for , Algorithm 3.2 reduces to Algorithm 3.1. For , Algorithm 3.2 collapses to the following implicit iterative method for solving the variational inequality (2.1).

Algorithm 3.3. For a given , find the approximate solution by the iterative scheme

For the convergence analysis of Algorithm 3.3, see Noor  and the references therein.

In order to implement Algorithm 3.2, we use the predictor-corrector technique. We use Algorithm 3.1 as the predictor and Algorithm 3.2 as the corrector. Consequently, we obtain the following two-step iterative method for solving the variational inequality (2.1).

Algorithm 3.4. For a given , find the approximate solution by the iterative schemes: Algorithm 3.4 is a new two-step iterative method for solving the variational inequality (2.1).
For , Algorithm 3.4 reduces to the following iterative method for solving the variational inequality (2.1).

Algorithm 3.5. For a given , find the approximate solution by the iterative schemes: which is known as the extragradient method and is due to Korpelevich .

For , Algorithm 3.4 reduces to the following iterative method for solving the variational inequality (2.1) and appears to be a new one.

Algorithm 3.6. For a given , find the approximate solution by the iterative schemes:

We would like to mention that one can deduce several iterative methods for solving the variational inequality and related optimization problems by choosing the appropriate and suitable value of the parameter . This clearly shows that Algorithm 3.4 is a unified implicit method and includes the previously known implicit and predictor-corrector methods as special cases.

We now consider the convergence criteria of Algorithm 3.4, and this is the main motivation of our next result.

Theorem 3.7. Let the operator be strongly monotone with constant and Lipschitz continuous with constant . If there exists a constant such that then the approximate solution obtained from Algorithm 3.4 converges strongly to the exact solution satisfying the variational inequality (2.1).

Proof. Let be a solution of (2.1), and let be the approximate solution obtained from Algorithm 3.3. Then, from (2.5) and (3.5), we have From the strongly monotonicity and the Lipschitz continuity of the operator , we obtain From (3.10) and (3.11), we obtain where Form (2.5), (3.6), (3.9), (3.12), and (3.13), we have where From (3.9), it follows that . Thus the fixed point problem (2.5) has a unique solution, and consequently the iterative solution obtained from Algorithm 3.3 converges to , the exact solution of (2.5).

For a given , we can rewrite (2.5) as This fixed point formulation (3.16) has been used to suggest and analyze the following unified proximal methods for solving the variational inequality (2.1).

Algorithm 3.8. For a given , find the approximate solution by the iterative scheme

For the convergence analysis of Algorithm 3.8, see Noor . For different and appropriate choice of the parameter , Algorithm 3.8 includes the extragradient method of Korpelevich  and other methods as special cases.

We would like to mention that if the operator is linear, then Algorithm 3.4 and Algorithm 3.8 are equivalent. In this case, one can easily prove that the convergence of Algorithm 3.4 requires only the partially relaxed strong monotonicity of the operator , which is a weaker condition.

#### 4. Conclusion

In this paper, we have used the equivalence between the variational inequality and the fixed point problem to suggest and analyze some new proximal point methods for solving the variational inequality. We have also shown that these new implicit methods include the extragradient method of Korpelevich  and the classical implicit method as special cases. We have also discussed the convergence criteria of the proposed new iterative methods under some suitable conditions. Results proved in this paper may inspire further research in this area. It is an open problem to consider the implementation of these new proximal methods and the comparison with other methods. Using the ideas and techniques of this, one can suggest and analyze several new proximal point methods for solving the general variational inequality and its variant form.

#### Acknowledgments

This research is supported by the Visiting Professor Program of King Saud University, Riyadh, Saudi Arabia, and the Research Grant no. KSU.VPP.108. The authors are also grateful to Dr. S. M. Junaid Zaidi, Rector, COMSATS Institute of Information Technology, Pakistan, for providing the excellent research facilities.

1. G. Stampacchia, “Formes bilineaires coercitives sur les ensembles convexes,” Comptes Rendus de L'Académie des Sciences, vol. 258, pp. 4413–4416, 1964. View at: Google Scholar | Zentralblatt MATH
2. G. M. Korpelevich, “An extragradient method for finding saddle points and for other problems,” Ekonomika i Matematicheskie Metody, vol. 12, no. 4, pp. 747–756, 1976. View at: Google Scholar
3. M. A. Noor, “On an implicit method for nonconvex variational inequalities,” Journal of Optimization Theory and Applications, vol. 147, no. 2, pp. 411–417, 2010.
4. F. Giannessi and A. Maugeri, Variational Inequalities and Network Equilibrium Problems, Plenum Press, New York, NY, USA, 1995.
5. F. Giannessi, A. Maugeri, and P. M. Pardalos, Equilibrium Problems: Nonsmooth Optimization and Variational Inequality Methods, Nonconvex Optimization and Its Applications, Kluwer Academic, Dordrecht, The Netherlands, 2001.
6. R. Glowinski, J. L. Lions, and R. Tremolieres, Numerical Analysis of Variational Inequalities, North-Holland, Amsterdam, Holland, 1981. View at: Publisher Site | Zentralblatt MATH
7. M. A. Noor, “General variational inequalities,” Applied Mathematics Letters, vol. 1, no. 2, pp. 119–121, 1988.
8. M. A. Noor, “Some developments in general variational inequalities,” Applied Mathematics and Computation, vol. 152, no. 1, pp. 199–277, 2004.
9. M. A. Noor, “Extended general variational inequalities,” Applied Mathematics Letters, vol. 22, no. 2, pp. 182–185, 2009.
10. M. A. Noor, On Unified Proximal Method for Variational Inequalities. In press, COMSATS Institute of Information Technology, Islamabad, Pakistan, 2011.
11. M. A. Noor, E. Al-Said, K. I. Noor, and Y. Yao, “Extragradient methods for solving nonconvex variational inequalities,” Journal of Computational and Applied Mathematics, vol. 235, no. 9, pp. 3104–3108, 2011.
12. M. A. Noor, K. I. Noor, S. Zainab, and E. Al-Said, “Proximal algorithms for solving mixed bifunction variational inequalities,” International Journal of Physical Sciences, vol. 6, no. 17, pp. 4203–4207, 2011. View at: Google Scholar
13. M. A. Noor, K. I. Noor, S. Zainab, and E. Al-Said, “Some iterative algorithms for solving regularized mixed quasi variational inequalities,” International Journal of Physical Sciences, vol. 6, 2011. View at: Google Scholar
14. M. A. Noor, K. I. Noor, and T. M. Rassias, “Some aspects of variational inequalities,” Journal of Computational and Applied Mathematics, vol. 47, no. 3, pp. 285–312, 1993.