Abstract
It is well known that the variational inequalities are equivalent to the fixed point problem. We use this alternative equivalent formulation to suggest and analyze some new proximal point methods for solving the variational inequalities. These new methods include the explicit, the implicit, and the extragradient methods as special cases. The convergence analysis of the new methods is considered under some suitable conditions. Results proved in this paper may stimulate further research in this direction.
1. Introduction
Variational inequalities, the origin of which can be traced back to Stampacchia [1], are being used to study a wide class of diverse unrelated problems arising in various branches of pure and applied sciences in a unified framework. It is well known that the variational inequalities are equivalent to the fixed point problem. This alternative equivalent formulation has played an important and fundamental role in the existence, numerical methods, and other aspects of the variational inequalities. This equivalent formulation has been used to suggest the projection iterative method, the implicit iterative method, and the extragradient method, which is due to Korpelevich [2], for solving the variational inequalities. It has been shown [3] that the implicit iterative method and the extragradient method are equivalent. We remark that the implicit iterative method and the explicit iterative method are two different and distinct methods. We use this alternative equivalent formulation to suggest and analyze some new proximal point methods, which include the implicit and explicit methods as special cases. This is the main motivation of this paper. We also consider its convergence criteria under suitable conditions. We hope that the ideas and techniques of this paper may stimulate further research in this area of pure and applied sciences.
2. Preliminaries
Let be a real Hilbert space, whose inner product and norm are denoted by and , respectively. Let be a nonempty, closed, and convex set in .
For a given nonlinear operator , we consider the problem of finding such that which is called the variational inequality, introduced and studied by Stampacchia [1].
For the applications, formulations, numerical methods, and other aspects of the equilibrium variational inequalities, see [1–14] and the references therein.
We now recall some well-known results and concepts.
Lemma 2.1. Let be a nonempty, closed, and convex set in . Then, for a given , satisfies the inequality if and only if where is the projection of onto the closed and convex set .
It is well known that the projection operator is nonexpansive, that is, This property plays a very important part in the studies of the variational inequalities and related optimization.
Using Lemma 2.1, one can easily show that the variational inequality (2.1) is equivalent to finding such that where is a constant.
Definition 2.2. An operator is said to be strongly monotone if and only if there exists a constant such that and Lipschitz continuous if there exists a constant such that
3. Main Results
In this section, we use the fixed point formulation (2.5) to suggest a new unified implicit method for solving the variational inequality (2.1), and this is the main motivation of this paper. Using the equivalent fixed point formulation, one can suggest the following iterative method for solving the variational inequality (2.1).
Algorithm 3.1. For a given , find the approximate solution by the iterative scheme
Algorithm 3.1 is known as the projection iterative method. For the convergence analysis of Algorithm 3.1, see Noor [8].
For a given , we can rewrite (2.5) as This fixed point formulation is used to suggest the following new proximal point iterative method for solving the variational inequality (2.1).
Algorithm 3.2. For a given , find the approximate solution by the iterative scheme
Note that Algorithm 3.2 is an implicit-type iterative method. It is clear that for , Algorithm 3.2 reduces to Algorithm 3.1. For , Algorithm 3.2 collapses to the following implicit iterative method for solving the variational inequality (2.1).
Algorithm 3.3. For a given , find the approximate solution by the iterative scheme
For the convergence analysis of Algorithm 3.3, see Noor [3] and the references therein.
In order to implement Algorithm 3.2, we use the predictor-corrector technique. We use Algorithm 3.1 as the predictor and Algorithm 3.2 as the corrector. Consequently, we obtain the following two-step iterative method for solving the variational inequality (2.1).
Algorithm 3.4. For a given , find the approximate solution by the iterative schemes:
Algorithm 3.4 is a new two-step iterative method for solving the variational inequality (2.1).
For , Algorithm 3.4 reduces to the following iterative method for solving the variational inequality (2.1).
Algorithm 3.5. For a given , find the approximate solution by the iterative schemes: which is known as the extragradient method and is due to Korpelevich [2].
For , Algorithm 3.4 reduces to the following iterative method for solving the variational inequality (2.1) and appears to be a new one.
Algorithm 3.6. For a given , find the approximate solution by the iterative schemes:
We would like to mention that one can deduce several iterative methods for solving the variational inequality and related optimization problems by choosing the appropriate and suitable value of the parameter . This clearly shows that Algorithm 3.4 is a unified implicit method and includes the previously known implicit and predictor-corrector methods as special cases.
We now consider the convergence criteria of Algorithm 3.4, and this is the main motivation of our next result.
Theorem 3.7. Let the operator be strongly monotone with constant and Lipschitz continuous with constant . If there exists a constant such that then the approximate solution obtained from Algorithm 3.4 converges strongly to the exact solution satisfying the variational inequality (2.1).
Proof. Let be a solution of (2.1), and let be the approximate solution obtained from Algorithm 3.3. Then, from (2.5) and (3.5), we have From the strongly monotonicity and the Lipschitz continuity of the operator , we obtain From (3.10) and (3.11), we obtain where Form (2.5), (3.6), (3.9), (3.12), and (3.13), we have where From (3.9), it follows that . Thus the fixed point problem (2.5) has a unique solution, and consequently the iterative solution obtained from Algorithm 3.3 converges to , the exact solution of (2.5).
For a given , we can rewrite (2.5) as This fixed point formulation (3.16) has been used to suggest and analyze the following unified proximal methods for solving the variational inequality (2.1).
Algorithm 3.8. For a given , find the approximate solution by the iterative scheme
For the convergence analysis of Algorithm 3.8, see Noor [10]. For different and appropriate choice of the parameter , Algorithm 3.8 includes the extragradient method of Korpelevich [2] and other methods as special cases.
We would like to mention that if the operator is linear, then Algorithm 3.4 and Algorithm 3.8 are equivalent. In this case, one can easily prove that the convergence of Algorithm 3.4 requires only the partially relaxed strong monotonicity of the operator , which is a weaker condition.
4. Conclusion
In this paper, we have used the equivalence between the variational inequality and the fixed point problem to suggest and analyze some new proximal point methods for solving the variational inequality. We have also shown that these new implicit methods include the extragradient method of Korpelevich [2] and the classical implicit method as special cases. We have also discussed the convergence criteria of the proposed new iterative methods under some suitable conditions. Results proved in this paper may inspire further research in this area. It is an open problem to consider the implementation of these new proximal methods and the comparison with other methods. Using the ideas and techniques of this, one can suggest and analyze several new proximal point methods for solving the general variational inequality and its variant form.
Acknowledgments
This research is supported by the Visiting Professor Program of King Saud University, Riyadh, Saudi Arabia, and the Research Grant no. KSU.VPP.108. The authors are also grateful to Dr. S. M. Junaid Zaidi, Rector, COMSATS Institute of Information Technology, Pakistan, for providing the excellent research facilities.