Abstract
The construction of a specific splitting-type preconditioner in block formulation applied to a class of group relaxation iterative methods derived from the centred and rotated (skewed) finite difference approximations has been shown to improve the convergence rates of these methods. In this paper, we present some theoretical convergence analysis on this preconditioner specifically applied to the linear systems resulted from these group iterative schemes in solving an elliptic boundary value problem. We will theoretically show the relationship between the spectral radiuses of the iteration matrices of the preconditioned methods which affects the rate of convergence of these methods. We will also show that the spectral radius of the preconditioned matrices is smaller than that of their unpreconditioned counterparts if the relaxation parameter is in a certain optimum range. Numerical experiments will also be presented to confirm the agreement between the theoretical and the experimental results.
1. Introduction
Consider the finite difference discretisation schemes for solving the following boundary value problem which is the two-dimensional Poisson equation with Dirichlet boundary conditions: Here, is a continuous unit square solution domain with boundary . This equation plays a very important role in the modelers of fluid flow phenomena and heat conduction problems. Let be discretised uniformly in both and directions with a mesh size , where is an integer. The simplest finite difference approximation of the Laplacian is Here, . Another approximation to (1.1) can be derived from the rotated five-point finite-difference approximation to give [1] Based on the latter approximation, improved point and group iterative schemes have been developed over the last few years in solving several types of partial differential equations [1–5]. In particular, the Modified Explicit Decoupled Group (MEDG) method [6, 7] was formulated as the latest addition to this family of four-point explicit group methods in solving the Poisson equation. This method has been shown to be the fastest among the existing explicit group methods due to its lesser computational complexity.
Since it is well known that preconditioners play a vital role in accelerating the convergence rates of iterative methods, several preconditioned strategies have been used for improving the convergence rate of the explicit group methods derived from the standard and skewed (rotated) finite difference operators [8–11]. In particular, Saeed and Ali [12–14] presented an -type preconditioning matrix applied to the systems obtained from the four-point Explicit Decoupled Group (EDG) and the Modified Explicit Decoupled Group (MEDG) methods for solving the elliptic partial differential equation, where is obtained by taking the first upper diagonal groups of the iteration matrix of the original system. The numerical experiments performed on these methods were seen to yield very encouraging results. However, no detailed studies of the spectral radius analysis of all these preconditioned systems have been done to confirm the superiority of this preconditioner.
The focus of this study is to establish the convergence properties of the preconditioned systems based on the splitting-type preconditioner for improving the performance and reliability of this family of explicit group methods derived from the rotated finite-difference formula. We will prove that this type of preconditioner applied to the MEDG SOR can minimize the most the spectral radius of the preconditioned matrix provided the relaxation parameter is in a certain optimum range. This paper is organised as follows: in Section 2, we give a presentation of the preconditioner applied to the system resulted from the EDG SOR method. A brief description of the application of the preconditioner in block formulation to the MEDG SOR is given in Section 3. The theoretical convergence analysis of these methods is discussed in Section 4. In Section 5, we give a numerical example to confirm the results obtained in Section 4. Finally, we report a brief conclusion in Section 6.
2. Preconditioned Explicit Decoupled Group SOR (EDG SOR)
For convenience, we will now briefly explain some of the definitions used in this paper.
Definition 2.1 (see [15]). A matrix of order has property if there exists two disjoint subsets and of such that if and if either and , then and or else and .
Definition 2.2 (see [3]). An ordered grouping of is a subdivision of into disjoint subsets such that . Given a matrix and an ordered grouping , we define the submatrices for as follows: is formed from deleting all rows except those corresponding to and all columns except those corresponding to .
Definition 2.3 (see [3]). Let be an ordered grouping with groups. A matrix has Property if the matrix defined by if or 1 if has Property .
Definition 2.4 (see [15]). A matrix of order is consistently ordered if for some t there exist disjoint subsets of such that and such that if and are associated, then if and if , where is the subset containing i.
Note that a matrix is a -consistently ordered matrix if the matrix in Definition 2.3 is consistently ordered.
From the discretisation of the EDG finite-difference formula in solving the Poisson equation, the linear system
is obtained with [1]
Let , be written as , where , and are strict block lower triangular, and strict block upper triangular parts of . Here, the diagonal entries are nonsingular and denotes the set of all matrices in which is of the form (2.1) relative to some given block partitioning . The block Jacobi iteration matrix is , where ,, the block Gauss-Seidel iteration matrix is , and the Block Successive Over-Relaxation method (BSOR) iteration matrix is
Since the matrix of (2.1) is a -consistently ordered and possesses property , therefore, the theory of block SOR is valid for this iterative method [1].
The theoretical optimum relaxation factor for implementing the group SOR iterative scheme can be computed from the formula:
where is the spectral radius of the group Jacobian iterative matrix. Yousif and Evans [16] gave a good estimate of the spectral radius for the EDG method:
In an effort to further accelerate the convergence rates of this method, Saeed and Ali [12] applied a preconditioner to the linear system (2.1) and transformed it into an equivalent system:
with , where is the identity matrix which has the same dimension as while is obtained by taking the first upper diagonal groups of in the original system above as the following:
Here, is a null matrix.
The system (2.1) becomes
Hence, we have the linear system of equations:
with
The SOR iteration matrix of this scheme is called the Modified Block Successive Over-Relaxation iteration matrix (MBSOR) and is given by
The matrix of (2.9) is -consistently ordered and possesses property [13].
3. Preconditioned-Modified Explicit Decoupled Group SOR (MEDG SOR)
Using the MEDG approximation formula in discretising the Poisson equation, the following system is obtained [6]: where It is observed that the partitioning of is in the following block form: with , where , , and . Let , where and are block matrices consisting of the block diagonal, strict block lower triangular, and strict block upper triangular parts of . Here, the diagonal entries are nonsingular. The block Jacobi iteration matrix is , where , , while the block Gauss-Seidel iteration matrix is . The Block Successive Over-Relaxation method (BSOR) iteration matrix is, therefore, Since the matrix of (3.3) is -consistently ordered and possesses property , the theory of block SOR is also valid for this iterative method and, therefore, is convergent [6].
Similarly, the theoretical optimum relaxation factor for implementing this group SOR-iterative scheme can be obtained from (2.4). In view of the fact that the grid spacing , an estimate of the spectral radius of the group Jacobian iterative matrix of the MEDG method may be obtained from (2.5) as Good agreement between the theoretical estimates and experimental values of the optimum relaxation parameters was observed in our numerical experiments. Upon applying the left-sided preconditioner to system (3.1), the following system is obtained [13]: with where is a null matrix.
The preconditioner is of the following form: Here, is a identity matrix and the system (3.7) becomes Hence, where The SOR iteration matrix will result in an Improved Modified Block Successive Over-Relaxation iteration matrix (IMBSOR) and is given by
4. Convergence Properties of the Preconditioned Group Methods
In this section, we will derive several properties related to the convergence of the preconditioned methods discussed in Sections 2 and 3. We will begin with the presentation of several preliminary relevant theorems and lemmas which are needed for the proof of the convergence properties. The spectral radius of a matrix is denoted by , which is defined as the largest of the moduli of the eigenvalues of the iteration matrix.
Theorem 4.1 (see [15]). If is a regular splitting of the matrix and , then Thus, an iterative method with coefficient matrix is convergent for any initial vector .
An accurate analysis of convergence properties of the SOR method is possible if the matrix is consistently ordered in the following sense (see [17]).
Definition 4.2. A matrix is a generalized -consistently ordered matrix (a GCO-matrix) if is independent of for all and for all . Here, and and are strictly lower and strictly upper triangular matrices, respectively, such that .
Definition 4.3 (see [17]). A matrix of the form (3.3) is said to be generally consistently ordered or simply GCO, where and are positive integers, if for the partitioning of , the diagonal submatrices ≥2, are nonsingular, and the eigenvalues of
are independent of , for all , where and are strict block lower and upper triangular parts of respectively.
For any matrix in , let denote the block matrix in with entries. Given the matrix
then denotes the spectral radius of the matrix:
so that
Lemma 4.4 (see [17]). Let of (4.4) be a GCO -matrix and . Then, for any real nonnegative constant , , and with satisfying: , the matrix is such that .
Lemma 4.5 (see [14]). Suppose is a GCO, where and are strictly lower and upper triangular matrices, respectively. Let be the block iteration matrix of the SOR method given by (2.3). If , then the block SOR method converges, that is, .
Theorem 4.6 (see [14]). Suppose is a GCO, where and are strictly lower and upper triangular matrices, respectively. Let and be the iteration matrices of the SOR method given by (2.3) and (2.11), respectively. If , then (i) if ,(ii) if ,(iii) if .Using the results and definitions stated above, we can prove the following lemma and theorems in relation to the spectral radius of the iteration matrices of the preconditioned group methods and their unpreconditioned counterparts.
Lemma 4.7. Suppose is a GCO, where and are strictly lower and upper triangular matrices, respectively. Let be the block iteration matrix of the SOR method given by (3.5). If , then the block SOR method converges, that is, .
Proof. Let the matrix with partitioning be given as in (3.3) and let the block SOR iteration matrix be given as in (3.5).
Set
Clearly, we can see that and hence we can conclude that .
Consider the matrix defined by
where and . It is easily seen that is nonsingular and . Moreover, since and , is a regular splitting of (cf.[11]). For satisfying the condition , Lemma 4.4 implies that . Therefore, recalling Theorem 4.1 above, we have . Hence, , which completes the proof.
The result of Lemma 4.7 enables us to prove the following theorem
Theorem 4.8. Suppose is a GCO, where and are strictly lower and upper triangular matrices, respectively. Let and be the iteration matrices of the SOR method given by (3.5) and (3.13), respectively. If , then (i) if ,(ii) if ,(iii) if .
Proof. From Lemma 4.7 and since the matrix of (3.3) is a GCO and , there exists a positive vector such that
where or equivalently
Also, since
we can write
Rearrange (4.11), we can get
But from (4.9), we have
Therefore, (4.12) can be written as
Hence, for and from [10], we can get(i), then and from Theorem 4.6 we have ,(ii), then and from Theorem 4.6 we have ,(iii), then and from Theorem 4.6 we have .
Thus, the proof is complete.
Theorem 4.9. Suppose and are GCO, where , , and are strictly lower and upper triangular matrices of and , respectively. Let , , and be the iteration matrices of the SOR method given by (2.3), (2.11), (3.5), and (3.13), respectively. If , then (i) if ,(ii) if ,(iii) if .
Proof. In the same manner of the proof of Theorem 4.8 and since the matrix of (2.9) is a GCO, see [13], and , there exists a positive vector such that
where
Equation (4.15) can be written as
Also, since , we can write
But, from (4.17) we have
Thus, from (4.19) and since of (3.3) is a GCO matrix, we can get
Equation (4.18) can then be written as
Hence, we can conclude that, for , if(a), then and from Lemma 4.7 we have ,(b), then and from Lemma 4.7 we have ,(c), then and from Lemma 4.7 we have .
In consequence of the above, for and from Theorems 4.6 and 4.8, we have(i) if ,(ii) if ,(iii) if ,and the theorem is proved.
In view of Theorem 4.9, the superiority of the preconditioned MEDG SOR over the unpreconditioned MEDG SOR, EDG SOR methods and also preconditioned EDG SOR are confirmed for certain relaxation parameters lying in an optimum range.
5. Numerical Experiments and Discussion of Results
To further confirm the results obtained in Theorems 4.8 and 4.9, several experiments were carried out on the following model problem with Dirichlet boundary conditions: This problem has an exact solution with the unit square as the solution domain. The values of were calculated using different mesh sizes, 34, 86, 118, 186, and 222. The tolerance was set to be . The experimental optimum relaxation parameter was obtained by running the programs repeatedly and choosing the values which gave the fastest rate of convergence. The computer processing unit was Intel(R) Core(TM) 2Duo with memory of 3Gb and the software used to implement and generate the results was Developer C++ Version 4.9.9.2. Tables 1 and 2 display the corresponding number of iterations , optimum execution times , and the maximum errors for the unpreconditioned and preconditioned methods of EDG SOR and MEDG SOR, respectively.
From the results in Table 1, it is obvious that the original MEDG SOR method is superior to the EDG SOR method in terms of the number of iterations and computing times. The superiority of the preconditioned MEDG SOR over the preconditioned EDG SOR was also depicted in Table 2. The preconditioned EDG SOR was also outperformed by the unpreconditioned MEDG as shown in Figure 1 since the spectral radius of the latter is smaller than the former as proven in Theorem 4.9. From the numerical results, it is also apparent that the preconditioned MEDG SOR scheme requires the least computing effort amongst the four methods in terms of number of iterations and execution times due to its smallest spectral radius value amongst the four schemes.
Figure 1 shows the number of iterations needed for convergence for the unpreconditioned and preconditioned methods which were shown to be in agreement with the theoretical results obtained in Theorem 4.9.
6. Conclusion
In this paper, we present a theoretical convergence analysis of a specific splitting-type preconditioner in block formulation applied to the linear systems resulted from a class of group iterative schemes specifically the EDG SOR and the MEDG SOR schemes. We have shown that the spectral radius of the iteration matrix of the preconditioned MEDG SOR method is the smallest compared to the unpreconditioned MEDG SOR, EDG SOR, and preconditioned EDG SOR methods provided that the relaxation parameter . This work confirms the superiority of the preconditioned MEDG SOR method theoretically and experimentally in terms of convergence rates among this class of group iterative methods.
Acknowledgment
The authors acknowledge the Fundamental Research Grant Scheme (203/PMATHS/6711188) for the completion of this article.