Abstract

An iterative algorithm is proposed for solving the least-squares problem of a general matrix equation , where () are to be determined centro-symmetric matrices with given central principal submatrices. For any initial iterative matrices, we show that the least-squares solution can be derived by this method within finite iteration steps in the absence of roundoff errors. Meanwhile, the unique optimal approximation solution pair for given matrices can also be obtained by the least-norm least-squares solution of matrix equation , in which . The given numerical examples illustrate the efficiency of this algorithm.

1. Introduction

Throughout this paper, we denote the set of all real matrices by . The symbol represents the transpose of matrix . and stand for the reverse unit matrix, and identity matrix, respectively. For , the inner product of matrices and is defined by , which leads to the Frobenius norm, that is, .

A matrix , is called centro-symmetric (centro-skew symmetric) if and only if which can also be characterized equivalently by . The set of all centro-symmetric (centro-skew symmetric) matrices is denoted by . This kind of matrices plays an important role in many applications (see, e.g., [14]), and has been frequently and widely investigated (see, e.g., [57]) by using generalized inverse, generalized singular value decomposition (GSVD) [8], and so forth. For more, we refer the readers to [916] and therein.

We firstly introduce the concept of the central principal submatrix which is originally put forward by Yin [17].

Definition 1. Let , if is even, a central principal submatrix of , denoted by , is obtained by deleting the first and last rows and columns of , namely, .

Evidently, a matrix with odd (even) order only has central principal submatrices of odd (even) order.

Now, the first problem to be studied here can be stated as follows.

Problem 2. Given , and (), . Find the least-squares solution of matrix equation in which with , , , and represents the central principal submatrix of .

Problem 2 is the submatrix constrained problem of matrix equation (2), which originally arises from a practical subsystem expansion problem, and has been deeply investigated (see, e.g., [7, 1822]). In these literatures, the generalized inverses or some complicated matrix decompositions such as canonical correlation decomposition (CCD) [23] and GSVD are employed. However, it is almost impossible to solve (2) by the above methods. The iterative method is an efficient approach. Recently, kinds of iteration methods have been constructed: Zhou and Duan [24] studied the generalized Sylvester matrix equation by so-called generalized Sylvester mapping that has pretty properties. Wu et al. [25] presented an finite iterative method for a class of complex matrix equations including conjugate and transpose of unknown solution. Motivated by the well-known Jacobi and Gauss-Seidel iterations methods, Ding and Chen, in [26], proposed a general family of iterative methods to solve linear matrix equations; meanwhile, these methods were also extended to solve the following coupled Sylvester matrix equations

Although these iterative algorithms are efficient, there still exist some handicaps when meeting the constrained matrix equation problem (i.e., to find the solution of matrix equation in some matrices sets with specifical structure, for instance, symmetric matrices, centro-symmetric matrices, and bi-symmetric matrices sets) and the submatrix constrained problem, since these methods cannot keep the special properties of the unknown matrix in the iterative process. Based on the classical conjugate gradient (CG) method, Peng et al. [27] gave an iterative method to find the bisymmetric solution of matrix equation (2). Similar method was constructed to solve matrix equations (4) with generalized bisymmetric in [28]. In particular, Li et al. [29] proposed an elegant algorithm for solving the generalized Sylvester (Lyapunov) matrix equation with bisymmetric and symmetric , the two unknown matrices include the given central principal submatrix and leading principal submatrix, respectively. This method shunned the difficulties in numerical instability and computational complexity, and solved the problem, completely. By borrowing the thinking of this iterative algorithm, we will solve Problem 2 by iteration method.

The second problem to be considered is the optimal approximation problem.

Problem 3. Let be the solutions set of Problem 2. For given matrices , find such that

This problem occurs frequently in experimental design (see for instance [30]). Here, the preliminary estimation of the unknown matrix can be obtained from experiments, but it may not satisfy the structural requirement and/or spectral requirement. The best estimation of , is the matrix that satisfies both requirements, which is the optimal approximation of (see, e.g., [31, 32]). About this problem, we also refer the authors to [911, 13, 15, 16, 2023, 2729, 3336] and therein.

The rest of this paper is outlined as follows. In Section 2, an iterative algorithm will be proposed to solve Problem 2, and the properties of which will be investigated. In Section 3, we will consider the optimal approximation Problem 3 by using the iterative algorithm. In Section 4, some numerical examples will be given to verify the efficiency of this algorithm.

2. The Algorithm for Problem 2 and Its Properties

According to the definition of centro-symmetric matrix, when is even, a centro-symmetric matrix can be divided into smaller submatrices, namely,where , , , , and .

Now, for some fixed positive integer , we define two matrix sets.

It is clear that both and are linear subspaces of .

In addition, for any matrix , it has uniquely decomposition in direct sum, that is, , here , . Furthermore, is also the direct sum decomposition of if , , since . Hence, we obtain the following.

Lemma 4. Consider .

Lemma 4 reveals that any matrix can be uniquely written as , where , , . Then, we can define the following linear projection operators: for .

According to the definition of , if and , we have This property will be employed frequently in the residual context.

The following theorem is essential for solving Problem 2, which transforms equivalently Problem 2 into solving the least-square problem of another matrix equation.

Theorem 5. Any solution group of Problem 2 can be obtained by where is the least-squares solution of matrix equation is the given central principal submatrix of in Problem 2.

Proof. Noting that the definition of , we have The proof is completed.

Remark 6. It follows, from Theorem 5, that Problem 2 can be solved completely by finding the least-squares solution of matrix equations (11) in subspaces .

In the next part of this section, we will establish an iterative algorithm for (11) and analysis its properties. For the convenience of expression, we define a matrix function then matrix equation (11) can be simplified as Moreover, we can easily verify that holds for arbitrary .

The iterative algorithm for the least squares problem of matrix equations (11) can be expressed as follows.

Algorithm 7. Consider the following.
Step 1. Let , , and for .
Input arbitrary matrices .
Step 2. Calculate
Step 3. Calculate
Step 4. Calculate
Step 5. If , stop. Otherwise, , go to Step 3.

From Algorithm 7, we can see that . In particular, is a least-squares solution group belonging to matrix equation (11) if for all . The following lemma gives voice to the reason.

Lemma 8. If hold simultaneously for some positive , then generated by Algorithm 7 is a solution group of matrix equation (11).

Proof. Let and . Obviously, . Then, from the Project Theorem, is a least-square solution group of matrix equation (11) if and only if . That is to say, for any matrices , noting that Lemma 4, we have , that is, which completes the proof.

In addition, the sequences , , generated by Algorithm 7 are self-orthogonal, that is, as follows.

Lemma 9. Suppose that the sequences , , generated by Algorithm 7 not equal null for , then where , .

Proof. In view of the symmetry of the inner product, we only prove (20)–(22) when . According to Algorithm 7, when , we have which also deduces that
Assume that (20), (21), and (22) hold for positive integer (), that is, for , Then, similar to the above proof, noting that the assumptions, we have Furthermore,
The last equal sign “” holds due to . In fact, from the hypothesis, we deduce Moreover, It follows from (29) that Therefore, for arbitrary integers number , the conclusions hold. The proof is completed.

Remark 10. We know from Lemma 4 that the matrices sequences are orthogonal to each other. Hence, it can be regarded as an orthogonal basis of matrix space . Hence, the iteration will be terminated at most steps in the absence of roundoff errors. Therefore, there exists a positive integer such that , in this case, can be regarded as a least-squares solution group of matrix equation (11).
In addition, we should point out that if or , the conclusions may not be true, and the iteration will break down before for . Actually, implies that , so for . While leads to , making inner product with by both sides, it follows from Algorithm 7 that which also implies the same situation as . Hence, if there exists a positive integer such that the coefficient or , then the corresponding matrix group is just the solution of matrix equation (11).

Together with the above analysis and Lemma 9, we can conclude the following theorem.

Theorem 11. For any initial iteration matrices , , the least-squares solution of matrix equation (11) can be obtained within finite iteration steps. Moreover, Suppose that is a least-squares solution group of (11), then the general solution to Problem 2 can be expressed as , where satisfy homogeneous equation as in Theorem 5.

In order to show the validity of Theorem 11, it is adequate to prove the following conclusion.

Proposition 12. The least-squares solution group of matrix equation (11) can be expressed as , where satisfy equality (33).

Proof. According to the assumptions, we obtain On the other hand, noting that , then The proof is completed.

Next, we will show that the unique least norm solution of matrix equation (11) can be derived by choosing a special kind of initial iteration matrices.

Theorem 13. Let the initial iteration matrices with arbitrary , , and then generated by Algorithm 7 is the least-norm least-squares solution group of matrix equation (11). Furthermore, the least-norm solution group to Problem 2 can be expressed by

Proof. From Algorithm 7 and Theorem 11, for initial iteration matrices , we can obtain a least-squares solution of matrix equation (11) and there exists a matrix such that . Hence, it is enough to prove that the is the least-norm solution. In fact, noting that (33) and Proposition 12, we have as required.

Theorems 11 and 13 display the efficiency of Algorithm 7. Actually, the iteration sequence converges smoothly to the solution , that is the minimization property of Algorithm 7.

Theorem 14. For any initial iteration matrices , the generated by Algorithm 7 satisfy the minimization problem where , .

Proof. From the definition of , there exist a series of real numbers such that .
Define a function of variables , that is, In addition, from Algorithm 7, we know that Noting that (22) and making the inner product with on both sides of (40) yield Hence, by simple calculation, (40) and (41), the function can be rewritten as Then,
Since only if , it follows from (29) that Combined with (43), we complete the proof.

Theorem 14 reveals that the sequence monotonically decreases with respect to increasing integer . The descent property of the residual norm of matrix equation (11) leads to the smoothly convergence of Algorithm 7.

3. The Solution of Problem 3

In this section, we discuss the optimal approximation Problem 3. Since the least squares problem is always consistent, it is easy to verify that the solution set of Problem 2 is a nonempty convex cone, so the optimal approximation solution is unique.

Without loss of generality, we can assume that the given matrices . In fact, from Lemma 4, arbitrary can be divided into Furthermore, if , then which meets the claim.

Denote , , then to solve Problem 3 is equivalent to find the least-norm solution of the new matrix equation Furthermore, similar to the construction of (11), Problem 2 is transformed equivalently into finding the least-norm least-squares solution of matrix equation in which .

Therefore, we can apply Algorithm 7 to derive the required solution of matrix equation (49). Virtually, it follows from Theorem 13 that if let the initial iteration matrices with arbitrary , or especially , then the iteration solutions consist of the least-norm least-squares solution of which. In this case, the unique optimal approximation solution to Problem 3 can be obtained by

4. Numerical Example

In this section, we illustrate the efficiency and reliability of Algorithm 7 by some numerical experiments.

All the numerical experiments are performed by using Matlab 6.5. In addition, because of the influence of the roundoff errors, may not equal zero within finite iteration steps, so the iteration will be terminated if , for example, let . At this time, can be regarded as a solution of matrix equation (11), and () consist of the solution group to Problem 2. In particular, let the initial iteration matrices , then we will obtain the least-norm solution by (36).

Example 1. Input matrices , , , , , , and as follows: where toeplitz, hilb, hankel, zeros and eye denote the Toeplitz matrix, Hilbert matrix, Hankel matrix, null matrix, identity matrix with orders , and the elements of matrix are one, represents tri-diagonal matrix produced by vector .

Let the given central principal matrices By using the Algorithm 7, we obtain the solution to Problem 2. To save space, we shall not report the explicit datum of the solution, but the bars graphs of the components for the solution matrices will be given. Let , Figure 1 shows the bars graphs of , , when we choose the initial iterative matrices and the terminal condition .

Moreover, when and , the convergence curves for the Frobenius norm of the residual denoted by and the termination condition denoted by are plotted in Figures 2 and 3, respectively.

From Figure 2, we can see that the residual norm of Algorithm 7 is monotonically decreasing, which is in accordance with the theory established in Theorem 14, namely, this algorithm is numerical stable. While Figure 3 shows that the terminated condition is oscillating back and forth and approaches to zero as iterative process. Hence, the iterative Algorithm 7 is efficient, but it lacks of smooth convergence. Of course, for a problem with large and sparse matrices, Algorithm 7 may not terminate in a finite number of steps because of roundoff errors. How to establish an efficient and smooth algorithm is an important problem which we should study in a future work.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The authors would like to express their sincere gratitude to the editor and two anonymous reviewers for their valuable comments and suggestions which have helped immensely improving the quality of the paper. Mao-lin Liang acknowledges the support of the scientific foundation of Tianshui Normal University (no. TSA1104). Young-hong Shen is supported by the “QingLan” Talent Engineering Funds of Tianshui Normal University.