Abstract

We give a sufficient condition (the solvability of two standard equations) of Sylvester matrix by using the displacement structure of the Sylvester matrix, and, according to the sufficient condition, we derive a new fast algorithm for the inversion of a Sylvester matrix, which can be denoted as a sum of products of two triangular Toeplitz matrices. The stability of the inversion formula for a Sylvester matrix is also considered. The Sylvester matrix is numerically forward stable if it is nonsingular and well conditioned.

1. Introduction

Let 𝑅[𝑥] be the space of polynomials over the real numbers. Given univariate polynomials 𝑓(𝑥),𝑔(𝑥)𝑅[𝑥], 𝑎10, where𝑓(𝑥)=𝑎1𝑥𝑛+𝑎2𝑥𝑛1++𝑎𝑛+1,𝑎10,𝑔(𝑥)=𝑏1𝑥𝑚+𝑏2𝑥𝑚1++𝑏𝑚+1,𝑏10.(1.1)

Let 𝑆 denote the Sylvester matrix of 𝑓(𝑥) and 𝑔(𝑥)𝑎𝑆=1𝑎2𝑎𝑛+1𝑎1𝑎2𝑎𝑛+1𝑎1𝑎2𝑎𝑛+1𝑏1𝑏2𝑏𝑚𝑏𝑚+1𝑏1𝑏2𝑏𝑚𝑏𝑚+1𝑏1𝑏2𝑏𝑚𝑏𝑚+1𝑚row𝑛row(1.2)

Sylvester matrix is applied in many science and technology fields. The solutions of Sylvester matrix equations and matrix inequations play an important role in the analysis and design of control systems. In determining the greatest common divisor of two polynomials, the Sylvester matrix plays a vital role, and the magnitude of the inverse of the Sylvester matrix is important in determining the distance to the closest polynomials which have a common root. Assuming that all principal submatrices of the matrix are nonsingular, in [1], Jing Yang et al. have given a fast algorithm for the inverse of Sylvester matrix by using the displacement structure of 𝑚+𝑛-order Sylvester matrix.

By using the displacement structure of the Sylvester matrix, in this paper, we give a sufficient condition (the solvability of two standard equations) of Toeplitz matrix, and, according to the sufficient condition, we derive a new fast algorithm for the inversion of a Sylvester matrix, which can be denoted as a sum of products of two triangular Toeplitz matrices. At last, the stability of the inversion formula for a Sylvester matrix is also considered. The Sylvester matrix is numerically forward stable if it is nonsingular and well conditioned.

In this paper, 2 always denotes the Euclidean or spectral norm and 𝐹 the Frobenius norm.

2. Preliminary Notes

In this section, we present a lemma that is important to our main results.

Lemma 2.1 (see [2, Section 2.4.8]). Let 𝐴,𝐵𝑛,𝑛 and 𝛼. Then for any floating-point arithmetic with machine precision 𝜀, one has that 𝑓𝑙(𝛼𝐴)=𝛼𝐴+𝐸,𝐸𝐹𝜀|𝛼|𝐴𝐹𝜀𝑛|𝛼|𝐴2,𝑓𝑙(𝐴+𝐵)=𝐴+𝐵+𝐸,𝐸𝐹𝜀𝐴+𝐵𝐹𝜀𝑛𝐴+𝐵2,𝑓𝑙(𝐴𝐵)=𝐴𝐵+𝐸,𝐸𝐹𝜀𝑛𝐴𝐹𝐵𝐹.(2.1) As usual, one neglects the errors of O(𝜀2), O(̃𝜀2), and O(𝜀̃𝜀).

3. Sylvester Inversion Formula

In this section, we present our main results.

Theorem 3.1. Let matrix 𝑆 be a Sylvester matrix; then it satisfies the formula 𝐾𝑆𝑆𝐾=𝑒𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇,(3.1) where 𝐾=(0𝑒1𝑒2𝑒𝑚+𝑛1) is a (𝑚+𝑛)×(𝑚+𝑛) shift matrix, 𝑏𝑓=1𝑏𝑚𝑏𝑚+1𝑎1𝑎𝑛𝑇,𝑔=00𝑏1𝑏𝑚𝑇.(3.2)

Proof. We have that 𝐾𝑆𝑆𝐾=0𝑎1𝑎𝑛+1𝑎001𝑎𝑛+1𝑏01𝑏𝑚+1𝑏001𝑏𝑚+100000𝑎1𝑎𝑛+1000𝑎1𝑎𝑛𝑏1𝑏𝑚+100𝑏1𝑏𝑚=𝑏00001𝑏𝑚𝑏𝑚+1𝑎1𝑎2𝑎𝑛00000𝑏1𝑏𝑚𝑒𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇=𝑏00001𝑏𝑚𝑏𝑚+1𝑎1𝑎2𝑎𝑛00000𝑏1𝑏𝑚.(3.3) So 𝐾𝑆𝑆𝐾=𝑒𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇.

Theorem 3.2. Let matrix 𝑆 be a Sylvester matrix and 𝑥=(𝑥1𝑥2𝑥𝑚+𝑛)𝑇, 𝑦=(𝑦1𝑦2𝑦𝑚+𝑛)𝑇, 𝜇=(𝜇1𝜇2𝜇𝑚+𝑛)T, and 𝑉=(𝑉1𝑉2𝑉𝑚+𝑛)T the solutions of the systems of equations 𝑆𝑥=𝑒𝑚, 𝑆𝑦=𝑒𝑚+𝑛, 𝑆𝑇𝜇=𝑓, and 𝑆𝑇𝑉=𝑔, respectively, where 𝑒𝑚 and 𝑒𝑚+𝑛 are both (𝑚+𝑛)×1 vectors; then (a)𝑆 is invertible, and the column vector 𝑤𝑗(𝑗=1,2,,𝑚+𝑛) of 𝑆1 satisfies the recurrence relation𝑤𝑚+𝑛𝑤=𝑦,𝑖1=𝐾𝑤𝑖+𝜇𝑖𝑥𝑉𝑖𝑦,𝑖=𝑚+𝑛,,3,2,(3.4)(b)  𝑆1=𝑆1𝑈1+𝑆2𝑈2=𝑦𝑚+𝑛𝑦𝑚+𝑛1𝑦2𝑦1𝑦𝑚+𝑛𝑦3𝑦2𝑦𝑚+𝑛𝑦𝑚+𝑛1𝑦𝑚+𝑛1𝑉𝑚+𝑛1𝑉3𝑉41𝑉2𝑉3𝑉𝑚+𝑛1+𝑥𝑚+𝑛𝑥𝑚+𝑛1𝑥2𝑥1𝑥𝑚+𝑛𝑥3𝑥2𝑥𝑚+𝑛𝑥𝑚+𝑛1𝑥𝑚+𝑛0𝜇𝑚+𝑛0𝜇3𝜇4𝜇02𝜇3𝜇𝑚+𝑛0.(3.5)

Proof. By Theorem 3.1  𝑆𝑥=𝑒𝑚 and 𝑆𝑦=𝑒𝑚+𝑛, we have that 𝐾𝑆=𝑆𝐾+𝑒𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇=𝑆𝐾+𝑆𝑥𝑓𝑇𝑆𝑦𝑔𝑇=𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇,(3.6) so 𝐾𝑖𝑆=𝐾𝑖1𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇=𝐾𝑖2𝐾𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇=𝐾𝑖2𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇2=𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇𝑖.(3.7) Hence, we have that 𝐾𝑖𝑒𝑚+𝑛=𝐾𝑖𝑆𝑦=𝑆𝐾+𝑥𝑓𝑇𝑦𝑔𝑇𝑖𝑦,𝑖=0,1,,𝑚+𝑛1.(3.8) Let 𝑤𝑚+𝑛𝑖=𝐾+𝑥𝑓𝑇𝑦𝑔𝑇𝑖𝑦,𝑖=0,1,,𝑚+𝑛1.(3.9) It is easy to see that 𝐾𝑖𝑒𝑚+𝑛=𝑒𝑚+𝑛𝑖, and by (3.8) 𝑆𝑤𝑚+𝑛𝑖=𝐾𝑖𝑒𝑚+𝑛=𝑒𝑚+𝑛𝑖,𝑖=0,1,,𝑚+𝑛1.(3.10) From 𝑆𝑦=𝑒𝑚+𝑛, we have that 𝑤𝑚+𝑛=𝑦. Let 𝑋=(𝑤1𝑤2𝑤𝑚+𝑛); then 𝑤𝑆𝑋=𝑆1𝑤2𝑤𝑚+𝑛=𝑆𝑤1𝑆𝑤2𝑆𝑤𝑚+𝑛=𝐼𝑚+𝑛,(3.11) so the matrix 𝑆 is invertible and the inverse of 𝑆 is the matrix 𝑋.
From (3.1), we have that𝑆1(𝐾𝑆𝑆𝐾)𝑆1=𝑆1𝑒𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇𝑆1,(3.12) and thus 𝑆1𝐾𝐾𝑆1=𝑥𝜇𝑇𝑦𝑉𝑇.(3.13)
So𝑆1𝐾𝐾𝑆1𝑒𝑖=𝑥𝜇𝑇𝑦𝑉𝑇𝑒𝑖.(3.14)
Since  𝐾𝑒𝑖=𝑒𝑖1, we have that𝑤𝑖1=𝐾𝑠𝑖+𝜇𝑖𝑥𝑉𝑖𝑦,𝑖=𝑚+𝑛,,3,2,(3.15) and hence 𝑤𝑚+𝑛𝑤=𝑦,𝑖1=𝐾𝑤𝑖+𝜇𝑖𝑥𝑉𝑖𝑦,𝑖=𝑚+𝑛,,3,2.(3.16)
For (b), by  (3.4)𝑤𝑚+𝑛𝑤=𝑦,𝑚+𝑛1=𝐾𝑦+𝜇𝑚+𝑛𝑥𝑉𝑚+𝑛𝑤𝑦,𝑚+𝑛2=𝐾2𝑦+𝐾𝜇𝑚+𝑛𝑥𝐾𝑉𝑚+𝑛𝑦+𝜇𝑚+𝑛1𝑥𝑉𝑚+𝑛1𝑤𝑦,1=𝐾𝑚+𝑛1𝑦+𝐾𝑚+𝑛2𝜇𝑚+𝑛𝑥𝐾𝑚+𝑛2𝑉𝑚+𝑛𝑦++𝜇2𝑥𝑉2𝑦.(3.17) So 𝑆1=𝑤1𝑤2𝑤𝑚+𝑛=𝐾𝑚+𝑛1𝑦𝐾𝑚+𝑛21𝑦𝐾𝑦𝑦𝑉𝑚+𝑛1𝑉3𝑉41𝑉2𝑉3𝑉𝑚+𝑛1+𝐾𝑚+𝑛1𝑥𝐾𝑚+𝑛20𝜇𝑥𝐾𝑥𝑥𝑚+𝑛0𝜇3𝜇4𝜇02𝜇3𝜇𝑚+𝑛0=𝑦𝑚+𝑛𝑦𝑚+𝑛1𝑦2𝑦1𝑦𝑚+𝑛𝑦3𝑦2𝑦𝑚+𝑛𝑦𝑚+𝑛1𝑦𝑚+𝑛1𝑉𝑚+𝑛1𝑉3𝑉41𝑉2𝑉3𝑉𝑚+𝑛1+𝑥𝑚+𝑛𝑥𝑚+𝑛1𝑥2𝑥1𝑥𝑚+𝑛𝑥3𝑥2𝑥𝑚+𝑛𝑥𝑚+𝑛1𝑥𝑚+𝑛0𝜇𝑚+𝑛0𝜇3𝜇4𝜇02𝜇3𝜇𝑚+𝑛0.(3.18) This completes the proof.

4. Stability Analysis

In this section, we will show that the Sylvester inversion formula presented in this paper is evaluation forward stable.

If, for all well conditioned problems, the computed solution ̃𝑥 is close to the true solution 𝑥, in the sense that the relative error 𝑥̃𝑥2/𝑥2 is small, then we call the algorithm forward stable (the author called this weakly in [3]). Round-off errors will occur in the matrix computation.

Theorem 4.1. Let matrix 𝑆 be a nonsingular Sylvester matrix and well conditioned; then the formula in Theorem 3.2 is forward stable.

Proof. Assume that we have computed the solutions ̃𝑥,̃𝑦,𝜇, and 𝑉 of 𝑆𝑥=𝑒𝑚, 𝑆𝑦=𝑒𝑚+𝑛, 𝑆𝑇𝜇=𝑓, and 𝑆𝑇𝑉=𝑔 in Theorem 3.2 which are perturbed by the normwise relative errors bounded by ̃𝜀, ̃𝑥2𝑥2(1+̃𝜀),̃𝑦2𝑦2(1+̃𝜀),𝜇2𝜇2𝑉(1+̃𝜀),2𝑉2(1+̃𝜀).(4.1)
Thus, we have that𝑆1𝐹𝑚+𝑛𝑦2,𝑆2𝐹𝑚+𝑛𝑥2,𝑈1𝐹𝑚+𝑛1+𝑉22,𝑈2𝐹𝑚+𝑛𝜇2.(4.2)
The inversion formula in Theorem 3.2, using the perturbed solutions ̃𝑥, ̃𝑦, 𝜇, and 𝑉, can be expressed as𝑆1𝑆=𝑓𝑙1𝑈1+𝑆2𝑈2𝑆=𝑓𝑙1+Δ𝑆1𝑈1+Δ𝑈1+𝑆2+Δ𝑆2𝑈2+Δ𝑈2=𝑆1+Δ𝑆1𝑈1+𝑆1Δ𝑈1+Δ𝑆2𝑈2+𝑆2Δ𝑈2+𝐸+𝐹.(4.3) Here, and in the sequel, 𝐸 is the matrix containing the error which results from computing the matrix products and 𝐹 contains the error from subtracting the matrices. For the error matrices Δ𝑆1, Δ𝑈1, Δ𝑆2, and Δ𝑈2, we have that Δ𝑆1𝐹𝑆̃𝜀1𝐹̃𝜀𝑚+𝑛𝑦2,Δ𝑆2𝐹𝑆̃𝜀2𝐹̃𝜀𝑚+𝑛𝑥2,Δ𝑈1𝐹𝑈̃𝜀1𝐹̃𝜀𝑚+𝑛1+𝑉22,Δ𝑈2𝐹𝑈̃𝜀2𝐹̃𝜀𝑚+𝑛𝜇2.(4.4)
By Lemma 2.1, we have the following bounds on 𝐸 and 𝐹:𝐸2𝐸𝐹𝑆𝜀(𝑚+𝑛)1𝐹𝑈1𝐹+𝑆2𝐹𝑈2𝐹𝜀(𝑚+𝑛)2𝑦21+𝑉22+𝑥2𝜇2𝜀(𝑚+𝑛)2𝑦21+𝑉22+𝑥2𝜇2𝐹2𝑆𝑚+𝑛𝜀12.(4.5)
Consequently, adding all these error bounds, by (4.3), we have that𝑆1𝑆12(𝑚+𝑛)(2̃𝜀+(𝑚+𝑛)𝜀)𝑦21+𝑉22+𝑥2𝜇2+𝑆𝑚+𝑛𝜀12.(4.6)
From the equations 𝑆𝑥=𝑒𝑚, 𝑆𝑦=𝑒𝑚+𝑛𝑆𝑇𝜇=𝑓, and 𝑆𝑇𝑉=𝑔 in Theorem 3.2, we have that𝑦2𝑆12,𝑉2𝑆12𝑔2,𝑥2𝑆12,𝜇2𝑆12𝑓2.(4.7)
Thus, the relative error is𝑆1𝑆12𝑆12(𝑚+𝑛)(2̃𝜀+(𝑚+𝑛)𝜀)𝑦21+𝑉22+𝑥2𝜇22𝑆12+𝜀𝑆𝑚+𝑛(𝑚+𝑛)(2̃𝜀+(𝑚+𝑛)𝜀)1+12𝑔2+𝑓2+𝜀𝑚+𝑛.(4.8)
Since 𝑆 is well conditioned, 𝑆12 is finite. It is easy to see that 𝑔2,𝑓2 are finite. Therefore, the formula presented in Theorem 3.2 is forward stable.
This completes the proof.

5. Numerical Example

This section gives an example to illustrate our results. All the following tests are performed by MATLAB 7.0.

Example 5.1. Given 𝑓(𝑥)=𝑥+1,𝑔(𝑥)=𝑥2+𝑥+1, that is, 𝑎1=1,𝑎2=1,𝑏1=1,𝑏2=1,𝑏3=1,𝑛=1,𝑚=2,𝑓=(1,1,0)T,𝑔=(0,1,1)T, and 𝑆=110011111.
So==,𝑒𝐾𝑆𝑆𝐾=010001000110011111110011111010001000011111000011001011000110011(5.1)𝑚𝑓𝑇𝑒𝑚+𝑛𝑔𝑇=010001=110011000110011.(5.2) Therefore, 𝐾𝑆𝑆𝐾=𝑒𝑚𝑓T𝑒𝑚+𝑛𝑔T.
By the condition of Theorem 3.2, we can get𝑥=(1,1,0)T,𝑦=(1,1,1)T,𝜇=(1,0,0)T,𝑉=(0,1,0)T.(5.3) Obviously, 𝑆 is invertible and 𝑆1=011111101. And it is easy to see that 𝑤3=𝑦, 𝑤1=01=10101110100010001+011=𝑆𝑤2+𝜇2𝑥𝑉2𝑤𝑦,2=10=11101110100010001+0101=𝑆𝑤3+𝜇3𝑥𝑉3𝑆𝑦,1==11+0001111110111111011010110100000=𝑆1𝑈1+𝑆2𝑈2.(5.4)

Acknowledgments

This work was supported financially by the Youth Foundation of Shandong Province, China (ZR2010EQ014), Shandong Province Natural Science Foundation (ZR2010AQ026), and National Natural Science Foundation of China (11026047).