 ORIGINAL ARTICLE
 Open Access
 Published:
A tridiagonal matrix construction by the quotient difference recursion formula in the case of multiple eigenvalues
Pacific Journal of Mathematics for Industry volume 6, Article number: 10 (2014)
Abstract
In this paper, we grasp an inverse eigenvalue problem which constructs a tridiagonal matrix with specified multiple eigenvalues, from the viewpoint of the quotient difference (qd) recursion formula. We also prove that the characteristic and the minimal polynomials of a constructed tridiagonal matrix are equal to each other. As an application of the qd formula, we present a procedure for getting a tridiagonal matrix with specified multiple eigenvalues. Examples are given through providing with four tridiagonal matrices with specified multiple eigenvalues.
1 Introduction
One of the important problems in linear algebra is to construct matrices with specified eigenvalues. This is an inverse eigenvalue problem which is classified in Structured Inverse Eigenvalue Problem (SIEP) called in [1]. The main purpose of this paper is to design a procedure for solving an SIEP in the case where the constructed matrix has tridiagonal form with multiple eigenvalues, through reconsidering the quotient difference (qd) formula. It is known that the qd formula has the applications to computing a continued fraction expansion of power series [5], zeros of polynomial [3], eigenvalues of a socalled Jacobi matrix [9] and so on. Though the book [9] refers to an aspect similar to in the following sections, it gives only an anticipated comment without proof in the case of multiple eigenvalues. There is no observation about numerical examples for verifying it. The key point for the purpose is to investigate the Hankel determinants appearing in the determinant solution to the qd formula with the help of the Jordan canonical form. In this paper, we give our focus on the unsettled case in order to design a procedure for constructing a tridiagonal matrix with specified multiple eigenvalues, based on the qd formula. The reason why the sequence of discussions was stopped is expected that multipleprecision arithmetic and symbolic computing around the published year of Rutishauser’s works for the qd formula were not sufficiently developed. The qd formula, strictly speaking the differential form of it, for computing tridiagonal eigenvalues acts with high relative accuracy in singleprecision or doubleprecision arithmetic [7], while, actually, that serving for constructing a tridiagonal matrix gives rise to no small errors. Thus, the qd formula serving for constructing a tridiagonal matrix is not so worth in singleprecision or doubleprecision arithmetic. In recent computers, it is not difficult to employ not only single or double precision arithmetic but also arbitraryprecision arithmetic or symbolic computing. In fact, an expression involving only symbolic quantities achieves exact arithmetic on the scientific computing software such as Wolfram Mathematica, Maple and so on. Numerical errors frequently occur in finiteprecision arithmetic, so that a constructed tridiagonal matrix probably does not have multiple eigenvalues without symbolic computing. The resulting procedure in this paper is assumed to be carried out on symbolic computing.
This paper is organized as follows. In Section 2, we first give a short explanation of some already obtained properties concerning the qd formula. In Section 3, we observe a tridiagonal matrix whose characteristic polynomial is associated with the minimal polynomial of a general matrix through reconsidering the qd formula. The tridiagonal matrix essentially differs from the Jacobi matrix in that it is not always symmetrized. We also discuss the characteristic and the minimal polynomials of a tridiagonal matrix in Section 4. In Section 5, we design a procedure for constructing a tridiagonal matrix with specified multiple eigenvalues, and then demonstrate four tridiagonal matrices as examples of the resulting procedure. Finally, in Section 6, we give conclusion.
2 Some properties for the qd recursion formula
In this section, we briefly review two theorems in [4] concerning the qd formula from the viewpoint of a generating function, the Hankel determinant and a tridiagonal matrix.
Let us introduce the Hankel determinants \(H_{1}^{(n)},H_{2}^{(n)},\dots \) given in terms of a complex sequence \(\{\,f_{n}\}_{0}^{\infty }\) as
where \(H_{1}^{(n)}=0\) and \(H_{0}^{(n)}=1\) for n=0,1,…. Moreover, let F(z) be a generating function associated with \(\{\,f_{n}\}_{0}^{\infty }\) as
Let us consider that F(z) is a rational function with respect to z with a pole of order l_{0}≥0 at infinity and finite poles z_{ k }≠0 of order l_{ k } for k=1,2,…,L. Then the sum of the orders of the finite poles is l=l_{1}+l_{2}+⋯+l_{ L }, and F(z) is factorized as
where G(z) is a polynomial of degree at most l, and G_{0}(z) is a polynomial of degree l_{0} if l_{0}>0, or G_{0}(z)=0 if l_{0}=0. The following theorem gives the determinant solution to the qd recursion formula
Theorem1.
([4], pp. 596, 603, 610) Let F(z) be factorized as in (3). Then it holds that
Let us assume that
Then the qd formula (4) with the initial settings
admits the determinant solution
From (9) with (5), it follows that \(e_{l}^{(n)}=0\) for n=0,1,…. Moreover, it turns out that \(q_{s}^{(n)}\) and \(e_{s}^{(n)}\) for s=l+1,l+2,… and n=0,1,… are not given in the same form as (8) and (9).
Let us introduce sbys tridiagonal matrices,
with the qd variables \(q_{s}^{(n)}\) and \(e_{s}^{(n)}\). Let I_{ s } be the sbys identity matrix. Then we obtain a theorem for the characteristic polynomial of \(T_{l}^{(n)}\).
Theorem2.
([4], pp. 626, 635) Let F(z) be factorized as in (3). Let us assume that \(H_{s}^{(n)}\) satisfies (6). For n=0,1,…, it holds that
3 Tridiagonal matrix associated with general matrix
In this section, from the viewpoint of the characteristic and the minimal polynomials, we associate a general MbyM complex matrix A with a tridiagonal matrix \(T_{l}^{(n)}\).
Let λ_{1},λ_{2},…,λ_{ N } be the distinct eigenvalues of A, which are numbered as λ_{1}≥λ_{2}≥⋯≥λ_{ N }. It is noted that some of λ_{1},λ_{2},…,λ_{ N } may equal to each other in the case where some of λ_{1},λ_{2},…,λ_{ N } are negative eigenvalues or complex eigenvalues. Let M_{ k } be the algebraic multiplicity of λ_{ k }, where M=M_{1}+M_{2}+⋯+M_{ N }. For the identity matrix \(I_{M}\in \mathbb {R}^{M\times M}\), let ϕ_{ A }(z)= det(zI_{ M }−A) be the characteristic polynomial of A, namely,
Let us prepare the sequence \(\{\,f_{n}\}_{0}^{\infty }\) given by
for some nonzero Mdimensional complex vectors u and w, where the superscript H denotes the Hermitian transpose. Originally, f_{0},f_{1},… were called the Schwarz constants, but they are usually today called the moments or the Markov parameters [2]. Since the matrix power series \(\sum _{n=0}^{\infty }(zA)^{n}\) is a Neumann series (cf. [6]), \(F(z)=\sum _{n=0}^{\infty }\boldsymbol {w}^{H}(zA)^{n}\boldsymbol {u}\) converges absolutely in the disk D:z<λ_{1}^{−1}. Moreover, we derive F(z)=w^{H}(I_{ M }−zA)^{−1}u which implies that F(z) is a rational function with the denominator det(I_{ M }−zA)=z^{M}ϕ_{ A }(z^{−1}) as follows.
where \(\tilde {G}(z)\) is some polynomial with respect to z. It is remarkable that the numerator \(\tilde {G}(z)\) may have the same factors as the denominator \(\phantom {\dot {i}\!}(1\lambda _{1}z)^{M_{1}}(1\lambda _{2}z)^{M_{2}}\cdots (1\lambda _{N}z)^{M_{N}}\). In other words, F(z) has the poles \(\lambda _{1}^{1},\lambda _{2}^{1},\dots,\lambda _{N}^{1}\) whose orders are equal to or less than M_{1},M_{2},…,M_{ N }, respectively.
Let us introduce the Jordan canonical form of A in order to investigate the poles of F(z) with (13) even in the case where A has multiple eigenvalues. Let \({\mathcal {M}}_{k}\) be the geometric multiplicity of λ_{ k } which indicates the dimension of eigenspace Ker(A−λ_{ k }I_{ M }). It is noted that \({\mathcal {M}}_{k}\) is equal to or less than the algebraic multiplicity M_{ k }. The matrix A has \({\mathcal {M}}_{k}\) eigenvectors corresponding to λ_{ k }, and then the eigenvectors, denoted by \(\boldsymbol {v}_{k,1},\boldsymbol {v}_{k,2},\dots,\boldsymbol {v}_{k,{\mathcal {M}}_{k}}\), satisfy
Hereinafter, for \(j=1,2,\dots,{\mathcal {M}}_{k}\), let v_{k,j}(1)=v_{k,j}. Moreover, for \(j=1,2,\dots,{\mathcal {M}}_{k}\), let v_{k,j}(2), v_{k,j}(3), …, v_{k,j}(m_{k,j}) denote the generalized eigenvectors associated with the eigenvectors v_{k,j}(1), where m_{k,j} is the maximal integer such that v_{k,j}(1), v_{k,j}(2), …, v_{k,j}(m_{k,j}) are linearly independent. Of course, \(m_{k,1}+m_{k,2}+\cdots +m_{k,{\mathcal {M}}_{k}}=M_{k}\). Then, the generalized eigenvectors v_{k,j}(2),v_{k,j}(3),…,v_{k,j}(m_{k,j}) satisfy
From (15) and (16), we derive the Jordan canonical form of A as
with the nonsingular matrix
and the block diagonal matrix
where
Without loss of generality, we may assume that \(m_{k,1}\ge m_{k,2}\ge \cdots \ge m_{k,{\mathcal {M}}_{k}}\).
Let \(m_{k}=\max \{m_{k,1},m_{k,2},\dots,m_{k,{\mathcal {M}}_{k}}\}\). Since \(m_{k,1}\ge m_{k,2}\ge \cdots \ge m_{k,{\mathcal {M}}_{k}}\), it is obvious that m_{ k }=m_{k,1}. With the help of the Jordan canonical form of A as in (17), we get a proposition for the sequence \(\{\,f_{n}\}_{0}^{\infty }\) in (13).
Proposition1.
Let u be the vector given by the linear combination of the eigenvectors and the generalized eigenvectors of A, namely, for some constants κ_{k,j,i},
Moreover, for a vector w, let
Then, the sequence \(\{f_{n}\}_{0}^{\infty }\) in (13) can be expressed by
where the binomial coefficients are 0 if n<i−1. Also, for suitable u and w, it holds that
Proof.
From V^{−1}AV=J in (17), it holds that A^{n}=VJ^{n}V^{−1}. By combining it with (13) and (24), we derive
Let ρ_{k,j,i} be the column number in which v_{k,j}(i) arranges. Then it is obvious that V^{−1}v_{k,j}(i)=e_{k,j}(i) where e_{k,j}(i) denotes a unit vector such that the ρ_{k,j,i}th entry is 1 and the others are 0. Thus, it follows that
Since J is the block diagonal matrix, the matrix J^{n} and its small blocks (J_{ k })^{n} are also so. It also turns out that (J_{k,j})^{n} is upper triangle. So, it is worth noting that J^{n}e_{k,j}(i) becomes the ρ_{k,j,i}th column vector of J^{n} and the zeroentries arrange in except for its ρ_{k,j,1}th, ρ_{k,j,2}th, …, ρ_{k,j,i}th rows. The Jordan blocks J_{k,j} can be decomposed as
It is emphasized that \(E_{m_{k,j}}\) is a nilpotent matrix whose i^{′}th power becomes the zeromatrix O for i^{′}≥m_{k,j}. Thus, (J_{k,j})^{n} can be expressed as
where \((E_{m_{k,j}})^{0}=I_{m_{k,j}}\). Let us introduce an m_{k,j}dimensional unit vector e(i) which is regarded as a part of e_{k,j}(i). Then, by taking account that \((E_{m_{k,j}})^{i^{\prime }1}\boldsymbol {e}(i)=\boldsymbol {e}(ii^{\prime }+1)\) in (32), we derive
Since it holds that Ve_{k,j}(i−i^{′}+1)=v_{k,j}(i−i^{′}+1), by combining it with (29) and (33), we therefore have
By writing down two summations, we get
Moreover, by paying our attention to the binomial coefficients, we can rewrite f_{ n } as
From m_{ k }≥m_{k,j} and \(\boldsymbol {w}^{H}\boldsymbol {v}_{k,j}(ii^{\prime }+1)=\boldsymbol {v}^{H}_{k,j}(ii^{\prime }+1)\boldsymbol {w}\), it follows that
The exchange of i for i^{′} in (35) brings us to (25) and (26).
For example, let us consider the case where the constants κ_{k,j,i} are all 1. Then u becomes the sum of all the eigenvectors and generalized eigenvectors. Moreover, let w=V^{−H}α in (25) where α is an Mdimensional vector with all the entries 1. Then it holds that \(\kappa _{k,j,i}\boldsymbol {v}_{k,j}^{H}(i^{\prime }i+1)\boldsymbol {w}=\boldsymbol {e}_{k,j}^{\top }(i^{\prime }i+1)\boldsymbol {\alpha }=1\). Thus, it is concluded that c_{k,i}≠0. The above discussion suggests that there exists at least a pair of u and w for satisfying (27).
Proposition 1 leads to a theorem concerning the generating function F(z) with the moments f_{ n }=w^{H}A^{n}u.
Theorem3.
Let F(z) be the generating function with the moments f_{ n }=w^{H}A^{n}u. Then, F(z) converges absolutely in the disk D:z<λ_{1}^{−1}, and F(z) is expressed as
Especially, if λ_{ N }=0, then F(z) is expressed as
Let us assume that (27) holds for suitable u and w. If λ_{ N }≠0, then F(z) has the finite poles \(\lambda _{1}^{1},\lambda _{2}^{1},\dots,\lambda _{N}^{1}\) of the orders m_{1},m_{2},…,m_{ N }, respectively, and the sum of the orders is m=m_{1}+m_{2}+⋯+m_{ N }. If λ_{ N }=0, then F(z) has the pole of the order m_{ N }−1 at infinity and the finite poles \(\lambda _{1}^{1},\lambda _{2}^{1},\dots,\lambda _{N1}^{1}\) of the orders m_{1},m_{2},…,m_{ N }, respectively, and the sum of the orders of all the finite poles is m−m_{ N }.
Proof.
By substituting f_{ n } in (26) into F(z) in (2), we get
By letting n=n^{′}+i−1, we derive
It is noted that, for z<1,
From (39) and (40), it turns out that F(z) converges absolutely in the disk D:z<λ_{1}^{−1}. Simultaneously, we have (36) for z∈D. It is obvious that (36) with λ_{ N }=0 becomes (37). Moreover, (36) and (37) immediately lead to the latter half concerning the poles of F(z).
Let ψ_{ A }(z) be the polynomial whose degree is the smallest such that ψ_{ A }(A)=O. Here ψ_{ A }(z) is called the minimal polynomial of A. Let us recall here that the maximal dimension of the Jordan blocks \(J_{k,1},J_{k,2},\dots,J_{k,{\mathcal {M}}_{k}}\) corresponding to λ_{ k } is m_{ k }. So, ψ_{ A }(z) is representable as
Therefore, we have the main theorem in this section for the relationship between the minimal polynomial of a general matrix A and the characteristic polynomial of a tridiagonal matrix \(T_{l}^{(n)}\).
Theorem4.
Let F(z) be given by the generating function with the moments f_{ n }=w^{H}A^{n}u. Let us assume that (6) and (27) hold for suitable u and w. If λ_{1}≠0,λ_{2}≠0,…,λ_{ N }≠0, then it holds that
otherwise,
Proof.
It is remarkable that three integers L,l,l_{ k } and a complex z_{ k } associated with the tridiagonal matrix \(T_{l}^{(n)}\) in Theorem 2 are given in terms of three integers N,m,m_{ k } and a complex λ_{ k } associated with a general matrix A. If λ_{ N }≠0, then it follows from the latter half of Theorem 3 that L=N, l=m, l_{0},l_{1}=m_{1},l_{2}=m_{2},…,l_{ N }=m_{ N } and \(z_{k}=\lambda _{k}^{1}\). So, from (11) and (41), we derive (42). Similarly, if λ_{ N }=0, then L=N−1, l=m−m_{ N }, l_{0}=m_{ N }−1,l_{1}=m_{1},l_{2}=m_{2},…,l_{N−1}=m_{N−1} and \(z_{k}=\lambda _{k}^{1}\). Thus (11) and (41) lead to (43).
Incidentally, the editors in ([9], pp. 444–445) give a simple example with short comments concerning the minimal polynomial, the Jordan canonical form of A and the multiple poles of F(z).
4 Minimal polynomial of tridiagonal matrix
In this section, with the help of the Jordan canonical form, we clarify the relationship of the characteristic polynomial of the tridiagonal matrix \(T_{l}^{(n)}\) to the minimal one.
For simplicity, let us here adopt the following abbreviations for matrices \(T_{s}^{(n)}\),
where l=m if λ_{ N }≠0 or l=m−m_{ N } if λ_{ N }=0. Let p_{0}(z)=1 and p_{ s }(z)= det(zI_{ s }−T_{ s }) for s=1,2,…,l. Then p_{ l }(z) is just the characteristic polynomial of T_{ l }, namely,
where L=N if λ_{ N }≠0 or L=N−1 if λ_{ N }=0. The following proposition gives the Jordan canonical form of the tridiagonal matrix T_{ l }.
Proposition2.
There exists a nonsingular matrix P such that
where J_{1,1},J_{2,1},…,J_{L,1} are of the same form as (23).
Proof.
The characteristic polynomials p_{0}(z), p_{1}(z), …, p_{ l }(z) satisfy
This is easily derived from the expansion of det(zI_{ s }−T_{ s }) by the sth row minors. By taking the 0th, the 1st, …, the (m_{ k }−1)th derivatives with respect to z in (48), we get
where D^{i}p_{ s }(z) denotes the ith derivative of p_{ s }(z) with respect to z. Let \(\boldsymbol {p}_{k,i}=(D^{i}p_{0}(\lambda _{k}),D^{i}p_{1}(\lambda _{k}),\dots,D^{i}p_{l1}(\lambda _{k}))^{\top }\in \mathbb {C}^{l}\). Then, by substituting z=λ_{ k } in (49) and by taking account that \(\phantom {\dot {i}\!}D^{i}p_{l}(\lambda _{k})=D^{i}(z\lambda _{1})^{m_{1}}(z\lambda _{2})^{m_{2}}\cdots (z\lambda _{l})^{m_{l}}_{z=\lambda _{k}}=0\) for i=0,1,…,m_{ k }−1, we obtain
Moreover, it follows that
where P_{k,i}=(1/i!)p_{k,i}. Thus, by letting \(P=(P_{1,0}\,P_{1,1}\,\cdots \,P_{1,m_{1}1}\,\,P_{2,0}\,P_{2,1}\cdots \,P_{2,m_{2}1}\,\cdots \,\,P_{L,0}\,P_{L,1}\,\cdots \,P_{L,m_{L}1})\in \mathbb {C}^{l\times l}\), we have \((T_{l})^{\top }P=P\hat {J}\).
Here, it remains to prove that P is nonsingular. Of course, P_{k,i}≠O since the (i+1)th row of P_{k,i} is D^{i}p_{ i }(λ_{ k })/i!=1. Let W_{k,i}=Ker((T_{ l })^{⊤}−λ_{ k }I_{ l })^{i} for i=1,2,…m_{ k }−1, which indicates the generalized eigenspace of (T_{ l })^{⊤} corresponding to λ_{ k }. Then it is obvious from (51) that ((T_{ l })^{⊤}−λ_{ k }I_{ l })P_{k,0}=O and P_{k,0}∈W_{k,1}. Eq. (51) with i=1 also leads to that ((T_{ l })^{⊤}−λ_{ k }I_{ l })^{2}P_{k,1}=O and P_{k,1}∈W_{k,2}. Simultaneously, it is observed that P_{k,1}∉W_{k,1}. Let us assume that P_{k,1}∈W_{k,1}, namely, (T_{ l })^{⊤}P_{k,1}=λ_{ k }P_{k,1}. Then, from (51), we derive P_{k,0}=O, which contradicts with P_{k,0}≠O. Thus, it follows that P_{k,1}∉W_{k,1}. Similarly, by induction for i=2,3,…,m_{ k }−1 in P_{k,i}, we have
From (52), it turns out that P_{k,i} for i=0,1,…,m_{ k }−1 and k=1,2,…,L are linearly independent. Therefore, it is concluded that P is nonsingular and the Jordan canonical form of (T_{ l })^{⊤} is given by (46).
Proposition 2 suggests that the minimal polynomial of (T_{ l })^{⊤} becomes
which is equal to the characteristic polynomial of T_{ l } in (45). If m_{1}=m_{2}=⋯=m_{ L }=1, then it is obvious that T_{ l } is diagonalizable. Otherwise, T_{ l } is not diagonalizable. This is because multiplicity of roots in minimal polynomial coincides with maximal size of the Jordan blocks. To sum up, we have a theorem for the properties of the tridiagonal matrix T_{ l }.
Theorem5.
The minimal polynomial of T_{ l } is equal to the characteristic one. Also, T_{ l } is diagonalizable tridiagonal matrix if and only if it has no multiple eigenvalues.
5 Procedure for constructing tridiagonal matrix and its examples
In this section, based on the discussions in the previous sections, we first design a procedure for constructing a tridiagonal matrix with specified multiple eigenvalues. We next give four kinds of examples for demonstrating that the resulting procedure can provide with tridiagonal matrices with multiple eigenvalues. Examples have been carried out with our computer with OS: Mac OS X 10.8.5, CPU: Intel Core i7 2 GHz, RAM: 8 GB. We also use the scientific computing software Wolfram Mathematica 9.0. In every example, all the entries of u are simply set to 1 and those of w are not artificial. The readers will realize that the settings of u and w are not so difficult for satisfying (6) and (27).
Let us here consider the relationship of five theorems in the previous sections. Theorem 2 shows that the eigenvalues of \(T_{l}^{(n)}\) in the tridiagonal form as (10) are equal to the poles of the generating function F(z) and the multiplicity of the eigenvalues coincide with the those of the poles of F(z). Theorems 3 and 4 claim that the minimal polynomial of a general matrix A, denoted by ψ_{ A }(z) is just the denominator of F(z) involving f_{ n }=w^{H}A^{n}u, and it coincides with the characteristic polynomial of \(T_{l}^{(n)}\) denoted by ϕ_{ T }(z), except for the factor corresponding to zeroeigenvalues. With the help of Theorem 1, we thus realize that the nonzero eigenvalues of \(T_{l}^{(0)}\) with the entries involving \(q_{1}^{(0)},q_{2}^{(0)},\dots,q_{l}^{(0)}\) and \(e_{1}^{(0)},e_{2}^{(0)},\dots,e_{l1}^{(0)}\) become roots of the minimal polynomial ψ_{ A }(z) in the case where \(q_{1}^{(0)},q_{2}^{(0)},\dots,q_{l}^{(0)}\) and \(e_{1}^{(0)},e_{2}^{(0)},\dots,e_{l1}^{(0)}\) are given by the qd formula (4) under the initial settings \(e_{0}^{(n)}=0\) and \(q_{1}^{(n)}=f_{n+1}/f_{n}\) with f_{ n }=w^{H}A^{n}u. See also Figure 1 for the diagram for getting \(q_{s}^{(n)}\) and \(e_{s}^{(n)}\) by the qd formula (4). A procedure for constructing \(T=T_{l}^{(0)}\) with the same nonzero eigenvalues as A is therefore as follows.

1:
Set l=m if λ_{ N }≠0 or l=m−m_{ N } if λ_{ N }=0.
 2:

3:
Compute f_{ n }=w^{H}A^{n}u for n=0,1,…,2l−1.

4:
Set \(e_{0}^{(n)}=0\) for n=0,1,…,2l−3.

5:
Compute \(q_{1}^{(n)}=f_{n+1}/f_{n}\) for n=0,1,…,2l−2.

6:
Repeat (a) and (b) for s=2,3,…,l.

(a)
Compute \(e_{s1}^{(n)}=q_{s1}^{(n+1)}+e_{s2}^{(n+1)}q_{s1}^{(n)}\) for n=0,1,…,2l−2s+1.

(b)
Compute \(q_{s}^{(n)} = q_{s1}^{(n+1)}e_{s1}^{(n+1)}/e_{s1}^{(n)}\) for n=0,1,…,2l−2s.

(a)

7:
Construct a tridiagonal matrix by arranging \(q_{1}^{(0)},q_{2}^{(0)},\dots,q_{l}^{(0)}\) and \(e_{1}^{(0)},e_{2}^{(0)},\dots,e_{l1}^{(0)}\).
According to Theorem 5, the minimal and the characteristic polynomials of the resulting tridiagonal matrix T are equal to each other. Moreover, T is diagonalizable if and only if it has no multiple eigenvalues.
It is necessary to control the eigenvalues of A for getting T as a tridiagonal matrix with specified eigenvalues. It is easy to specify the eigenvalues of the diagonal matrix and those of the Jordan matrix.
First, in the procedure, let us consider the case where
which is a diagonal matrix with two eigenvalues 1 and 2 each of multiplicity 3. Obviously, the characteristic and the minimal polynomials are factorized as (z−1)^{3}(z−2)^{3} and (z−1)(z−2), respectively. So, the integers l and m are immediately determined as l=2 and m=6. Moreover, by letting u=(1,1,1,1,1,1)^{⊤} and w=(1,1,1,1,1,1)^{⊤}, we derive a tridiagonal matrix as
whose characteristic and minimal polynomials are both factorized as (z−1)(z−2). The tridiagonal matrix T is a diagonalizable matrix with the distinct eigenvalues 1 and 2.
Next, let us adopt a bidiagonal matrix, which can be regarded as the Jordan matrix, with eigenvalues 2 of multiplicity 6 as A, namely,
in the procedure. Since the characteristic polynomial of A is equal to the minimal one, the integers l and m are determined as l=m=6. Then the procedure with u=(1,1,1,1,1,1)^{⊤} and w=(1,1,0,1,0,1)^{⊤} constructs a tridiagonal matrix, which can not be symmetrized,
The characteristic and the minimal polynomials of A and T are all the same polynomial with respect to z, which is factored as (z−2)^{6}. So, the tridiagonal matrix T is not diagonalizable.
Let us prepare the Jordan matrix
The matrix A has multiple eigenvalues such as λ_{1}=3, λ_{2}=3, λ_{3}=3, λ_{4}=3, λ_{5}=3, λ_{6}=3, λ_{7}=2, λ_{8}=2. It is noted that λ_{1}=λ_{2}=λ_{3}=λ_{4}=λ_{5}=λ_{6}>λ_{7}=λ_{8}>0. The characteristic and the minimal polynomials of A are factorized as (z−2)^{2}(z−3)^{6} and (z−2)^{2}(z−3)^{3}, respectively. So, let l=5 and m=8 in the procedure. Then, the settings u=(1,1,1,1,1,1,1,1)^{⊤} and w=(1,1,1,1,1,1,1,1)^{⊤} bring us to a tridiagonal matrix, which can not be symmetrized,
whose characteristic and minimal polynomials are both factorized as (z−2)^{2}(z−3)^{3}, which is just equal to the minimal one of A. The tridiagonal matrix T is not a diagonalizable matrix with eigenvalues 2 and 3 of multiplicity 2 and 3, respectively.
Finally, let us A be set as the Jordan matrix with complex eigenvalues 2+i and 2−i each of multiplicity 2 and distinct real eigenvalues 1 and 2, namely,
in this procedure. By taking account that the characteristic and the minimal polynomials of A are equal to each other, let l=m=6 in the procedure. Under the settings u=(1,1,1,1,1,1)^{⊤} and w=(1,1,1,1,1,1)^{⊤}, the resulting matrix T is a real tridiagonal matrix, which can not be symmetrized,
The characteristic and the minimal polynomials of A and T are all the same polynomial with respect to z, which is factorized as (z−2+i)^{2}(z−2−i)^{2}(z−2)(z−1). So, the tridiagonal matrix T is not a diagonalizable matrix with the same complex multiple eigenvalues and real distinct ones as A.
6 Conclusion
In this paper, we clarify that the qd recursion formula is applicable to constructing a tridiagonal matrix with specified multiple eigenvalues. We first investigate the denominator of the generating function associated with the sequence given from two suitable vectors and the powers of a general matrix A, through considering the Jordan canonical form of A. Accordingly, it is observed that the minimal polynomial of A coincides with the characteristic polynomial of a tridiagonal matrix T, denoted by ϕ_{ T }(z), or the polynomial \(\phantom {\dot {i}\!}z^{m_{L}}\phi _{T}(z)\) for the multiplicity m_{ L } of the zeroeigenvalues of A. Next, by taking account of the Jordan canonical form of T, we show that the characteristic and the minimal polynomials of T are equal to each other. We finally present a procedure for constructing a tridiagonal matrix with specified multiple eigenvalues, and then give four examples for the resulting procedure.
References
Chu, MT., Golub, GH.: Inverse Eigenvalue Problem: Theory, Algorithms, and Applications. Oxford University Press, New York (2005).
Gutknecht, MH., Parlett, BN.: From qd to LR, or, how were the qd and LR algorithms discovered?IMA J. Numer. Anal. 31, 741–754 (2011).
Henrici, P., Watkins, BO.: Finding zeros of a polynomial by the QD algorithm. Commut. ACM. 8, 570–574 (1965).
Henrici, P.: Applied and, Computational Complex Analysis, Vol. 1. John Wiley, New York (1974).
Henrici, P.: Applied and Computational Complex Analysis, Vol. 2. John Wiley, New York (1977).
Meyer, CD.: Matrix Analysis and Applied Linear Algebra. SIAM, Philadelphia (2000).
Parlett, BN.: The new qd algorithms. Acta Numerica. 4, 459–491 (1995).
Rutishauser, H.: Bestimmung der Eigenwerte und Eigenvektoren einer Matrix mit Hilfe des QuotientenDifferenzenAlgorithmus. Z. Angew. Math. Phys. 6, 387–401 (1955).
Rutishauser, H.: Lectures on Numerical Mathematics. Birkhäuser, Boston (1990).
Acknowledgements
The authors would like to thank the reviewer for his/her careful reading and beneficial suggestions. This work is supported by JSPS KAKENHI Grant Number 23654032.
Author information
Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Akaiwa, K., Iwasaki, M., Kondo, K. et al. A tridiagonal matrix construction by the quotient difference recursion formula in the case of multiple eigenvalues. Pac. J. Math. Ind. 6, 10 (2014). https://doi.org/10.1186/s4073601400100
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1186/s4073601400100
Keywords
 Quotient difference formula
 Tridiagonal matrix
 Multiple eigenvalues
 Characteristic polynomial
 Minimal polynomial