On The Distance From A Normal Matrix Pol
On The Distance From A Normal Matrix Pol
Abstract
Consider an n × n matrix polynomial P (λ). An upper bound for a spectral norm
distance from P (λ) to the set of n×n matrix polynomials that have a given scalar µ ∈ C
as a multiple eigenvalue was recently obtained by Papathanasiou and Psarrakos (2008).
This paper concerns a refinement of this result for the case of weakly normal matrix
polynomials. A modification method is implemented and its efficiency is verified by
an illustrative example.
1 Introduction
Let A be an n × n complex matrix and µ be a complex number, and denote by Mµ the
set of n × n complex matrices that have µ ∈ C as a multiple eigenvalue. Malyshev [10]
obtained the following formula for the spectral norm distance from A to Mµ :
A − µI γIn
min kA − Bk2 = max s2n−1 ,
B∈Mµ γ≥0 0 A − µI
where k · k2 denotes the spectral matrix norm (i.e., that norm subordinate to the euclidean
vector norm) and s1 (·) ≥ s2 (·) ≥ s3 (·) ≥ · · · are the singular values of the corresponding
matrix in nonincreasing order. Malyshev’s work can be considered as a theoretical solution
to Wilkinson’s problem, that is, the calculation of the distance from a matrix A ∈ Cn×n
∗
Department of Mathematics, Faculty of Science, Yazd University, Yazd, Iran ([email protected],
[email protected]).
†
Department of Mathematics, National Technical University of Athens, Zografou Campus, 15780
Athens, Greece ([email protected]).
1
that has all its eigenvalues simple to the n × n matrices with multiple eigenvalues. Wilkin-
son introduced this distance in [17], and some bounds for it were computed by Ruhe [15],
Wilkinson [18–21] and Demmel [1].
However, in the non-generic case where A is a normal matrix, Malyshev’s formula is
not directly applicable. In 2004, Ikramov and Nazari [4] showed this point and obtained
an extension of Malyshev’s method for normal matrices. Moreover, Malyshev’s results
were extended by Lippert [9] and Gracia [3]; in particular, they computed a spectral norm
distance from A to the set of matrices that have two prescribed eigenvalues and studied
a nearest matrix with the two desired eigenvalues. Nazari and Rajabi [12] refined the
method obtained by Lippert and Gracia for the case of normal matrices.
In 2008, Papathanasiou and Psarrakos [14] introduced and studied a spectral norm
distance from a n × n matrix polynomial P (λ) to the set of n × n matrix polynomials
that have a scalar µ ∈ C as a multiple eigenvalue. In particular, generalizing Malyshev’s
methodology, they computed lower and upper bounds for this distance, constructing an
associated perturbation of P (λ) for the upper bound. Motivated by the above, in this note,
we study the case of weakly normal matrix polynomials. In the next section, we give some
definitions and present briefly some of the results of [13,14]. We also give an example of a
normal matrix polynomial where the method described in [14] for the computation of the
upper bound is not directly applicable. In Section 3, we prove that the methodology of [14]
for the computation of the upper bound is indeed not directly applicable to weakly normal
matrix polynomials, and in Section 4, we obtain a modified procedure for weakly normal
matrix polynomials to improve the method. The same numerical example is considered
to illustrate the validity of the proposed technique.
2 Preliminaries
For A0 , A1 , . . . , Am ∈ Cn×n , with det(Am ) 6= 0, and a complex variable λ, we define the
matrix polynomial
The study of matrix polynomials, especially with regard to their spectral analysis, has
received a great deal of attention and has been used in several applications [2, 6, 7, 11, 16].
Standard references for the theory of matrix polynomials are [2,11]. Here, some definitions
of matrix polynomials are briefly reviewed.
If for a scalar λ0 ∈ C and some nonzero vector x0 ∈ Cn , it holds that P (λ0 )x0 = 0,
then the scalar λ0 is called an eigenvalue of P (λ) and the vector x0 is known as a (right)
eigenvector of P (λ) corresponding to λ0 . The spectrum of P (λ), denoted by σ(P ), is
the set of all eigenvalues of P (λ). Since the leading matrix-coefficient Am is nonsingular,
2
the spectrum σ(P ) contains at most mn distinct finite elements. The multiplicity of an
eigenvalue λ0 ∈ σ(P ) as a root of the scalar polynomial det P (λ) is said to be the algebraic
multiplicity of λ0 , and the dimension of the null space of the (constant) matrix P (λ0 ) is
known as the geometric multiplicity of λ0 . Algebraic multiplicity of an eigenvalue is always
greater than or equal to its geometric multiplicity. An eigenvalue is called semisimple if
its algebraic and geometric multiplicities are equal; otherwise, it is known as defective.
Definition 2.1. Let P (λ) be a matrix polynomial as in (1). If there exists a unitary
matrix U ∈ Cn×n such that U ∗ P (λ)U is a diagonal matrix polynomial, then P (λ) is said
to be weakly normal. If, in addition, all the eigenvalues of P (λ) are semisimple, then P (λ)
is called normal.
The suggested references on weakly normal and normal matrix polynomials, and their
properties are [8, 13]. Some of the results of [13] are summarized in the next proposition.
(iv) There exists a unitary matrix U ∈ Cn×n such that U ∗ Aj U is diagonal for every
j = 0, 1, . . . , m.
where the matrices ∆0 , ∆1 , . . . , ∆m ∈ Cn×n are arbitrary. For a given parameter ǫ > 0
and a given set of nonnegative weights w = {w0 , w1 , . . . , wm } with w0 > 0, define the class
of admissible perturbed matrix polynomials
and the scalar polynomial w(λ) = wm λm + wm−1 λm−1 + · · · + w1 λ + w0 . Note that the
weights w0 , w1 , . . . , wm allow freedom in how perturbations are measured.
3
For any real number γ ∈ [0, +∞), we define the 2n × 2n matrix polynomial
P (λ) 0
F [P (λ); γ] = ,
γP ′ (λ) P (λ)
(2) the n × 2 matrices U (γ∗ ) = [u1 (γ∗ ) u2 (γ∗ )] and V (γ∗ ) = [v1 (γ∗ ) v2 (γ∗ )] satisfy
U ∗ (γ∗ )U (γ∗ ) = V ∗ (γ∗ )V (γ∗ ).
Moreover, it is remarkable that (1) implies (2) (see the proof of Lemma 17 in [14]).
′
(|µ|) µ̄
Consider the quantity φ = ww(|µ|) µ̄
|µ| , where, by convention, we set |µ| = 0 whenever
µ = 0. Let alsoV (γ∗ )† be the
Moore-Penrose
pseudoinverse of V (γ∗ ). For the pair of
u1 (γ∗ ) v1 (γ∗ )
singular vectors , ∈ C2n of Lemma 2.3, define the n × n matrix
u2 (γ∗ ) v2 (γ∗ )
1 −γ∗ φ
∆γ∗ = −s∗ U (γ∗ ) V (γ∗ )† .
0 1
Theorem 2.4. [14, Theorem 19] Let P (λ) be a matrix polynomial as in (1), and let
w = {w0 , w1 , . . . , wm }, with w0 > 0, be a set of nonnegative weights. Suppose that
µ ∈ C\σ(P ′ ), γ∗ > 0 is a point where the singular value s2n−1 (F [P (µ); γ]) attains its
maximumvalue,
and s∗ = s2n−1 (F [P (µ); γ∗ ]) > 0. Then, for the pair of singular vectors
u1 (γ∗ ) v1 (γ∗ )
, ∈ C2n of Lemma 2.3, we have
u2 (γ∗ ) v2 (γ∗ )
4
s∗
V (γ∗ ) 1 −γ∗ φ V (γ∗ )†
, w and has µ
lies on the boundary of the set B P, w(|µ|)
0 1
as a (multiple) defective eigenvalue.
Some numerical examples in Section 8 of [14] illustrate the effectiveness of the upper
bound of Theorem 2.4. In all these examples,
s∗ is a simple singular value, and con-
u1 (γ∗ ) v1 (γ∗ )
sequently, the singular vectors , ∈ C2n of Lemma 2.3 are directly
u2 (γ∗ ) v2 (γ∗ )
computable (due to their essential uniqueness). Let us now consider the normal (in par-
ticular, diagonal) matrix polynomial
1 0 0 −3 0 0 2 0 0
P (λ) = 0 1 0 λ2 + 0 −1 0 λ + 0 0 0 (4)
0 0 1 0 0 3 0 0 2
that is borrowed from [13, Section 3]. Let also the set of weights w = {1, 1, 1} and the scalar
µ = −4. The singular value s5 (F [P (−4); γ]) attains its maximum value at γ∗ = 2.0180,
and at this point, we have s∗ = s5 (F [P (−4); 2.0180]) = s4 (F [P (−4); 2.0180]) = 12.8841;
i.e., s∗ is a multiple singular value of matrix F [P (−4); 2.0180]. A left and a right singular
vectors of F [P (−4); 2.0180] corresponding to s∗ are
0 0
0.8407 0.5416
u1 (γ∗ ) 0 v1 (γ∗ ) 0
= and = ,
u2 (γ∗ )
0
v2 (γ∗ )
0
0.5416 0.8407
0 0
respectively, and they yield the perturbed matrix polynomial (see (3))
1 0 0 −3 0 0 2 0 0
Qγ∗ (λ) = 0 0.0664 0 λ2 + 0 −0.0664 0 λ + 0 −0.9336 0 .
0 0 1 0 0 3 0 0 2
One can see that µ = −4 is not a multiple eigenvalue of Qγ∗ (λ). Moreover, proper-
ties (1) and (2) of Lemma 2.3 do not hold since u∗2 (γ∗ )P ′ (µ)v1 (γ∗ ) = −2.6396 6= 0 and
kU ∗ (γ∗ )U (γ∗ ) − V ∗ (γ∗ )V (γ∗ )k2 = 0.4134 6= 0.
Clearly, this example verifies that the computation of appropriate singular vectors
which satisfy (1) and (2) of Lemma 2.3 is still an open problem when s∗ is a multiple
singular value. In the next section, we obtain that for weakly normal matrix polynomials,
s∗ is always a multiple singular value, and in Section 4, we solve the problem of calculation
of the desired singular vectors of Lemma 2.3.
5
3 Weakly normal matrix polynomials
In this section, by extending the analysis performed in [5], we prove that s∗ is always a
multiple singular value of F [P (µ); γ∗ ] when P (λ) is a weakly normal matrix polynomial.
Let P (λ) be a weakly normal matrix polynomial, and let µ ∈ C\σ(P ′ ). By Proposi-
tion 2.2 (iv), it follows that there exists a unitary matrix U ∈ Cn×n such that all matrices
U ∗ A0 U, U ∗ A1 U, . . . , U ∗ Am U are diagonal. Hence, U ∗ P (µ)U and U ∗ P (µ)′ U are also diag-
onal matrices; in particular,
where all scalars ξ1 , ξ2 , . . . , ξn ∈ C are nonzero (recall that P ′ (µ) is nonsingular) and,
without loss of generality, we assume that
As a consequence,
∗
U∗ 0
U 0 U 0 P (µ) 0 U 0
∗ F [P (µ); γ] =
0 U 0 U 0 U∗ ′
γP (µ) P (µ) 0 U
diag{ζ1 , ζ2 , . . . , ζn } 0
= .
γ diag{ξ1 , ξ2 , . . . , ξn } diag{ζ1 , ζ2 , . . . , ζn }
The fact that singular values of a matrix are invariant under unitary similarity implies
that the 2n × 2n matrices
ζ1 0 ζ2 0 ζn 0
F [P (µ); γ] and ⊕ ⊕ ··· ⊕
γ ξ1 ζ1 γ ξ2 ζ2 γ ξn ζn
6
ζi 0
For any i = 1, 2, . . . , n, let si,1 (γ) ≥ si,2 (γ) be the singular values of , and
γ ξi ζi
consider the characteristic polynomial of matrix
∗
|ζi |2 + γ 2 |ξi |2 γ ξ i ζi
ζi 0 ζi 0
= ,
γ ξi ζi γ ξi ζi γ ξi ζ i |ζi |2
that is,
|ζi |2 + γ 2 |ξi |2 γ ξ i ζi
det tI − = t2 − 2 |ζi |2 + γ 2 |ξi |2 t + |ζi |4 .
γ ξi ζ i |ζi |2
|ζi |2 + γ |ξi |2 γ ξ i ζi
The positive square roots of the eigenvalues of matrix are the
γ ξi ζ i |ζi |2
ζi 0
singular values of matrix , namely,
γ ξi ζi
v s
u
γ 2 |ξi |2 γ 2 |ξi |2
u
si,1 (γ) = |ζi |2 + + γ |ξi | |ζi |2 + ,
t
2 4
and
v s
u
γ 2 |ξi |2 γ 2 |ξi |2
u
2 2
si,2 (γ) = |ζi | + − γ |ξi | |ζi | + .
t
2 4
As γ ≥ 0 increases, si,1 (γ) increases and lim si,1 (γ) = +∞, while si,2 (γ) decreases and
γ→+∞
lim si,2 (γ) = 0 (recall that |ξi | > 0, i = 1, 2, . . . , n). Also, it is apparent that
γ→+∞
si,2 (γ) ≤ |ζi | ≤ si,1 (γ), and si,1 (0) = si,2 (0) = |ζi | .
7
decrease to 0. Let (γ0 , s0 ) be the first point in R2 where the graph of the increasing function
sn,1 (γ) intersects the graph of one of the n−1 decreasing functions s1,2 (γ), s2,2 (γ), . . . , sn−1,2 (γ),
say sκ,2 (γ) (for some κ ∈ {1, 2, . . . , n − 1}). Note that by the definition of si,1 (γ) and
si,2 (γ) (i = 1, 2, . . . , n), s0 lies in the open interval (0, |ζn−1 |) and the graph of sn,1 (γ)
cannot intersect the graph of one of the increasing functions s1,1 (γ), s2,1 (γ), . . . , sn−1,1 (γ)
for γ ≤ γ0 .
Since sn,2 (γ) and sκ,2 (γ) are both decreasing functions in γ ≥ 0, it follows that (see
Fig. 1 below, where κ = n − 1 = 2)
γ∗ = γ0 and s∗ = s0 = s2n−1 (F [P (µ); γ∗ ]) = sn,1 (γ∗ ) = sκ,2 (γ∗ ) = s2n−2 (F [P (µ); γ∗ ]).
Hence, when |ζn | < |ζn−1 |, γ∗ is the minimum positive root of one of the equations
sn,1 (γ) = sn−1,2 (γ), sn,1 (γ) = sn−2,2 (γ), . . . , sn,1 (γ) = s1,2 (γ)
sn,1 (0) = sn,2 (0) = sn−1,1 (0) = sn−1,2 (0) = |ζn | = |ζn−1 | ,
i.e.,
s2n (F [P (µ); 0]) = s2n−1 (F [P (µ); 0]) = s2n−2 (F [P (µ); 0]) = s2n−3 (F [P (µ); 0]) = |ζn | = |ζn−1 | .
Since sn,2 (γ) and sn−1,2 (γ)) are decreasing functions in γ ≥ 0, s2n−1 (F [P (µ); γ]) attains
its maximum value s∗ at γ = 0 = γ∗ , and s∗ is a multiple singular value of F [P (µ); 0]. In
this non-generic case, an upper bound and an associate perturbed matrix polynomial can
be computed by the method described in Section 6 of [14].
Hence, we have the following result.
Theorem 3.1. Let P (λ) in (1) be a weakly normal matrix polynomial, and let µ ∈
C\σ(P ′ ). If γ∗ > 0 is a point where the singular value s2n−1 (F [P (µ); γ]) attains its maxi-
mum value, then s∗ = s2n−1 (F [P (µ); γ∗ ]) > 0 is a multiple singular value of F [P (µ); γ∗ ].
8
4.1 The case of multiplicity 2
First we consider the case where γ∗ > 0 and the multiplicity of the singular value s∗ > 0
is equal to 2, and we work on the example of Section 2.
Recall that for the normal matrix polynomial P (λ) in (4) and for µ = −4, the singular
value s2n−1 (F [P (µ); γ]) = s5 (F [P (−4); γ]) attains its maximum value at γ∗ = 2.0180
and s∗ = s5 (F [P (−4); 2.0180]) = s4 (F [P (−4); 2.0180]) = 12.8841 (i.e., s∗ is a double
singular value of F [P (−4); 2.0180]). Two pairs of left and a right singular vectors of
F [P (−4); 2.0180] corresponding to s∗ , which do not satisfy properties (1) and (2) of Lemma
2.3 are
0 0
0.8407 0.5416
u1 (γ∗ ) 0 v1 (γ∗ ) 0
= , = ,
u2 (γ∗ )
0
v2 (γ∗ )
0
0.5416 0.8407
0 0
and
0 0
0
0
û1 (γ∗ ) −0.4222 v̂1 (γ∗ ) −0.9065
= , = .
û2 (γ∗ )
0
v̂2 (γ∗ )
0
0 0
0.9065 0.4222
In particular, we have
u2 (γ∗ )∗ P ′ (−4)v1 (γ∗ ) = −2.6396 6= 0 and û2 (γ∗ )∗ P ′ (−4)v̂1 (γ∗ ) = 4.1089 6= 0.
s2n−1 (F [P (µ); γ]) = s5 (F [P (−4); γ]) and s2n−2 (F [P (µ); γ]) = s4 (F [P (−4); γ])
are plotted for γ ∈ [0, 10], and their common point (γ∗ , s∗ ) = (2.0180, 12.8841) is marked
with “◦”. With respect to the discussion in the previous section, it is worth noting that
in this example, the graph of s2,2 (γ) (that is, sn−1,2 (γ)) is the graph of the decreasing
functions s1,2 (γ) and s2,2 (γ) that intersects first the graph of the increasing function s3,1 (γ)
(that is, sn,1 (γ)). Moreover, it is apparent that s2n−1 (F [P (µ); γ]) and s2n−2 (F [P (µ); γ])
are non-differentiable functions at γ∗ .
9
20
s2n−1(γ)
18 s (γ)
2n−2
16
14
o
12
(γ ,s )
* *
10
4
0 1 2 3 4 5 6 7 8 9 10
γ
Fig 1: The singular values s2n−1 (F [P (µ); γ]) (solid line) and s2n−2 (F [P (µ); γ]) (dashed line).
such that
where the scalars α, β ∈ C satisfy |α|2 + |β|2 = 1. By substituting the unknown singular
vectors of (5) into (6), we obtain
α
α β M = 0, (7)
β
where
u2 (γ∗ )∗ P ′ (µ)v1 (γ∗ ) u2 (γ∗ )∗ P ′ (µ)v̂1 (γ∗ )
M= . (8)
û2 (γ∗ )∗ P ′ (µ)v1 (γ∗ ) û2 (γ∗ )∗ P ′ (µ)v̂1 (γ∗ )
10
Proof. Recall that γ∗ and s∗ are positive. By the proof of Lemma 17 in [14], it follows
that the diagonal entries of matrix M are real.
By the definition of the pairs of singular vectors
u1 (γ∗ ) v1 (γ∗ ) û1 (γ∗ ) v̂1 (γ∗ )
, and ,
u2 (γ∗ ) v2 (γ∗ ) û2 (γ∗ ) v̂2 (γ∗ )
or equivalently,
P (µ)v1 (γ∗ ) = s∗ u1 (γ∗ ),
γ∗ P ′ (µ)v1 (γ∗ ) + P (µ)v2 (γ∗ ) = s∗ u2 (γ∗ ),
(9)
P (µ)v̂1 (γ∗ ) = s∗ û1 (γ∗ ),
γ∗ P ′ (µ)v̂1 (γ∗ ) + P (µ)v̂2 (γ∗ ) = s∗ û2 (γ∗ ),
and
∗ ∗ P (µ) 0
= s∗ v1 (γ∗ )∗ v2 (γ∗ )∗ ,
u1 (γ∗ ) u2 (γ∗ )
′
γ∗ P (µ) P (µ)
P (µ) 0
û1 (γ∗ )∗ û2 (γ∗ )∗ = s∗ v̂1 (γ∗ )∗ v̂2 (γ∗ )∗ ,
′
γ∗ P (µ) P (µ)
or equivalently,
By multiplying the fourth equation in (9) by u2 (γ∗ )∗ from the left, and the second
equation of (10) by v̂2 (γ∗ ) from the right, we obtain
γ∗ u2 (γ∗ )∗ P ′ (µ)v̂1 (γ∗ ) + u2 (γ∗ )∗ P (µ)v̂2 (γ∗ ) = s∗ u2 (γ∗ )∗ û2 (γ∗ ), (11)
and
11
respectively. As a consequence,
γ∗ u2 (γ∗ )∗ P ′ (µ)v̂1 (γ∗ ) = s∗ (u2 (γ∗ )∗ û2 (γ∗ ) − v2 (γ∗ )∗ v̂2 (γ∗ )) . (13)
γ∗ û2 (γ∗ )∗ P ′ (µ)v1 (γ∗ ) = s∗ (û2 (γ∗ )∗ u2 (γ∗ ) − v̂2 (γ∗ )∗ v2 (γ∗ )) . (14)
Clearly, equations (13) and (14) imply that the non-diagonal entries of matrix M are
complex conjugate.
By Lemma 2.3 (1), equation (7) has always a nontrivial (i.e., nonzero) solution, and
hence, the hermitian matrix M in (8) cannot be (positive or negative) definite. In our
numerical example, M has a negative and a positive diagonal entries (namely, −2.6396
and 4.1089), and thus, it is an indefinite hermitian matrix.
To derive an explicit solution of (7), suppose that η1 , η2 ∈ C are the (real) eigenvalues
of matrix M , with η1 > 0 > η2 , and let w1 , w2 ∈ C2 be unit eigenvectors of M corre-
sponding to η1 and η2 , respectively. Then, it is straightforward to see (keeping in mind
the orthogonality of the eigenvectors) that the unit vector
s s
α |η2 | |η1 |
= w1 + w2
β |η1 | + |η2 | |η1 | + |η2 |
satisfies
α |η1 | η2 |η2 | η1
α β M = + = 0.
β |η1 | + |η2 | |η1 | + |η2 |
Finally, in order to verify the validity of this refinement, we return again to the
normal matrix polynomial P (λ) in (4), and by applying the above methodology, we obtain
α = 0.6254 and β = 0.7803. Consequently, the desired vectors in (5) are (approximately)
0 0
0.6560 0.4226
ũ1 (γ∗ ) −0.2640 ṽ1 (γ∗ ) −0.5669
= and = .
ũ2 (γ∗ )
0
ṽ2 (γ∗ )
0
0.4226 0.6560
0.5669 0.2640
12
and for the n × 2 matrices Ũ (γ∗ ) = [ũ1 (γ∗ ) ũ2 (γ∗ )] and Ṽ (γ∗ ) = [ṽ1 (γ∗ ) ṽ2 (γ∗ )], we have
∗ ∗ −6
Ũ (γ )Ũ (γ ) − Ṽ (γ )Ṽ (γ∗
= 1.1383 · 10 .
)
∗ ∗ ∗
2
In addition, the lower bound 0.4031 of the distance is given by Theorem 11 in [14]. (All
computations were performed in Matlab with 16 significant digits.)
and
" # " # " #
(1) (2) (r)
v1 (γ∗ ) v1 (γ∗ ) v1 (γ∗ )
(1) , (2) , ..., (r)
v2 (γ∗ ) v2 (γ∗ ) v2 (γ∗ )
be orthonormal bases of the left and right singular subspaces of F [P (µ); γ∗ ] corresponding
to s∗ , respectively. Then, we are looking for a pair of unit vectors
r r
" # " #
X (j) X (j)
ũ1 (γ∗ ) u1 (γ∗ ) ṽ1 (γ∗ ) v1 (γ∗ )
= αj (j) , = αj (j) (15)
ũ2 (γ∗ ) u2 (γ∗ ) ṽ2 (γ∗ ) v2 (γ∗ )
j=1 j=1
13
such that
satisfies (17).
References
[1] J.W. Demmel, On condition numbers and the distance to the nearest ill-posed prob-
lem, Numer. Math., 51 (1987), 251–289.
[2] I. Gohberg, P. Lancaster and L. Rodman, Matrix Polynomials, Academic Press, New
York, 1982.
[3] J.M. Gracia, Nearest matrix with two prescribed eigenvalues, Linear Algebra Appl.,
401 (2005), 277–294.
14
[4] Kh.D. Ikramov and A.M. Nazari, Computational aspects of the use of Malyshev’s
formula, Comput. Math. Math. Phys., 44(1) (2004), 1–5.
[5] Kh.D. Ikramov and A.M. Nazari, On a remarkable implication of the Malyshev for-
mula, Dokl. Akad. Nauk., 385 (2002), 599–600.
[8] P. Lancaster and P. Psarrakos, Normal and seminormal eigenvalues of matrixp oly-
nomials. Integral Equations Operator Theory, 41 (2001), 331–342.
[9] R.A. Lippert, Fixing two eigenvalues by a minimal perturbation, Linear Algebra
Appl., 406 (2005), 177–200.
[10] A.N. Malyshev, A formula for the 2-norm distance from a matrix to the set of matrices
with a multiple eigenvalue, Numer. Math., 83 (1999), 443–454.
[11] A.S. Markus, Introduction to the Spectral Theory of Polynomial Operator Pencils,
Amer. Math. Society, Providence, RI, Translations of Mathematical Monographs,
Vol. 71, 1988.
[12] A.M. Nazari and D. Rajabi, Computational aspect to the nearest matrix with two
prescribed eigenvalues, Linear Algebra Appl., 432 (2010), 1–4.
[14] N. Papathanasiou and P. Psarrakos, The distance from a matrix polynomial to matrix
polynomials with a prescribed multiple eigenvalue, Linear Algebra Appl., 429 (2008),
1453–1477.
[16] F. Tisseur and K. Meerbergen, The quadratic eigenvalue problem, SIAM Rev., 43
(2001), 235–286.
[17] J.H. Wilkinson, The Algebraic Eigenvalue Problem, Claredon Press, Oxford, 1965.
[18] J.H. Wilkinson, Note on matrices with a very ill-conditioned eigenproblem, Numer.
Math., 19 (1972), 175–178.
15
[19] J.H. Wilkinson, On neighbouring matrices with quadratic elementary divisors, Nu-
mer. Math., 44 (1984), 1–21.
[21] J.H. Wilkinson, Sensitivity of eigenvalues II, Util. Math., 30 (1986), 243–286.
16