* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Let v denote a column vector of the nilpotent matrix Pi(A)(A − λ iI)ni
Covariance and contravariance of vectors wikipedia , lookup
System of linear equations wikipedia , lookup
Matrix completion wikipedia , lookup
Rotation matrix wikipedia , lookup
Capelli's identity wikipedia , lookup
Determinant wikipedia , lookup
Matrix (mathematics) wikipedia , lookup
Linear least squares (mathematics) wikipedia , lookup
Principal component analysis wikipedia , lookup
Orthogonal matrix wikipedia , lookup
Singular-value decomposition wikipedia , lookup
Jordan normal form wikipedia , lookup
Non-negative matrix factorization wikipedia , lookup
Eigenvalues and eigenvectors wikipedia , lookup
Ordinary least squares wikipedia , lookup
Four-vector wikipedia , lookup
Gaussian elimination wikipedia , lookup
Matrix multiplication wikipedia , lookup
Cayley–Hamilton theorem wikipedia , lookup
Let v denote a column vector of the nilpotent matrix Pi (A)(A − λi I)ni −1 where ni is the so called nilpotency. Theorem 3 in [1] shows that APi (A)(A − λi I)ni −1 = λi Pi (A)(A − λi I)ni −1 . which means a column vector v of the matrix is an eigenvector corresponding to the eigenvalue λi . The symbols are explained in [1]. However it is worth noting that Pi (A)(A − λi I). That is 6 0 Pi (A)(A − λi I)ni −1 = ni but Pi (A)(A − λi I) = 0. We now use the formula for At as derived in [2] t A = k X Pi (A)λti i=1 m i −1 X j=0 Γ(t + 1) A − λi I j ( ) , assuming I t = I j!Γ(t − j + 1) λi (0.1) t to compute A v by, k X At Ps (A)(A − λs I)ns −1 = [ Pi (A)λti i=1 = Ps (A)λts m i −1 X j=0 m i −1 X j=0 by Ps (A)Pi (A) = Ps (A)δis Γ(t + 1) A − λi I j ( ) ][Ps (A)(A − λs I)ns −1 ] j!Γ(t − j + 1) λi Γ(t + 1) 1 j+ns −1 ] j (A − λs I) j!Γ(t − j + 1) λs Note that by the property of the nilpotency, ns , Ps (A)(A − λs I)ns = 0. Therefore the only nonzero term in the last sum is the leading term corresponding to j = 0. It follows then At Ps (A)(A − λs I)ns −1 = λts Ps (A)(A − λs I)ns −1 1 which proves for each column vector, v, of the matrix Ps (A)(A − λs I)ns −1 the result you mentioned At v =λts v. We will use Example 3 in [1] to help explain the notations used here and to illustrate the result. Let ⎡ ⎤ −3 5 −5 ⎥ A=⎢ ⎣ 3 −1 3 ⎦ . 8 −8 10 The characteristic polynomial is p(λ) = (λ − 2)3 and only one projection matrix is P1 (A) = I so that At = 2t I + t2t−1 (A − 2I) + t(t − 1) t−2 2 (A − 2I)2 . 2 More explicitly ⎡ ⎤ 2 − 5t 5t −5t 3t 2 − 3t 3t ⎥ At = 2t−1 ⎢ ⎣ ⎦. 8t −8t 2 + 8t The nilpotent matrix, Pi (A)(A − λi I) in this case is ⎡ ⎤ ⎡ ⎤2 −5 5 −5 ⎢ A − 2I = ⎣ 3 −3 3 ⎥ ⎦ and 8 −8 8 (A − 2I)2 −5 5 −5 ⎢ ⎥ = ⎣ 3 −3 3 ⎦ = 0 8 −8 8 That means ni = 2 and then Ps (A)(A − λs I)ns −1 = A − 2I 2 an eigenvector is ⎡ ⎤ −5 ⎢ v=⎣ 3 ⎥ ⎦ 8 If we choose t = 1/2 in the above formula for At , we find ⎡ − 12 1 3 A1/2 = √ ⎢ ⎣ 2 2 4 ⎡ 5 2 1 2 −5 2 3 2 −4 6 ⎤ ⎥ ⎦ ⎤ −1 5 −5 1 ⎢ 1 3 ⎥ = √ ⎣ 3 ⎦ 2 2 8 −8 12 ⎡ ⎤⎡ ⎤ −1 5 −5 −5 1 ⎢ ⎥⎢ 1/2 √ 1 3 ⎦⎣ 3 ⎥ A v = ⎣ 3 ⎦ 2 2 8 −8 12 8 ⎡ ⎤ ⎡ ⎤ −20 √ ⎢ −5 ⎥ 1 ⎥ 12 = √ ⎢ = 2⎣ 3 ⎦ ⎣ ⎦ 2 2 32 8 1/2 = λ1 v. References [1] D. V. Ho, The power & exponential of a matrix, www.math.gatech.edu/~ho [2] D. V. Ho, Real power and logarithm of a matrix, www.math.gatech.edu/~ho 3