* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Section 7.2
Factorization wikipedia , lookup
Tensor operator wikipedia , lookup
Capelli's identity wikipedia , lookup
System of linear equations wikipedia , lookup
Bra–ket notation wikipedia , lookup
Cartesian tensor wikipedia , lookup
Quadratic form wikipedia , lookup
Rotation matrix wikipedia , lookup
Invariant convex cone wikipedia , lookup
Basis (linear algebra) wikipedia , lookup
Fundamental theorem of algebra wikipedia , lookup
Symmetry in quantum mechanics wikipedia , lookup
Linear algebra wikipedia , lookup
Determinant wikipedia , lookup
Matrix (mathematics) wikipedia , lookup
Non-negative matrix factorization wikipedia , lookup
Four-vector wikipedia , lookup
Gaussian elimination wikipedia , lookup
Matrix calculus wikipedia , lookup
Singular-value decomposition wikipedia , lookup
Matrix multiplication wikipedia , lookup
Perron–Frobenius theorem wikipedia , lookup
Cayley–Hamilton theorem wikipedia , lookup
EIGENVALUES & EIGENVECTORS We say λ is an eigenvalue of a square matrix A if Ax = λx for some x = 0. The vector x is called an eigenvector of A, associated with the eigenvalue λ. Note that if x is an eigenvector, then any multiple αx is also an eigenvector. EXAMPLES: −7 13 −16 1 1 13 −10 13 −1 = −36 −1 −16 13 −7 1 1 1 0 −1 1 1 0 1 = 0 1 1 −1 −1 1 0 1 1 Knowing the eigenvalues and eigenvectors of a matrix A will often give insight as to what is happening when solving systems or problems involving A. THE CHARACTERISTIC POLYNOMIAL For an n × n matrix A, solving Ax = λx for a vector x= 0 is equivalent to solving the homogeneous linear system (A − λI )x = 0 This has a nonzero solution if and only if det (A − λI ) = 0 fA(λ) ≡ det a1,1 − λ · · · a1,n .. .. ... =0 an,1 · · · an,n − λ We can expand this determinant by minors, obtaining fA (λ) = a1,1 − λ · · · (an,n − λ) + terms of degree ≤ n − 2 = (−1)nλn n−1 +(−1) a1,1 + · · · + an,n λn−1 + terms of degree ≤ n − 2 We call fA(λ) the characteristic polynomial of A; and fA (λ) = 0 is called the characteristic equation for A. Since fA(λ) is a polynomial of degree n: 1. The matrix A has at least one eigenvalue. 2. A has at most n distinct eigenvalues. The multiplicity of λ as a root of fA(λ) = 0 is called the algebraic multiplicity of λ. The number of independent eigenvectors associated with λ is called the geometric multiplicity of λ. EXAMPLES: 1 0 0 1 has the eigenvalue λ = 1 with both the algebraic and geometric multiplicity equal to 2. 1 1 0 1 has the eigenvalue λ = 1 with algebraic multiplicity 2 and geometric multiplicity 1. PROPERTIES Let A and B be similar, meaning that for some nonsingular matrix P , B = P −1AP Being similar means that A and B represent the same linear transformation of the vector space Rn or Cn, but with respect to different bases. Then fB (λ) = det (B − λI ) = det P −1AP − λI = det P −1 [A − λI ] P = det P −1 det (A − λI ) [det P ] = det (A − λI ) = fA (λ) det P −1 det P P −1P since = det = 1. This says that the similar matrices have the same eigenvalues. For the eigenvectors, write Ax = λx P −1AP P −1x = λP −1x Bz = λz with z = P −1x Thus there is a simple connection with the eigenvectors of similar matrices. Since fB (λ) = fA(λ) for similar matrices A and B , we have that their coefficients are invariant. In particular, note that fA(0) = det A which is the constant term in fA(λ). Thus similar matrices have the same determinant. In particular, introduce trace A = a1,1 + · · · + an,n It is the coefficient of λn−1, except for sign. Similar matrices have the same trace and determinant. Let λ1, ..., λn be the eigenvalues of A, repeated according to their multiplicity. Then fA(λ) = (λ1 − λ) · · · (λn − λ) = (−1)nλn + (−1)n−1 (λ1 + · · · λn) λn−1 + · · · + (λ1 · · · λn) Thus trace A = λ1 + · · · λn , det A = λ1 · · · λn EXAMPLE : The eigenvalues of A= 1 2 3 4 satisfy λ1 + λ2 = 5, λ1λ2 = −2 CANONICAL FORMS By use of similarity transformations, we change a matrix A into a similar matrix which is “simpler” in its form. These are often called canonical forms; and there is a large literature on such forms. The ones most used in numerical analysis are: The Schur normal form The principal axes form The Jordan form The singular value decomposition The Schur normal form is less well-known, but is a powerful tool in deriving results in matrix algebra. SCHUR’S NORMAL FORM. Let A be a square matrix of order n with coefficients from C. Then there is a nonsingular unitary matrix U for which U ∗AU = T, U ∗U = I with T an upper triangular matrix. Since U is unitary, U ∗ = U −1, and thus T and A are similar. A proof by induction on n is given in the text, on page 474. Note that for a triangular matrix T , the characteristic polynomial is fT (λ) = det = t1,1 − λ t1,2 · · · t1,n ... 0 .. .. ... 0 · · · 0 tn,n − λ t1,1 − λ · · · (tn,n − λ) Thus the diagonal elements ti,i are the eigenvalues of T . PRINCIPAL AXES THEOREM. Let A be Hermitian (or self-adjoint), meaning A∗ = A Then there is a nonsingular unitary matrix U for which U ∗AU = D a diagonal matrix with only real entries. If the matrix A is real (and thus symmetric), the matrix U can be chosen as a real orthogonal matrix. PROOF: Using the Schur normal form, U ∗AU = T Now form the conjugate transpose of both sides T∗ = = = = = (U ∗AU )∗ U ∗A∗ (U ∗)∗ U ∗A∗U since (B ∗)∗ = B in general U ∗AU since A∗ = A T T∗ T Since = T , the diagonal elements of T are real. Moreover, T = T ∗ implies T must be a diagonal matrix, as asserted, and we write D = diag [λ1, ..., λn ] I will not show here that A real implies U real. Consequences: Again, assume A (and U ) are n × n; and let V = Rn or Cn , depending on whether A is real or complex. Write U = [u1, ..., un] with u1, ..., un the orthogonal columns of U . Since these are elements of the n-dimensional space V , {u1, ..., un} is an orthogonal basis of V . In addition, re-write U ∗AU = D as AU = U D A [u1, ..., un] = [u1, ..., un] λ1 0 · · · 0 .. 0 · .. · 0 0 · · · 0 λn Matching corresponding columns on the two sides of this equation, we have Auj = λj uj , j = 1, ..., n Thus the columns u1, ..., un are orthogonal eigenvectors of A; and they form a basis for V . For a Hermitian matrix A, the eigenvalues are all real; and there is an orthogonal basis for the associated vector space V consisting of eigenvectors of A. In dealing with such a matrix A in a problem, the basis {u1, ..., un} is often used in place of the original basis (or coordinate system) used in setting up the problem. With this basis, the use of A is clear: A (c1u1 + · · · + cnun) = c1Au1 + · · · + cnAun = c1λ1u1 + · · · + cnλnun SINGULAR VALUE DECOMPOSITION. This is a canonical form for general rectangular matrices. Let A have order n × m. Then there are unitary matrices U and V , of respective orders m × m and n × n, for which V ∗AU = F with F of order n × m and of the form F = µ1 0 · · · 0 µ2 0 · · · .. ... µr 0 ... The numbers µi are all real, with µ1 ≥ µ2 ≥ · · · ≥ µr > 0 This is proven in the text (p. 478). JORDAN CANONICAL FORM This is probably the best known canonical form of linear algebra textbooks. Begin by introducing matrices Jm(λ) = λ 1 0 λ .. . . . 0 ··· 0 ··· ... ... ... 1 0 λ The order is m × m, and all elements are zero except for λ on the diagonal and 1 on the superdiagonal. We refer to this matrix as a Jordan block. Let A have order n, with elements chosen from C. Then there is a nonsingular matrix P for which P −1AP = Jn1 (λ1) 0 ··· 0 .. Jn2 (λ2) . . . 0 .. ... ... 0 0 ··· 0 Jnr (λr ) The numbers λ1, ..., λr are the eigenvalues of A, and they need not be distinct. Note that the Jordan block Jm(λ) satisfies Jm(0)m = 0 Such a matrix is called nilpotent. We sometimes write the Jordan canonical form as P −1AP = D + N with D containing the diagonal elements of P −1AP and N containing the remaining nonzero elements. In this case, N is nilpotent with Nq = 0 with q = max {n1, ..., nr } ≤ n. The proof of the Jordan canonical form is very complicated, and it can be found in more advanced level books on linear algebra.