Diagonalization of Matrices

Diagonalization Theory: Part 1 and Part 2

The Diagonalization Theorem (Part 1)

  • Statement: If a square matrix AA has an eigenvector basis, then AA can be expressed as a special product of matrices: A=PDP1A = PDP^{-1}.

    • PP is a matrix whose columns are the eigenvectors of AA. These eigenvectors form the eigenvector basis.

    • DD is a diagonal matrix. Its diagonal entries are the eigenvalues of AA. The eigenvalues are repeated according to their algebraic multiplicity.

    • Correspondence: The ii-th column of PP (an eigenvector) must correspond to the ii-th diagonal entry of DD (its respective eigenvalue).

  • Benefit: This factorization simplifies calculations involving powers of AA.

    • For example, Ak=(PDP1)k=PDP1PDP1PDP1=PDkP1A^k = (PDP^{-1})^k = PD P^{-1} P D P^{-1} … P D P^{-1} = P D^k P^{-1} (where intermediate P1PP^{-1}P terms cancel out).

    • Calculating DkD^k for a diagonal matrix DD is straightforward: if D=diag(λ<em>1,λ</em>2,,λ<em>n)D = \text{diag}(\lambda<em>1, \lambda</em>2, …, \lambda<em>n), then Dk=diag(λ</em>1k,λ<em>2k,,λ</em>nk)D^k = \text{diag}(\lambda</em>1^k, \lambda<em>2^k, …, \lambda</em>n^k).

  • Example 1: Triangular Matrix

    • Given matrix A=(2amp;0amp;1 0amp;2amp;2 0amp;0amp;3)A = \begin{pmatrix} 2 &amp; 0 &amp; 1 \ 0 &amp; 2 &amp; 2 \ 0 &amp; 0 &amp; 3 \end{pmatrix}.

    • Eigenvalues: Since AA is a triangular matrix, its eigenvalues are the diagonal entries: λ<em>1=2\lambda<em>1 = 2, λ</em>2=2\lambda</em>2 = 2, λ3=3\lambda_3 = 3.

    • Eigenvectors: (Provided) v<em>1=(1 0 0)v<em>1 = \begin{pmatrix} 1 \ 0 \ 0 \end{pmatrix}, v</em>2=(0 1 0)v</em>2 = \begin{pmatrix} 0 \ 1 \ 0 \end{pmatrix}, v3=(1 2 1)v_3 = \begin{pmatrix} 1 \ 2 \ 1 \end{pmatrix}.

    • Factorization: A=PDP1A = P D P^{-1}

      • P=(1amp;0amp;1 0amp;1amp;2 0amp;0amp;1)P = \begin{pmatrix} 1 &amp; 0 &amp; 1 \ 0 &amp; 1 &amp; 2 \ 0 &amp; 0 &amp; 1 \end{pmatrix} (columns are eigenvectors).

      • D=(2amp;0amp;0 0amp;2amp;0 0amp;0amp;3)D = \begin{pmatrix} 2 &amp; 0 &amp; 0 \ 0 &amp; 2 &amp; 0 \ 0 &amp; 0 &amp; 3 \end{pmatrix} (diagonal entries are corresponding eigenvalues).

      • P1P^{-1} must be calculated (not immediately obvious).

  • Example 2: Markov Chain Transition Matrix

    • Eigenvectors: v<em>1=(3 5)v<em>1 = \begin{pmatrix} 3 \ 5 \end{pmatrix} (for λ</em>1=1\lambda</em>1 = 1) and v<em>2=(1 1)v<em>2 = \begin{pmatrix} 1 \ -1 \end{pmatrix} (for λ</em>2=0.92\lambda</em>2 = 0.92).

    • Factorization: T=(3amp;1 5amp;1)(1amp;0 0amp;0.92)P1T = \begin{pmatrix} 3 &amp; 1 \ 5 &amp; -1 \end{pmatrix} \begin{pmatrix} 1 &amp; 0 \ 0 &amp; 0.92 \end{pmatrix} P^{-1}. The matrix P1P^{-1} would need to be calculated explicitly.

The Diagonalization Theorem (Part 2)

  • Statement: Let AA be an n×nn \times n matrix. If AA can be expressed as A=PDP1A = PDP^{-1} where PP is an invertible matrix and DD is a diagonal matrix, then:

    • The columns of PP form an eigenvector basis for AA.

    • The corresponding eigenvalues are the diagonal entries of DD.

  • Significance: This allows us to construct a matrix AA with pre-chosen eigenvalues and eigenvectors. It is the reverse of Part 1.

  • Example: Constructing a Matrix

    • Choose eigenvectors: v<em>1=(1 4)v<em>1 = \begin{pmatrix} 1 \ 4 \end{pmatrix} and v</em>2=(3 8)v</em>2 = \begin{pmatrix} 3 \ 8 \end{pmatrix}. (These must be linearly independent to form a basis for R2R^2).

    • Choose eigenvalues: λ<em>1=πe\lambda<em>1 = \pi^e and λ</em>2=101010\lambda</em>2 = -10^{10^{10}}

    • Construct P=(1amp;3 4amp;8)P = \begin{pmatrix} 1 &amp; 3 \ 4 &amp; 8 \end{pmatrix} and D=(πeamp;0 0amp;101010)D = \begin{pmatrix} \pi^e &amp; 0 \ 0 &amp; -10^{10^{10}} \end{pmatrix}.

    • The matrix A=PDP1A = PDP^{-1} will then have these exact chosen eigenvalues and eigenvectors.

Core Concepts and Terminology

  • Diagonalizable Matrix: A square matrix AA is called diagonalizable if it can be expressed in the form A=PDP1A = PDP^{-1} where PP is invertible and DD is diagonal.

  • Equivalence: A square matrix AA is diagonalizable if and only if there exists an eigenvector basis with respect to AA. This means the existence of A=PDP1A=PDP^{-1} and the existence of an eigenvector basis are two sides of the same coin.

  • Rephrased Equivalence: An n×nn \times n matrix is diagonalizable if and only if there exist nn linearly independent eigenvectors of that matrix.

    • Note: All eigenvectors must have nn entries (i.e., be vectors in RnR^n) for the matrix multiplication to be valid.

  • Diagonalization of A: The process of writing AA in the form PDP1PDP^{-1}.

  • Diagonalization vs. Row Reduction: These are not the same. Row reduction typically changes the eigenvalues of a matrix, while diagonalization preserves them while transforming the matrix into a simpler form via its eigenvectors.

Determining if a Matrix is Diagonalizable

Case 1: Matrix with Distinct Eigenvalues
  • Principle: Eigenvectors corresponding to distinct eigenvalues are linearly independent.

  • Theorem: If an n×nn \times n matrix has nn distinct eigenvalues, then it is diagonalizable.

  • Example: A=(2amp;1 0amp;1)A = \begin{pmatrix} -2 &amp; 1 \ 0 &amp; 1 \end{pmatrix}.

    • Eigenvalues: λ<em>1=2\lambda<em>1 = -2, λ</em>2=1\lambda</em>2 = 1 (from diagonal of triangular matrix).

    • Since AA is a 2×22 \times 2 matrix and has two distinct eigenvalues, it is immediately diagonalizable. We don't need to find the eigenvectors explicitly to know this.

  • Example: A 3×33 \times 3 triangular matrix with eigenvalues 0,3,50, 3, 5 is diagonalizable because it has 33 distinct eigenvalues.

Case 2: Matrix with Repeated Eigenvalues
  • If a matrix does not have nn distinct eigenvalues, it might still be diagonalizable, but further investigation is required.

  • Example (Not Diagonalizable): B=(1amp;1 0amp;1)B = \begin{pmatrix} 1 &amp; 1 \ 0 &amp; 1 \end{pmatrix}.

    • Eigenvalue: λ1=1\lambda_1 = 1 (from diagonal of triangular matrix). Its algebraic multiplicity is 22 (since it's a 2×22 \times 2 matrix).

    • Finding Eigenvectors: For λ<em>1=1\lambda<em>1 = 1, we solve (B1I)x=0(B - 1I)x = 0. This is (0amp;1 0amp;0)(x</em>1 x2)=(0 0)\begin{pmatrix} 0 &amp; 1 \ 0 &amp; 0 \end{pmatrix} \begin{pmatrix} x</em>1 \ x_2 \end{pmatrix} = \begin{pmatrix} 0 \ 0 \end{pmatrix}.

      • This implies x<em>2=0x<em>2 = 0 and x</em>1x</em>1 is a free variable.

      • The eigenvectors are of the form t(1 0)t \begin{pmatrix} 1 \ 0 \end{pmatrix} (for t0t \ne 0).

    • Conclusion: There is only one linearly independent eigenvector (e.g., (1 0)\begin{pmatrix} 1 \ 0 \end{pmatrix}) for this matrix. We cannot form a basis of R2R^2 consisting of eigenvectors. Thus, BB is not diagonalizable.

  • Key Criterion: Geometric vs. Algebraic Multiplicity

    • Algebraic Multiplicity (AM): The number of times an eigenvalue appears as a root of the characteristic polynomial.

    • Geometric Multiplicity (GM): The dimension of the eigenspace corresponding to an eigenvalue (GM(λ)=dim(Eλ)GM(\lambda) = \dim(E_{\lambda})). This is the maximum number of linearly independent eigenvectors for that eigenvalue.

    • Bounds: For any eigenvalue λ<em>i\lambda<em>i, always: 1GM(λ</em>i)AM(λi)1 \le GM(\lambda</em>i) \le AM(\lambda_i).

    • Theorem (General Diagonalizability Condition): A square matrix AA is diagonalizable if and only if for every eigenvalue λ<em>i\lambda<em>i of AA, its geometric multiplicity equals its algebraic multiplicity (GM(λ</em>i)=AM(λi)GM(\lambda</em>i) = AM(\lambda_i)).

      • This means each eigenspace must be