MathIsimple
LA-6.3
Available
Core Topic

Diagonalization

Diagonalization is one of the most powerful techniques in linear algebra. When a matrix can be written asA=PDP1A = PDP^{-1} with DD diagonal, computations like matrix powers become trivial: Ak=PDkP1A^k = PD^kP^{-1}.

This module covers the diagonalizability criterion, the step-by-step algorithm, and key applications to differential equations, Markov chains, and dynamical systems.

Learning Objectives
  • Define diagonalizability for matrices and linear operators
  • State and apply the diagonalizability criterion (g = a for all eigenvalues)
  • Master the step-by-step diagonalization algorithm
  • Compute matrix powers efficiently using A^k = PD^kP^{-1}
  • Recognize non-diagonalizable (defective) matrices
  • Apply diagonalization to differential equations and dynamical systems
  • Understand the relationship between diagonalization and similarity
  • Handle special cases: symmetric, orthogonal, and normal matrices
Prerequisites
  • Eigenvalue definition and eigenvectors (LA-6.1)
  • Characteristic polynomial and multiplicities (LA-6.2)
  • Matrix inverse computation (LA-4.3)
  • Change of basis and similarity (LA-4.1)
  • Linear independence (LA-2.3)

1. What is Diagonalization?

Diagonalization expresses a matrix in terms of its eigenvalues and eigenvectors, revealing its essential structure.

Definition 6.8: Diagonalizable Matrix

An n×nn \times n matrix AA is diagonalizable if there exists an invertible matrix PP and a diagonal matrix DD such that:

A=PDP1A = PDP^{-1}

Equivalently: D=P1APD = P^{-1}AP (A is similar to D).

Definition 6.9: Diagonalizable Linear Operator

A linear operator T:VVT: V \to V is diagonalizable if there exists a basis B\mathcal{B} of VV such that [T]B[T]_{\mathcal{B}} is diagonal.

Remark 6.3: The P and D Matrices
  • PP: Columns are eigenvectors of AA
  • DD: Diagonal entries are eigenvalues (in corresponding order)
  • Column ii of PP is an eigenvector for eigenvalue diid_{ii}
Theorem 6.20: Equivalent Conditions

For an n×nn \times n matrix AA, the following are equivalent:

  1. AA is diagonalizable
  2. AA has nn linearly independent eigenvectors
  3. Rn\mathbb{R}^n (or Cn\mathbb{C}^n) has a basis of eigenvectors of AA
  4. The sum of eigenspace dimensions equals nn
Proof:

(1) ⟺ (2): If A=PDP1A = PDP^{-1}, then the columns of PP are eigenvectors and PP invertible means they're independent. Conversely, if we have nn independent eigenvectors, form PP from them.

(2) ⟺ (3): nn independent vectors in an nn-dimensional space form a basis.

(2) ⟺ (4): Eigenspaces for distinct eigenvalues are in direct sum, so total dimension equals number of independent eigenvectors.

Example 6.25: Verifying A = PDP^{-1}

For A=(5423)A = \begin{pmatrix} 5 & 4 \\ 2 & 3 \end{pmatrix}:

Step 1: Find eigenvalues: χA=λ28λ+7=(λ1)(λ7)\chi_A = \lambda^2 - 8\lambda + 7 = (\lambda-1)(\lambda-7)

Eigenvalues: λ1=1,λ2=7\lambda_1 = 1, \lambda_2 = 7

Step 2: Find eigenvectors:

For λ=1\lambda = 1: ker(AI)=span{(1,1)T}\ker(A - I) = \text{span}\{(-1, 1)^T\}

For λ=7\lambda = 7: ker(A7I)=span{(2,1)T}\ker(A - 7I) = \text{span}\{(2, 1)^T\}

Step 3: Form P and D:

P=(1211),D=(1007)P = \begin{pmatrix} -1 & 2 \\ 1 & 1 \end{pmatrix}, \quad D = \begin{pmatrix} 1 & 0 \\ 0 & 7 \end{pmatrix}

Verify: PDP1=APDP^{-1} = A

2. The Diagonalizability Criterion

The key question: when is a matrix diagonalizable? The answer involves comparing algebraic and geometric multiplicities.

Theorem 6.21: Complete Diagonalizability Criterion

An n×nn \times n matrix AA is diagonalizable if and only if:

  1. The characteristic polynomial splits completely (factors into linear terms)
  2. For each eigenvalue λ\lambda: g(λ)=a(λ)g(\lambda) = a(\lambda)

Equivalently: ig(λi)=n\sum_{i} g(\lambda_i) = n

Proof:

(⟹): If AA is diagonalizable, then A=PDP1A = PDP^{-1} where DD has eigenvalues on diagonal. The char poly of AA equals that of DD, which splits completely. Each eigenspace has dimension equal to the multiplicity of that eigenvalue in DD.

(⟸): If the char poly splits and g=ag = a for all eigenvalues, then the sum of eigenspace dimensions is gi=ai=n\sum g_i = \sum a_i = n. Taking a basis from each eigenspace gives nn independent eigenvectors.

Theorem 6.22: Distinct Eigenvalues Suffice

If an n×nn \times n matrix AA has nn distinct eigenvalues, then AA is diagonalizable.

Proof:

Each distinct eigenvalue has a=1a = 1, and always g1g \geq 1. Combined with gag \leq a, we get g=1=ag = 1 = a for each. Thus the criterion is satisfied.

Corollary 6.8: Generic Matrices

"Most" matrices are diagonalizable: the set of diagonalizable matrices is dense in the space of all matrices.

Example 6.26: Checking Diagonalizability

Is A=(210021002)A = \begin{pmatrix} 2 & 1 & 0 \\ 0 & 2 & 1 \\ 0 & 0 & 2 \end{pmatrix} diagonalizable?

Eigenvalues: λ=2\lambda = 2 with a=3a = 3 (from triangular form)

Geometric multiplicity: g=dimker(A2I)=dimker(010001000)=1g = \dim\ker(A - 2I) = \dim\ker\begin{pmatrix} 0 & 1 & 0 \\ 0 & 0 & 1 \\ 0 & 0 & 0 \end{pmatrix} = 1

Conclusion: g=13=ag = 1 \neq 3 = a, so NOT diagonalizable.

Example 6.27: Diagonalizable Despite Repeated Eigenvalue

Is A=(200020003)A = \begin{pmatrix} 2 & 0 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 3 \end{pmatrix} diagonalizable?

Eigenvalues: λ1=2\lambda_1 = 2 with a=2a = 2, λ2=3\lambda_2 = 3 with a=1a = 1

For λ=2\lambda = 2: g=2g = 2 (two free variables in ker)

For λ=3\lambda = 3: g=1g = 1

Conclusion: g=ag = a for both, so diagonalizable (it's already diagonal!).

Remark 6.4: Defective Matrices

A matrix that is NOT diagonalizable is called defective. For defective matrices, we use Jordan normal form instead.

3. The Diagonalization Algorithm

Step-by-Step Algorithm

  1. Find eigenvalues: Solve det(AλI)=0\det(A - \lambda I) = 0
  2. Check multiplicities: Verify g(λ)=a(λ)g(\lambda) = a(\lambda) for each
  3. Find eigenvectors: For each λ\lambda, find basis of ker(AλI)\ker(A - \lambda I)
  4. Form P: Columns are eigenvectors (any order, but remember it!)
  5. Form D: Diagonal entries are eigenvalues (matching order with P)
  6. Verify: Check AP=PDAP = PD or A=PDP1A = PDP^{-1}
Example 6.28: Complete 3×3 Diagonalization

Diagonalize A=(122021122)A = \begin{pmatrix} 1 & 2 & 2 \\ 0 & 2 & 1 \\ -1 & 2 & 2 \end{pmatrix}.

Step 1: Characteristic polynomial: χA=(λ1)(λ2)(λ2)=(λ1)(λ2)2\chi_A = (\lambda - 1)(\lambda - 2)(\lambda - 2) = (\lambda - 1)(\lambda - 2)^2

Eigenvalues: λ1=1\lambda_1 = 1 (a=1), λ2=2\lambda_2 = 2 (a=2)

Step 2: Check multiplicities...

For λ=1\lambda = 1: g = 1 = a ✓

For λ=2\lambda = 2: need g = 2. Find ker(A2I)\ker(A - 2I)...

If g = 2: diagonalizable. If g = 1: NOT diagonalizable.

Remark 6.5: Verification Shortcut

Instead of computing P1P^{-1}, you can verify AP=PDAP = PD (column by column: Avi=λiviAv_i = \lambda_i v_i).

4. Computing Matrix Powers

One of the most important applications of diagonalization: computing AkA^k efficiently.

Theorem 6.23: Matrix Powers via Diagonalization

If A=PDP1A = PDP^{-1}, then for any integer k0k \geq 0:

Ak=PDkP1A^k = PD^kP^{-1}

where Dk=diag(λ1k,λ2k,,λnk)D^k = \text{diag}(\lambda_1^k, \lambda_2^k, \ldots, \lambda_n^k).

Proof:

By induction or direct computation:

A2=(PDP1)(PDP1)=PD(P1P)DP1=PD2P1A^2 = (PDP^{-1})(PDP^{-1}) = PD(P^{-1}P)DP^{-1} = PD^2P^{-1}

In general: Ak=PDkP1A^k = PD^kP^{-1}.

Example 6.29: Computing A^{100}

For A=(3102)A = \begin{pmatrix} 3 & 1 \\ 0 & 2 \end{pmatrix}, compute A100A^{100}.

Step 1: Eigenvalues: 3, 2 (from diagonal of triangular matrix)

Step 2: Eigenvectors: v1=(1,0)Tv_1 = (1,0)^T, v2=(1,1)Tv_2 = (1,-1)^T

Step 3: P=(1101)P = \begin{pmatrix} 1 & 1 \\ 0 & -1 \end{pmatrix}, D=(3002)D = \begin{pmatrix} 3 & 0 \\ 0 & 2 \end{pmatrix}

Step 4: D100=(3100002100)D^{100} = \begin{pmatrix} 3^{100} & 0 \\ 0 & 2^{100} \end{pmatrix}

Step 5: A100=PD100P1A^{100} = PD^{100}P^{-1}

Corollary 6.9: Matrix Functions

If A=PDP1A = PDP^{-1} and ff is a function defined on eigenvalues:

f(A)=Pdiag(f(λ1),,f(λn))P1f(A) = P \cdot \text{diag}(f(\lambda_1), \ldots, f(\lambda_n)) \cdot P^{-1}

This defines matrix exponential, logarithm, square root, etc.

Example 6.30: Matrix Exponential Preview

For diagonal D=diag(λ1,,λn)D = \text{diag}(\lambda_1, \ldots, \lambda_n):

eD=diag(eλ1,,eλn)e^D = \text{diag}(e^{\lambda_1}, \ldots, e^{\lambda_n})

So eA=PeDP1e^A = Pe^DP^{-1} when A is diagonalizable.

5. Special Classes of Matrices

Theorem 6.24: Symmetric Matrices

Every real symmetric matrix is diagonalizable. Moreover, it is orthogonally diagonalizable:

A=ATA=QDQTA = A^T \Rightarrow A = QDQ^T

where QQ is orthogonal (Q1=QTQ^{-1} = Q^T).

Theorem 6.25: Normal Matrices

A complex matrix is unitarily diagonalizable iff it is normal (AA=AAAA^* = A^*A).

Remark 6.6: Classes Always Diagonalizable
  • Symmetric: A=ATA = A^T ⟹ diagonalizable (real eigenvalues)
  • Skew-symmetric: A=ATA = -A^T ⟹ diagonalizable (imaginary eigenvalues)
  • Orthogonal: ATA=IA^TA = I ⟹ diagonalizable (eigenvalues on unit circle)
  • Normal: AA=AAAA^* = A^*A ⟹ unitarily diagonalizable
  • Distinct eigenvalues: Always diagonalizable
Example 6.31: Orthogonal Diagonalization

For symmetric A=(2112)A = \begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix}:

Eigenvalues: 1, 3

Orthonormal eigenvectors: 12(1,1)T,12(1,1)T\frac{1}{\sqrt{2}}(1,-1)^T, \frac{1}{\sqrt{2}}(1,1)^T

Q=12(1111),A=QDQTQ = \frac{1}{\sqrt{2}}\begin{pmatrix} 1 & 1 \\ -1 & 1 \end{pmatrix}, \quad A = QDQ^T

6. Common Mistakes

Order of eigenvectors matters

Column ii of PP must correspond to diagonal entry ii of DD! Mismatching gives wrong result.

Not all matrices are diagonalizable

Always check g=ag = a for each eigenvalue before proceeding. Defective matrices need Jordan form.

Forgetting the inverse

It's A=PDP1A = PDP^{-1}, not A=PDA = PD. The inverse is essential for the similarity transform.

Real vs Complex diagonalization

A matrix might not be diagonalizable over R\mathbb{R} but is over C\mathbb{C}. Specify the field!

7. Applications

Differential Equations

System x=Ax\mathbf{x}' = A\mathbf{x} has solution x(t)=eAtx0\mathbf{x}(t) = e^{At}\mathbf{x}_0. With diagonalization: eAt=PeDtP1e^{At} = Pe^{Dt}P^{-1}.

Markov Chains

Long-term behavior: PnP^n as nn \to \infty. Diagonalization reveals stationary distribution.

Recurrence Relations

Fibonacci: Fn=ϕnϕ^n5F_n = \frac{\phi^n - \hat{\phi}^n}{\sqrt{5}} comes from diagonalizing the companion matrix.

Quadratic Forms

Diagonalizing symmetric AA simplifies xTAx\mathbf{x}^TA\mathbf{x} to sum of squares.

Example 6.32: System of Differential Equations

Solve {x=3x+yy=x+3y\begin{cases} x' = 3x + y \\ y' = x + 3y \end{cases}.

Matrix form: x=Ax\mathbf{x}' = A\mathbf{x} where A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix}

Diagonalize: A=PDP1A = PDP^{-1} with eigenvalues 2, 4

Solution: x(t)=c1e2tv1+c2e4tv2\mathbf{x}(t) = c_1 e^{2t} v_1 + c_2 e^{4t} v_2

Example 6.33: Fibonacci via Diagonalization

The Fibonacci matrix A=(1110)A = \begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} has eigenvalues ϕ,ϕ^\phi, \hat{\phi} (golden ratio).

Since (Fn+1Fn)=An(10)\begin{pmatrix} F_{n+1} \\ F_n \end{pmatrix} = A^n \begin{pmatrix} 1 \\ 0 \end{pmatrix}, we get:

Fn=ϕnϕ^n5F_n = \frac{\phi^n - \hat{\phi}^n}{\sqrt{5}}

8. More Worked Examples

Example 6.34: Non-Diagonalizable Matrix

Show A=(1101)A = \begin{pmatrix} 1 & 1 \\ 0 & 1 \end{pmatrix} is not diagonalizable.

Eigenvalues: χA=(λ1)2\chi_A = (\lambda - 1)^2, so λ=1\lambda = 1 with a=2a = 2

Geometric: ker(AI)=ker(0100)=span{(1,0)T}\ker(A - I) = \ker\begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix} = \text{span}\{(1,0)^T\}

So g=12=ag = 1 \neq 2 = a. NOT diagonalizable.

Example 6.35: Complex Eigenvalues

Diagonalize A=(0110)A = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} over C\mathbb{C}.

Eigenvalues: λ2+1=0λ=±i\lambda^2 + 1 = 0 \Rightarrow \lambda = \pm i

Eigenvectors:

For λ=i\lambda = i: v1=(1,i)Tv_1 = (1, -i)^T

For λ=i\lambda = -i: v2=(1,i)Tv_2 = (1, i)^T

P=(11ii),D=(i00i)P = \begin{pmatrix} 1 & 1 \\ -i & i \end{pmatrix}, \quad D = \begin{pmatrix} i & 0 \\ 0 & -i \end{pmatrix}
Example 6.36: Projection Matrix

Diagonalize projection P=(1/21/21/21/2)P = \begin{pmatrix} 1/2 & 1/2 \\ 1/2 & 1/2 \end{pmatrix}.

Eigenvalues: 0, 1 (projections always have eigenvalues 0 and 1)

Eigenvectors:

λ=1\lambda = 1: (1,1)T(1, 1)^T (image of P)

λ=0\lambda = 0: (1,1)T(1, -1)^T (kernel of P)

Pmat=(1111),D=(1000)P_{mat} = \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}, \quad D = \begin{pmatrix} 1 & 0 \\ 0 & 0 \end{pmatrix}

9. Conceptual Understanding

What does diagonalization mean geometrically?

In the eigenvector basis, the linear transformation is just scaling along each axis. No rotation, no shearing—pure stretching.

Why do repeated eigenvalues cause problems?

Repeated eigenvalues mean the eigenspace might not have enough dimensions. We need dim = multiplicity for each eigenvalue.

Is the zero matrix diagonalizable?

Yes! 0=I0I10 = I \cdot 0 \cdot I^{-1}. It's already diagonal (all zeros).

Can I always find P using any eigenvectors?

Yes, any basis of eigenvectors works. Different choices give different P but all valid diagonalizations.

Key Takeaways

The Formula

A=PDP1A = PDP^{-1}

P = eigenvectors, D = eigenvalues

Matrix Powers

Ak=PDkP1A^k = PD^kP^{-1}

Powers become trivial!

The Criterion

g(λ)=a(λ)g(\lambda) = a(\lambda) for all λ\lambda

Geometric = algebraic multiplicity

Sufficient Condition

nn distinct eigenvalues

⟹ Always diagonalizable

Quick Reference

Algorithm Steps

  1. Find eigenvalues
  2. Check g = a for each
  3. Find eigenvector bases
  4. Form P and D
  5. Verify AP = PD

Always Diagonalizable

  • • Distinct eigenvalues
  • • Symmetric matrices
  • • Normal matrices
  • • Diagonal matrices

10. Additional Practice

Problem 1

Diagonalize A=(4211)A = \begin{pmatrix} 4 & -2 \\ 1 & 1 \end{pmatrix}.

Answer: Eigenvalues 2, 3. P and D can be formed from eigenvectors.

Problem 2

Is A=(210020003)A = \begin{pmatrix} 2 & 1 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 3 \end{pmatrix} diagonalizable?

Answer: Check if g(2) = 2. If not, NOT diagonalizable.

Problem 3

If A=PDP1A = PDP^{-1} with D=diag(2,3)D = \text{diag}(2, 3), find A10A^{10} in terms of P.

Answer: A10=Pdiag(210,310)P1A^{10} = P \cdot \text{diag}(2^{10}, 3^{10}) \cdot P^{-1}

Problem 4

Prove: If A is diagonalizable and all eigenvalues are 1, then A = I.

Hint: A=PDP1=PIP1=IA = PDP^{-1} = PIP^{-1} = I

Problem 5 (Challenge)

Show that if A is diagonalizable and B commutes with A (AB = BA), then A and B are simultaneously diagonalizable.

11. Challenge Problems

Challenge 1

Prove that if AA is diagonalizable and invertible, then A1A^{-1} is also diagonalizable.

Challenge 2

If A and B are both diagonalizable and AB = BA, prove they share a common eigenvector.

Challenge 3

Find all 2×2 matrices that are diagonalizable and satisfy A2=IA^2 = I.

Hint: Eigenvalues must satisfy λ2=1\lambda^2 = 1.

Challenge 4

Prove: A matrix is diagonalizable iff its minimal polynomial has no repeated roots.

Challenge 5

For A=(010001100)A = \begin{pmatrix} 0 & 1 & 0 \\ 0 & 0 & 1 \\ 1 & 0 & 0 \end{pmatrix}, diagonalize A over C\mathbb{C}.

Hint: This is a cyclic permutation; eigenvalues are cube roots of unity.

12. Exam Preparation

What You Should Know

  • • Definition of diagonalizability
  • • The complete diagonalizability criterion
  • • Step-by-step diagonalization algorithm
  • • Computing matrix powers via Ak=PDkP1A^k = PD^kP^{-1}
  • • Recognizing always-diagonalizable classes
  • • Verifying diagonalization with AP = PD

Common Exam Questions

  • • Diagonalize a given 2×2 or 3×3 matrix
  • • Determine if a matrix is diagonalizable
  • • Compute AnA^n using diagonalization
  • • True/False on diagonalizability properties
  • • Apply to differential equations
  • • Work with complex eigenvalues

Verification Checklist

After diagonalizing, verify:

P is invertible (det ≠ 0)
D is diagonal
Columns of P are eigenvectors
AP = PD (column by column)
Order matches: column i ↔ entry (i,i)
Eigenvalues match char poly

Diagonalizability Summary

Matrix TypeDiagonalizable?Notes
n distinct eigenvaluesAlways ✓Most common case
Symmetric (A=ATA = A^T)Always ✓Orthogonally diagonalizable
Normal (AA=AAAA^* = A^*A)Always ✓Unitarily diagonalizable
DiagonalAlways ✓Already diagonal!
Identity multiple (cI)Always ✓Already diagonal
Repeated eigenvaluesCheck g = aMay or may not be
Nilpotent (non-zero)Never ✗Only eigenvalue is 0 with g < a
Jordan block (size > 1)Never ✗g = 1 but a = block size

Study Tips

  • Start with eigenvalues: Can't diagonalize without them!
  • Check multiplicities early: If any g < a, stop—not diagonalizable.
  • Use AP = PD: Easier to verify than computing P1P^{-1}.
  • Remember order matters: Column i of P ↔ entry (i,i) of D.
  • Know the shortcuts: Symmetric, distinct eigenvalues → always works.
  • Practice 2×2 first: Master small cases before 3×3 and larger.

Module Summary

Diagonalization A=PDP1A = PDP^{-1} reveals a matrix's structure by expressing it in terms of eigenvalues and eigenvectors. The key criterion is g(λ)=a(λ)g(\lambda) = a(\lambda) for all eigenvalues. When satisfied, matrix powers become trivial, differential equations simplify, and the geometry becomes pure scaling along eigenvector directions.

25+

Theorems

36+

Examples

12

Quiz Questions

8

FAQs

Historical Notes

Cauchy (1829): First systematic study of eigenvalue problems for symmetric matrices in the context of quadratic forms.

Jacobi (1846): Developed iterative methods for finding eigenvalues of symmetric matrices.

Sylvester (1852): Developed matrix algebra and studied invariants under similarity transformations.

Weierstrass (1858): Complete theory of canonical forms, including Jordan form for non-diagonalizable cases.

Spectral Theorem: The culmination—every symmetric/normal matrix is diagonalizable in a particularly nice way with orthonormal eigenvectors.

13. Geometric Interpretation

Diagonalization has a beautiful geometric meaning: it finds directions where the linear transformation acts simply.

The Key Insight

In the standard basis, a matrix transformation can stretch, rotate, and shear. But in the eigenvector basis:

  • Each eigenvector direction is simply scaled by its eigenvalue
  • No rotation between eigenvector directions
  • No shearing—pure stretching/compression
Remark 6.7: Change of Basis View

A=PDP1A = PDP^{-1} says: "To apply A, first convert to eigenvector coordinates (P1P^{-1}), then scale (DD), then convert back (PP)."

Example 6.37: Geometric Visualization

Consider A=(2103)A = \begin{pmatrix} 2 & 1 \\ 0 & 3 \end{pmatrix} with eigenvectors (1,0)T(1,0)^T (for λ=2\lambda=2) and (1,1)T(1,1)^T (for λ=3\lambda=3).

In standard coordinates: A shears and stretches.

In eigenvector basis: A just stretches by 2 along (1,0)(1,0) and by 3 along (1,1)(1,1).

14. Simultaneous Diagonalization

Theorem 6.26: Commuting Diagonalizable Matrices

If AA and BB are both diagonalizable and commute (AB=BAAB = BA), then they are simultaneously diagonalizable:

P:A=PDAP1,B=PDBP1\exists P: A = PD_AP^{-1}, \quad B = PD_BP^{-1}

The same PP works for both!

Corollary 6.10: Functions of Commuting Matrices

If AB=BAAB = BA and both are diagonalizable, then f(A)g(B)=g(B)f(A)f(A)g(B) = g(B)f(A) for any functions f,gf, g.

Example 6.38: Simultaneous Diagonalization

If A=(1002)A = \begin{pmatrix} 1 & 0 \\ 0 & 2 \end{pmatrix} and B=(3004)B = \begin{pmatrix} 3 & 0 \\ 0 & 4 \end{pmatrix}:

Both are already diagonal in the same basis (standard basis).

They commute: AB=BA=(3008)AB = BA = \begin{pmatrix} 3 & 0 \\ 0 & 8 \end{pmatrix}

15. Spectral Decomposition

Theorem 6.27: Spectral Decomposition

If AA is diagonalizable with distinct eigenvalues λ1,,λk\lambda_1, \ldots, \lambda_k and projection matrices P1,,PkP_1, \ldots, P_k onto eigenspaces:

A=λ1P1+λ2P2++λkPkA = \lambda_1 P_1 + \lambda_2 P_2 + \cdots + \lambda_k P_k

where PiPj=0P_i P_j = 0 for iji \neq j and Pi=I\sum P_i = I.

Remark 6.8: Why "Spectral"

The "spectrum" of a matrix is its set of eigenvalues. Spectral decomposition expresses A as a sum over its spectrum.

Example 6.39: Spectral Form

For A=(3005)A = \begin{pmatrix} 3 & 0 \\ 0 & 5 \end{pmatrix}:

A=3(1000)+5(0001)A = 3 \cdot \begin{pmatrix} 1 & 0 \\ 0 & 0 \end{pmatrix} + 5 \cdot \begin{pmatrix} 0 & 0 \\ 0 & 1 \end{pmatrix}

16. Connection to Minimal Polynomial

Theorem 6.28: Diagonalizability via Minimal Polynomial

A matrix AA is diagonalizable if and only if its minimal polynomial has no repeated roots:

mA(λ)=(λλ1)(λλ2)(λλk)m_A(\lambda) = (\lambda - \lambda_1)(\lambda - \lambda_2) \cdots (\lambda - \lambda_k)

where λ1,,λk\lambda_1, \ldots, \lambda_k are distinct.

Proof:

(Sketch): If the minimal polynomial has no repeated roots, it splits into distinct linear factors. By the primary decomposition theorem, the space decomposes into eigenspaces. Each factor contributes dimension equal to its multiplicity in the characteristic polynomial.

Example 6.40: Comparing Char and Min Poly

For A=(2002)A = \begin{pmatrix} 2 & 0 \\ 0 & 2 \end{pmatrix}:

Characteristic: χA=(λ2)2\chi_A = (\lambda - 2)^2 (repeated root)

Minimal: mA=λ2m_A = \lambda - 2 (NO repeated root)

⟹ Diagonalizable (it's 2I, already diagonal!)

Example 6.41: Minimal Polynomial with Repeated Root

For A=(2102)A = \begin{pmatrix} 2 & 1 \\ 0 & 2 \end{pmatrix}:

Characteristic: χA=(λ2)2\chi_A = (\lambda - 2)^2

Minimal: mA=(λ2)2m_A = (\lambda - 2)^2 (HAS repeated root)

⟹ NOT diagonalizable

17. More Theoretical Results

Theorem 6.29: Diagonalizability of Polynomials

If AA is diagonalizable and p(x)p(x) is any polynomial, then p(A)p(A) is also diagonalizable with:

p(A)=Pdiag(p(λ1),,p(λn))P1p(A) = P \cdot \text{diag}(p(\lambda_1), \ldots, p(\lambda_n)) \cdot P^{-1}
Proof:

If A=PDP1A = PDP^{-1}, then Ak=PDkP1A^k = PD^kP^{-1}. Therefore:

p(A)=kckAk=kckPDkP1=P(kckDk)P1=Pp(D)P1p(A) = \sum_k c_k A^k = \sum_k c_k PD^kP^{-1} = P \left(\sum_k c_k D^k\right) P^{-1} = Pp(D)P^{-1}

And p(D)=diag(p(λ1),,p(λn))p(D) = \text{diag}(p(\lambda_1), \ldots, p(\lambda_n)).

Theorem 6.30: Eigenvalues of Polynomial

If λ\lambda is an eigenvalue of AA, then p(λ)p(\lambda) is an eigenvalue of p(A)p(A).

Example 6.42: Computing p(A)

For A=(2003)A = \begin{pmatrix} 2 & 0 \\ 0 & 3 \end{pmatrix} and p(x)=x24x+3p(x) = x^2 - 4x + 3:

p(A)=diag(p(2),p(3))=diag(48+3,912+3)=diag(1,0)p(A) = \text{diag}(p(2), p(3)) = \text{diag}(4-8+3, 9-12+3) = \text{diag}(-1, 0)

Corollary 6.11: Invertibility

If AA is diagonalizable with all eigenvalues non-zero, then:

A1=Pdiag(λ11,,λn1)P1A^{-1} = P \cdot \text{diag}(\lambda_1^{-1}, \ldots, \lambda_n^{-1}) \cdot P^{-1}
Theorem 6.31: Trace and Determinant

For diagonalizable A=PDP1A = PDP^{-1}:

  • tr(A)=λ1+λ2++λn\text{tr}(A) = \lambda_1 + \lambda_2 + \cdots + \lambda_n
  • det(A)=λ1λ2λn\det(A) = \lambda_1 \cdot \lambda_2 \cdots \lambda_n
  • tr(Ak)=λ1k+λ2k++λnk\text{tr}(A^k) = \lambda_1^k + \lambda_2^k + \cdots + \lambda_n^k

18. Real-World Applications

PageRank Algorithm

Google's PageRank uses the dominant eigenvector of a stochastic matrix to rank web pages by importance.

Quantum Mechanics

Observable quantities correspond to eigenvalues of Hermitian operators. Diagonalization finds energy levels.

Principal Component Analysis

PCA diagonalizes the covariance matrix to find directions of maximum variance in data.

Vibration Analysis

Normal modes of vibrating systems are eigenvectors; natural frequencies are square roots of eigenvalues.

Example 6.43: Population Model

A Leslie matrix models population dynamics with age groups:

L=(0210.50000.60)L = \begin{pmatrix} 0 & 2 & 1 \\ 0.5 & 0 & 0 \\ 0 & 0.6 & 0 \end{pmatrix}

Population after nn years: pn=Lnp0\mathbf{p}_n = L^n \mathbf{p}_0.

Diagonalization gives long-term growth rate = dominant eigenvalue.

Quick Examples Gallery

✓ Diagonalizable

(1002)\begin{pmatrix} 1 & 0 \\ 0 & 2 \end{pmatrix} — Already diagonal

(0110)\begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} — Distinct eigenvalues ±1

(2222)\begin{pmatrix} 2 & 2 \\ 2 & 2 \end{pmatrix} — Symmetric

✗ NOT Diagonalizable

(1101)\begin{pmatrix} 1 & 1 \\ 0 & 1 \end{pmatrix} — Jordan block

(0100)\begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix} — Nilpotent

(210021002)\begin{pmatrix} 2 & 1 & 0 \\ 0 & 2 & 1 \\ 0 & 0 & 2 \end{pmatrix} — g=1, a=3

Diagonalizability Decision Tree

Step 1: Is the matrix symmetric (or normal)?

If YES → DIAGONALIZABLE (Spectral Theorem)

Step 2: Does it have n distinct eigenvalues?

If YES → DIAGONALIZABLE

Step 3: For repeated eigenvalues, compute geometric multiplicity

For each eigenvalue λ: find g(λ)=dimker(AλI)g(\lambda) = \dim\ker(A - \lambda I)

Step 4: Compare multiplicities

If g(λ)=a(λ)g(\lambda) = a(\lambda) for ALL λ → DIAGONALIZABLE

If g(λ)<a(λ)g(\lambda) < a(\lambda) for ANY λ → NOT DIAGONALIZABLE

The Big Picture

Diagonalization is about finding the "natural coordinates" for a linear transformation—the eigenvector basis where the transformation is simplest (just scaling).

When it works, A=PDP1A = PDP^{-1} gives us:

  • Easy computation of powers: Ak=PDkP1A^k = PD^kP^{-1}
  • Solutions to differential equations: eAt=PeDtP1e^{At} = Pe^{Dt}P^{-1}
  • Understanding of long-term dynamics via dominant eigenvalue
  • Geometric insight: A acts as pure scaling in eigenvector directions

Memorization Aids

The Formula

"P diagonalizes A" = A=PDP1A = PDP^{-1}

The Criterion

"geo = alg for all" = g=ag = a for each λ\lambda

Always Works

"Distinct eigenvalues" or "Symmetric"

Powers

"Power the D, keep the P"

P columns

"Eigenvectors in order"

D entries

"Eigenvalues matching P"

Learning Path

Previous

6.2 Characteristic Polynomial

Current

6.3 Diagonalization

Next

6.4 Jordan Normal Form

What's Next?

With diagonalization mastered, you're ready for:

  • Jordan Normal Form: The canonical form for non-diagonalizable matrices—"almost diagonal" with 1s on superdiagonal
  • Cayley-Hamilton Theorem: Every matrix satisfies its own characteristic equation: χA(A)=0\chi_A(A) = 0
  • Spectral Theorem: Orthogonal diagonalization for symmetric matrices with orthonormal eigenvectors
  • Matrix Exponential: Using eA=PeDP1e^A = Pe^DP^{-1} for systems of differential equations

Module Summary

In this comprehensive module on diagonalization, you learned:

  • The definition: A=PDP1A = PDP^{-1} with P containing eigenvectors and D containing eigenvalues
  • The criterion: diagonalizable ⟺ g(λ)=a(λ)g(\lambda) = a(\lambda) for all eigenvalues
  • The algorithm: find eigenvalues, check multiplicities, construct P and D
  • Applications: matrix powers, differential equations, Markov chains
  • Special cases: symmetric matrices are always orthogonally diagonalizable
  • Connections: minimal polynomial, spectral decomposition, similarity theory

Related Topics

Eigenvalues
Characteristic Polynomial
Jordan Form
Spectral Theorem
Matrix Powers
Similarity
Differential Equations
Minimal Polynomial
Matrix Exponential
Cayley-Hamilton
Diagonalization Practice
12
Questions
0
Correct
0%
Accuracy
1
A matrix is diagonalizable iff it has:
Easy
Not attempted
2
If A=PDP1A = PDP^{-1} with DD diagonal, then Ak=A^k =
Medium
Not attempted
3
A matrix with nn distinct eigenvalues is:
Medium
Not attempted
4
Geometric mult = Algebraic mult for all eigenvalues implies:
Hard
Not attempted
5
Which matrix is NOT diagonalizable?
Medium
Not attempted
6
In A=PDP1A = PDP^{-1}, the columns of PP are:
Easy
Not attempted
7
If AA is diagonalizable, so is:
Medium
Not attempted
8
A real matrix with eigenvalues 2±3i2 \pm 3i is:
Hard
Not attempted
9
For diagonal matrix DD, DkD^k equals:
Easy
Not attempted
10
Every real symmetric matrix is:
Medium
Not attempted
11
If A=PDP1A = PDP^{-1}, then det(A)=\det(A) =
Medium
Not attempted
12
A nilpotent matrix (non-zero) is:
Hard
Not attempted

Frequently Asked Questions

How do I check if a matrix is diagonalizable?

1) Find all eigenvalues via det(A - λI) = 0. 2) For each eigenvalue, compute geometric multiplicity g = dim(ker(A - λI)). 3) Compare g to algebraic multiplicity a (root multiplicity). 4) Diagonalizable ⟺ g = a for all eigenvalues.

What if a matrix isn't diagonalizable?

Use Jordan normal form instead. Every matrix over ℂ is similar to a Jordan form, which is 'almost diagonal' with 1s on the superdiagonal in Jordan blocks. This is the 'best approximation' to diagonal form.

Why is diagonalization useful?

Matrix powers become trivial: A^k = PD^kP^{-1}, and D^k just raises each diagonal entry to power k. This is essential for: 1) Solving systems of differential equations x' = Ax, 2) Analyzing Markov chains, 3) Computing matrix exponentials, 4) Understanding long-term behavior of dynamical systems.

Are symmetric matrices always diagonalizable?

Yes! Real symmetric matrices are not just diagonalizable, but orthogonally diagonalizable: A = QDQ^T where Q is orthogonal. This is the Spectral Theorem, one of the most important results in linear algebra.

What's the difference between P and D in A = PDP^{-1}?

P contains eigenvectors as columns (in some order). D is diagonal with corresponding eigenvalues on the diagonal. Column i of P is an eigenvector for the eigenvalue in position (i,i) of D.

Can a matrix be diagonalizable over ℂ but not over ℝ?

Yes! If a real matrix has complex eigenvalues (like rotation matrices), it's not diagonalizable over ℝ, but is over ℂ since complex eigenvalues are still eigenvalues with eigenvectors.

Is the diagonalization unique?

No. The eigenvalues in D can be in any order (as long as P's columns match). Also, eigenvectors can be scaled by any non-zero constant. For repeated eigenvalues, there's even more freedom in choosing eigenvector bases.

How do I find P^{-1}?

You can use row reduction, adjugate formula, or for orthogonal P (from symmetric matrices), P^{-1} = P^T. In practice, for computation you often don't need P^{-1} explicitly—just solve systems.