Diagonalization is one of the most powerful techniques in linear algebra. When a matrix can be written as with diagonal, computations like matrix powers become trivial: .
This module covers the diagonalizability criterion, the step-by-step algorithm, and key applications to differential equations, Markov chains, and dynamical systems.
Diagonalization expresses a matrix in terms of its eigenvalues and eigenvectors, revealing its essential structure.
An matrix is diagonalizable if there exists an invertible matrix and a diagonal matrix such that:
Equivalently: (A is similar to D).
A linear operator is diagonalizable if there exists a basis of such that is diagonal.
For an matrix , the following are equivalent:
(1) ⟺ (2): If , then the columns of are eigenvectors and invertible means they're independent. Conversely, if we have independent eigenvectors, form from them.
(2) ⟺ (3): independent vectors in an -dimensional space form a basis.
(2) ⟺ (4): Eigenspaces for distinct eigenvalues are in direct sum, so total dimension equals number of independent eigenvectors.
For :
Step 1: Find eigenvalues:
Eigenvalues:
Step 2: Find eigenvectors:
For :
For :
Step 3: Form P and D:
Verify: ✓
The key question: when is a matrix diagonalizable? The answer involves comparing algebraic and geometric multiplicities.
An matrix is diagonalizable if and only if:
Equivalently:
(⟹): If is diagonalizable, then where has eigenvalues on diagonal. The char poly of equals that of , which splits completely. Each eigenspace has dimension equal to the multiplicity of that eigenvalue in .
(⟸): If the char poly splits and for all eigenvalues, then the sum of eigenspace dimensions is . Taking a basis from each eigenspace gives independent eigenvectors.
If an matrix has distinct eigenvalues, then is diagonalizable.
Each distinct eigenvalue has , and always . Combined with , we get for each. Thus the criterion is satisfied.
"Most" matrices are diagonalizable: the set of diagonalizable matrices is dense in the space of all matrices.
Is diagonalizable?
Eigenvalues: with (from triangular form)
Geometric multiplicity:
Conclusion: , so NOT diagonalizable.
Is diagonalizable?
Eigenvalues: with , with
For : (two free variables in ker)
For :
Conclusion: for both, so diagonalizable (it's already diagonal!).
A matrix that is NOT diagonalizable is called defective. For defective matrices, we use Jordan normal form instead.
Diagonalize .
Step 1: Characteristic polynomial:
Eigenvalues: (a=1), (a=2)
Step 2: Check multiplicities...
For : g = 1 = a ✓
For : need g = 2. Find ...
If g = 2: diagonalizable. If g = 1: NOT diagonalizable.
Instead of computing , you can verify (column by column: ).
One of the most important applications of diagonalization: computing efficiently.
If , then for any integer :
where .
By induction or direct computation:
In general: .
For , compute .
Step 1: Eigenvalues: 3, 2 (from diagonal of triangular matrix)
Step 2: Eigenvectors: ,
Step 3: ,
Step 4:
Step 5:
If and is a function defined on eigenvalues:
This defines matrix exponential, logarithm, square root, etc.
For diagonal :
So when A is diagonalizable.
Every real symmetric matrix is diagonalizable. Moreover, it is orthogonally diagonalizable:
where is orthogonal ().
A complex matrix is unitarily diagonalizable iff it is normal ().
For symmetric :
Eigenvalues: 1, 3
Orthonormal eigenvectors:
Column of must correspond to diagonal entry of ! Mismatching gives wrong result.
Always check for each eigenvalue before proceeding. Defective matrices need Jordan form.
It's , not . The inverse is essential for the similarity transform.
A matrix might not be diagonalizable over but is over . Specify the field!
System has solution . With diagonalization: .
Long-term behavior: as . Diagonalization reveals stationary distribution.
Fibonacci: comes from diagonalizing the companion matrix.
Diagonalizing symmetric simplifies to sum of squares.
Solve .
Matrix form: where
Diagonalize: with eigenvalues 2, 4
Solution:
The Fibonacci matrix has eigenvalues (golden ratio).
Since , we get:
Show is not diagonalizable.
Eigenvalues: , so with
Geometric:
So . NOT diagonalizable.
Diagonalize over .
Eigenvalues:
Eigenvectors:
For :
For :
Diagonalize projection .
Eigenvalues: 0, 1 (projections always have eigenvalues 0 and 1)
Eigenvectors:
: (image of P)
: (kernel of P)
What does diagonalization mean geometrically?
In the eigenvector basis, the linear transformation is just scaling along each axis. No rotation, no shearing—pure stretching.
Why do repeated eigenvalues cause problems?
Repeated eigenvalues mean the eigenspace might not have enough dimensions. We need dim = multiplicity for each eigenvalue.
Is the zero matrix diagonalizable?
Yes! . It's already diagonal (all zeros).
Can I always find P using any eigenvectors?
Yes, any basis of eigenvectors works. Different choices give different P but all valid diagonalizations.
P = eigenvectors, D = eigenvalues
Powers become trivial!
for all
Geometric = algebraic multiplicity
distinct eigenvalues
⟹ Always diagonalizable
Problem 1
Diagonalize .
Answer: Eigenvalues 2, 3. P and D can be formed from eigenvectors.
Problem 2
Is diagonalizable?
Answer: Check if g(2) = 2. If not, NOT diagonalizable.
Problem 3
If with , find in terms of P.
Answer:
Problem 4
Prove: If A is diagonalizable and all eigenvalues are 1, then A = I.
Hint:
Problem 5 (Challenge)
Show that if A is diagonalizable and B commutes with A (AB = BA), then A and B are simultaneously diagonalizable.
Challenge 1
Prove that if is diagonalizable and invertible, then is also diagonalizable.
Challenge 2
If A and B are both diagonalizable and AB = BA, prove they share a common eigenvector.
Challenge 3
Find all 2×2 matrices that are diagonalizable and satisfy .
Hint: Eigenvalues must satisfy .
Challenge 4
Prove: A matrix is diagonalizable iff its minimal polynomial has no repeated roots.
Challenge 5
For , diagonalize A over .
Hint: This is a cyclic permutation; eigenvalues are cube roots of unity.
After diagonalizing, verify:
| Matrix Type | Diagonalizable? | Notes |
|---|---|---|
| n distinct eigenvalues | Always ✓ | Most common case |
| Symmetric () | Always ✓ | Orthogonally diagonalizable |
| Normal () | Always ✓ | Unitarily diagonalizable |
| Diagonal | Always ✓ | Already diagonal! |
| Identity multiple (cI) | Always ✓ | Already diagonal |
| Repeated eigenvalues | Check g = a | May or may not be |
| Nilpotent (non-zero) | Never ✗ | Only eigenvalue is 0 with g < a |
| Jordan block (size > 1) | Never ✗ | g = 1 but a = block size |
Diagonalization reveals a matrix's structure by expressing it in terms of eigenvalues and eigenvectors. The key criterion is for all eigenvalues. When satisfied, matrix powers become trivial, differential equations simplify, and the geometry becomes pure scaling along eigenvector directions.
25+
Theorems
36+
Examples
12
Quiz Questions
8
FAQs
Cauchy (1829): First systematic study of eigenvalue problems for symmetric matrices in the context of quadratic forms.
Jacobi (1846): Developed iterative methods for finding eigenvalues of symmetric matrices.
Sylvester (1852): Developed matrix algebra and studied invariants under similarity transformations.
Weierstrass (1858): Complete theory of canonical forms, including Jordan form for non-diagonalizable cases.
Spectral Theorem: The culmination—every symmetric/normal matrix is diagonalizable in a particularly nice way with orthonormal eigenvectors.
Diagonalization has a beautiful geometric meaning: it finds directions where the linear transformation acts simply.
In the standard basis, a matrix transformation can stretch, rotate, and shear. But in the eigenvector basis:
says: "To apply A, first convert to eigenvector coordinates (), then scale (), then convert back ()."
Consider with eigenvectors (for ) and (for ).
In standard coordinates: A shears and stretches.
In eigenvector basis: A just stretches by 2 along and by 3 along .
If and are both diagonalizable and commute (), then they are simultaneously diagonalizable:
The same works for both!
If and both are diagonalizable, then for any functions .
If and :
Both are already diagonal in the same basis (standard basis).
They commute:
If is diagonalizable with distinct eigenvalues and projection matrices onto eigenspaces:
where for and .
The "spectrum" of a matrix is its set of eigenvalues. Spectral decomposition expresses A as a sum over its spectrum.
For :
A matrix is diagonalizable if and only if its minimal polynomial has no repeated roots:
where are distinct.
(Sketch): If the minimal polynomial has no repeated roots, it splits into distinct linear factors. By the primary decomposition theorem, the space decomposes into eigenspaces. Each factor contributes dimension equal to its multiplicity in the characteristic polynomial.
For :
Characteristic: (repeated root)
Minimal: (NO repeated root)
⟹ Diagonalizable (it's 2I, already diagonal!)
For :
Characteristic:
Minimal: (HAS repeated root)
⟹ NOT diagonalizable
If is diagonalizable and is any polynomial, then is also diagonalizable with:
If , then . Therefore:
And .
If is an eigenvalue of , then is an eigenvalue of .
For and :
If is diagonalizable with all eigenvalues non-zero, then:
For diagonalizable :
Google's PageRank uses the dominant eigenvector of a stochastic matrix to rank web pages by importance.
Observable quantities correspond to eigenvalues of Hermitian operators. Diagonalization finds energy levels.
PCA diagonalizes the covariance matrix to find directions of maximum variance in data.
Normal modes of vibrating systems are eigenvectors; natural frequencies are square roots of eigenvalues.
A Leslie matrix models population dynamics with age groups:
Population after years: .
Diagonalization gives long-term growth rate = dominant eigenvalue.
— Already diagonal
— Distinct eigenvalues ±1
— Symmetric
— Jordan block
— Nilpotent
— g=1, a=3
Step 1: Is the matrix symmetric (or normal)?
If YES → DIAGONALIZABLE (Spectral Theorem)
Step 2: Does it have n distinct eigenvalues?
If YES → DIAGONALIZABLE
Step 3: For repeated eigenvalues, compute geometric multiplicity
For each eigenvalue λ: find
Step 4: Compare multiplicities
If for ALL λ → DIAGONALIZABLE
If for ANY λ → NOT DIAGONALIZABLE
Diagonalization is about finding the "natural coordinates" for a linear transformation—the eigenvector basis where the transformation is simplest (just scaling).
When it works, gives us:
The Formula
"P diagonalizes A" =
The Criterion
"geo = alg for all" = for each
Always Works
"Distinct eigenvalues" or "Symmetric"
Powers
"Power the D, keep the P"
P columns
"Eigenvectors in order"
D entries
"Eigenvalues matching P"
Previous
6.2 Characteristic Polynomial
Current
6.3 Diagonalization
Next
6.4 Jordan Normal Form
With diagonalization mastered, you're ready for:
In this comprehensive module on diagonalization, you learned:
1) Find all eigenvalues via det(A - λI) = 0. 2) For each eigenvalue, compute geometric multiplicity g = dim(ker(A - λI)). 3) Compare g to algebraic multiplicity a (root multiplicity). 4) Diagonalizable ⟺ g = a for all eigenvalues.
Use Jordan normal form instead. Every matrix over ℂ is similar to a Jordan form, which is 'almost diagonal' with 1s on the superdiagonal in Jordan blocks. This is the 'best approximation' to diagonal form.
Matrix powers become trivial: A^k = PD^kP^{-1}, and D^k just raises each diagonal entry to power k. This is essential for: 1) Solving systems of differential equations x' = Ax, 2) Analyzing Markov chains, 3) Computing matrix exponentials, 4) Understanding long-term behavior of dynamical systems.
Yes! Real symmetric matrices are not just diagonalizable, but orthogonally diagonalizable: A = QDQ^T where Q is orthogonal. This is the Spectral Theorem, one of the most important results in linear algebra.
P contains eigenvectors as columns (in some order). D is diagonal with corresponding eigenvalues on the diagonal. Column i of P is an eigenvector for the eigenvalue in position (i,i) of D.
Yes! If a real matrix has complex eigenvalues (like rotation matrices), it's not diagonalizable over ℝ, but is over ℂ since complex eigenvalues are still eigenvalues with eigenvectors.
No. The eigenvalues in D can be in any order (as long as P's columns match). Also, eigenvectors can be scaled by any non-zero constant. For repeated eigenvalues, there's even more freedom in choosing eigenvector bases.
You can use row reduction, adjugate formula, or for orthogonal P (from symmetric matrices), P^{-1} = P^T. In practice, for computation you often don't need P^{-1} explicitly—just solve systems.