MathIsimple
LA-7.5
Available
Fundamental

The Spectral Theorem

The spectral theorem is one of the crown jewels of linear algebra. It states that self-adjoint operators (symmetric matrices) can be orthogonally diagonalized—they have an orthonormal basis of eigenvectors with real eigenvalues. This has profound implications for physics, statistics, and engineering.

Learning Objectives
  • Define self-adjoint (symmetric/Hermitian) operators
  • Prove eigenvalues of self-adjoint operators are real
  • Show eigenvectors for distinct eigenvalues are orthogonal
  • State and prove the spectral theorem
  • Orthogonally diagonalize symmetric matrices
  • Apply the spectral theorem to quadratic forms
  • Understand normal operators and their spectral theory
  • Connect to physical applications (quantum mechanics, vibrations)
Prerequisites
  • Orthonormal bases and projections (LA-7.2-7.4)
  • Eigenvalues and eigenvectors (LA-6.1-6.2)
  • Diagonalization (LA-6.3)
  • Inner products and adjoints

1. Self-Adjoint Operators

Definition 7.16: Adjoint Operator

Let T:VVT: V \to V be a linear operator on inner product space VV. The adjoint TT^* is the unique operator satisfying:

Tx,y=x,Tyfor all x,yV\langle Tx, y \rangle = \langle x, T^* y \rangle \quad \text{for all } x, y \in V
Remark 7.22: Matrix Representation

For matrices with standard inner product:

  • Real: A=ATA^* = A^T (transpose)
  • Complex: A=AˉT=AHA^* = \bar{A}^T = A^H (conjugate transpose)
Definition 7.17: Self-Adjoint Operator

An operator TT is self-adjoint (or Hermitian) if:

T=TT = T^*

Equivalently: Tx,y=x,Ty\langle Tx, y \rangle = \langle x, Ty \rangle for all x,yx, y.

Example 7.35: Symmetric Matrices are Self-Adjoint

A real matrix AA is self-adjoint iff A=ATA = A^T (symmetric).

A=(2113)A = \begin{pmatrix} 2 & 1 \\ 1 & 3 \end{pmatrix}

is symmetric, hence self-adjoint.

Example 7.36: Hermitian Matrix

A complex matrix is Hermitian (self-adjoint) iff A=AˉTA = \bar{A}^T:

A=(21i1+i3)A = \begin{pmatrix} 2 & 1-i \\ 1+i & 3 \end{pmatrix}

Note: diagonal entries must be real.

Theorem 7.28a: Existence of Adjoint

For any linear operator TT on a finite-dimensional inner product space, the adjoint TT^* exists and is unique.

Proof:

For fixed yy, the map xTx,yx \mapsto \langle Tx, y \rangle is a linear functional.

By the Riesz Representation Theorem, there exists unique zyz_y such that Tx,y=x,zy\langle Tx, y \rangle = \langle x, z_y \rangle.

Define Ty=zyT^* y = z_y. This TT^* is linear and unique.

Remark 7.22a: Properties of Adjoint
  • (T)=T(T^*)^* = T
  • (S+T)=S+T(S + T)^* = S^* + T^*
  • (αT)=αˉT(\alpha T)^* = \bar{\alpha} T^*
  • (ST)=TS(ST)^* = T^* S^* (order reverses)
Example 7.36a: Computing Adjoint

For T:R2R2T: \mathbb{R}^2 \to \mathbb{R}^2 with matrix A=(1234)A = \begin{pmatrix} 1 & 2 \\ 3 & 4 \end{pmatrix}:

A=AT=(1324)A^* = A^T = \begin{pmatrix} 1 & 3 \\ 2 & 4 \end{pmatrix}
Definition 7.17a: Skew-Adjoint Operator

An operator TT is skew-adjoint (or anti-Hermitian) if:

T=TT^* = -T

Skew-adjoint operators have purely imaginary eigenvalues.

Example 7.36b: Skew-Symmetric Matrix
A=(0110)A = \begin{pmatrix} 0 & 1 \\ -1 & 0 \end{pmatrix}

This is skew-symmetric (AT=AA^T = -A). Eigenvalues are ±i\pm i.

Definition 7.17b: Normal Operator

An operator TT is normal if:

TT=TTTT^* = T^*T

Self-adjoint, skew-adjoint, and unitary operators are all normal.

Remark 7.22b: Why Normal Matters

Normal operators share the key property with self-adjoint ones:

  • They have an orthonormal basis of eigenvectors
  • The Complex Spectral Theorem applies
  • But eigenvalues may be complex (for non-self-adjoint normal operators)

2. Key Properties

Theorem 7.29: Eigenvalues are Real

All eigenvalues of a self-adjoint operator are real.

Proof:

Let Tv=λvTv = \lambda v with v0v \neq 0. Then:

λv2=λv,v=λv,v=Tv,v=v,Tv=v,λv=λˉv2\lambda \|v\|^2 = \lambda \langle v, v \rangle = \langle \lambda v, v \rangle = \langle Tv, v \rangle = \langle v, Tv \rangle = \langle v, \lambda v \rangle = \bar{\lambda} \|v\|^2

Since v2>0\|v\|^2 > 0, we have λ=λˉ\lambda = \bar{\lambda}, so λR\lambda \in \mathbb{R}.

Theorem 7.30: Eigenvectors are Orthogonal

Eigenvectors corresponding to distinct eigenvalues of a self-adjoint operator are orthogonal.

Proof:

Let Tv1=λ1v1Tv_1 = \lambda_1 v_1 and Tv2=λ2v2Tv_2 = \lambda_2 v_2 with λ1λ2\lambda_1 \neq \lambda_2.

λ1v1,v2=Tv1,v2=v1,Tv2=λ2v1,v2\lambda_1 \langle v_1, v_2 \rangle = \langle Tv_1, v_2 \rangle = \langle v_1, Tv_2 \rangle = \lambda_2 \langle v_1, v_2 \rangle

So (λ1λ2)v1,v2=0(\lambda_1 - \lambda_2) \langle v_1, v_2 \rangle = 0. Since λ1λ2\lambda_1 \neq \lambda_2, we have v1,v2=0\langle v_1, v_2 \rangle = 0.

Corollary 7.6: Orthonormal Eigenvectors

We can always choose orthonormal eigenvectors for a self-adjoint operator.

Theorem 7.29a: Invariant Subspace Theorem

If TT is self-adjoint and WW is a TT-invariant subspace, then WW^\perp is also TT-invariant.

Proof:

Let vWv \in W^\perp. For any wWw \in W:

Tv,w=v,Tw=0\langle Tv, w \rangle = \langle v, Tw \rangle = 0

since TwWTw \in W (W is T-invariant) and vWv \perp W. Thus TvWTv \perp W, so TvWTv \in W^\perp.

Example 7.36c: Checking Self-Adjoint

Verify A=(123245356)A = \begin{pmatrix} 1 & 2 & 3 \\ 2 & 4 & 5 \\ 3 & 5 & 6 \end{pmatrix} is self-adjoint:

Check A=ATA = A^T: ✓ (symmetric)

Eigenvalues will be real, eigenvectors orthogonalizable.

Remark 7.22c: Geometric Interpretation

Self-adjoint operators preserve angles in a specific sense:

Tx,y=x,Ty\langle Tx, y \rangle = \langle x, Ty \rangle

The "shadow" of TxTx on yy equals the shadow of xx on TyTy.

Theorem 7.29b: Positive Semi-Definiteness

A self-adjoint operator TT is positive semi-definite iff:

Tv,v0for all v\langle Tv, v \rangle \geq 0 \quad \text{for all } v

Equivalently: all eigenvalues are non-negative.

Definition 7.18: Positive Definite

TT is positive definite if Tv,v>0\langle Tv, v \rangle > 0 for all v0v \neq 0.

Equivalently: all eigenvalues are strictly positive.

Example 7.36d: Positive Definite Test

Is A=(2112)A = \begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix} positive definite?

Eigenvalues: λ=3,1\lambda = 3, 1 (both positive) ✓

Or: Av,v=2x2+2xy+2y2=2(x+y/2)2+(3/2)y2>0\langle Av, v \rangle = 2x^2 + 2xy + 2y^2 = 2(x + y/2)^2 + (3/2)y^2 > 0

3. The Spectral Theorem

Theorem 7.31: Spectral Theorem (Finite Dimensions)

Let TT be a self-adjoint operator on finite-dimensional inner product space VV. Then:

  1. All eigenvalues of TT are real
  2. VV has an orthonormal basis of eigenvectors of TT

Equivalently, TT is orthogonally diagonalizable.

Theorem 7.32: Matrix Form

A real matrix AA is symmetric iff there exists orthogonal QQ and diagonal DD with:

A=QDQTA = QDQ^T

where columns of QQ are orthonormal eigenvectors and DD has real eigenvalues.

Example 7.37: Orthogonal Diagonalization

Diagonalize A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix}.

Eigenvalues: det(AλI)=(3λ)21=0\det(A - \lambda I) = (3-\lambda)^2 - 1 = 0

λ1=4,λ2=2\lambda_1 = 4, \lambda_2 = 2

Eigenvectors:

v1=12(1,1)T,v2=12(1,1)Tv_1 = \frac{1}{\sqrt{2}}(1, 1)^T, \quad v_2 = \frac{1}{\sqrt{2}}(1, -1)^T

Diagonalization:

A=12(1111)(4002)(1111)A = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix} \begin{pmatrix} 4 & 0 \\ 0 & 2 \end{pmatrix} \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}
Proof:

Proof of Spectral Theorem: By induction on dimension. Base case: dimV=1\dim V = 1 is trivial.

Induction: Let λ1\lambda_1 be an eigenvalue of TT (exists over C\mathbb{C}, real for self-adjoint).

Let v1v_1 be a unit eigenvector for λ1\lambda_1. Set W=span{v1}W = \text{span}\{v_1\}^\perp.

By the Invariant Subspace Theorem, WW is TT-invariant.

Apply induction to TWT|_W: get orthonormal eigenbasis for WW.

Together with v1v_1, this gives orthonormal eigenbasis for VV.

Example 7.37a: Complete 3×3 Example

Orthogonally diagonalize A=(100011011)A = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 1 \\ 0 & 1 & 1 \end{pmatrix}:

Step 1: Characteristic polynomial: det(AλI)=(1λ)[(1λ)21]=0\det(A - \lambda I) = (1 - \lambda)[(1-\lambda)^2 - 1] = 0

Eigenvalues: λ=1,2,0\lambda = 1, 2, 0

Step 2: Find eigenvectors:

  • λ=1\lambda = 1: v1=(1,0,0)Tv_1 = (1, 0, 0)^T
  • λ=2\lambda = 2: v2=(0,1,1)T/2v_2 = (0, 1, 1)^T / \sqrt{2}
  • λ=0\lambda = 0: v3=(0,1,1)T/2v_3 = (0, 1, -1)^T / \sqrt{2}

Step 3: Form Q and D.

Theorem 7.31a: Complex Spectral Theorem

A linear operator TT on a complex inner product space is normal iff VV has an orthonormal basis of eigenvectors of TT.

Remark 7.23a: Real vs Complex
  • Real case: Self-adjoint ⟹ orthonormal eigenbasis, real eigenvalues
  • Complex case: Normal ⟹ orthonormal eigenbasis, complex eigenvalues
Corollary 7.7: Powers and Functions

If A=QDQTA = QDQ^T, then:

An=QDnQT,f(A)=Qf(D)QTA^n = QD^n Q^T, \quad f(A) = Qf(D)Q^T

where f(D)f(D) applies ff to diagonal entries.

Example 7.37b: Computing A¹⁰⁰

For A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix} with A=QDQTA = QDQ^T:

A100=QD100QT=Q(4100002100)QTA^{100} = Q D^{100} Q^T = Q \begin{pmatrix} 4^{100} & 0 \\ 0 & 2^{100} \end{pmatrix} Q^T

4. Spectral Decomposition

Theorem 7.33: Spectral Decomposition

A self-adjoint operator TT can be written as:

T=i=1kλiPiT = \sum_{i=1}^{k} \lambda_i P_i

where λi\lambda_i are distinct eigenvalues and PiP_i is the orthogonal projection onto the λi\lambda_i-eigenspace.

Remark 7.23: Projection Properties
  • Pi2=PiP_i^2 = P_i (idempotent)
  • PiPj=0P_i P_j = 0 for iji \neq j (orthogonal projections)
  • Pi=I\sum P_i = I (complete)
Example 7.38: Computing Spectral Decomposition

For A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix} with λ1=4,λ2=2\lambda_1 = 4, \lambda_2 = 2:

P1=12(1111),P2=12(1111)P_1 = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & 1 \end{pmatrix}, \quad P_2 = \frac{1}{2}\begin{pmatrix} 1 & -1 \\ -1 & 1 \end{pmatrix}
A=4P1+2P2A = 4P_1 + 2P_2
Theorem 7.33a: Functional Calculus

For any function ff defined on the spectrum of TT:

f(T)=i=1kf(λi)Pif(T) = \sum_{i=1}^{k} f(\lambda_i) P_i

This allows computing matrix functions like eAe^A, A\sqrt{A}, logA\log A.

Example 7.38a: Matrix Exponential

Compute eAe^A for A=(0110)A = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}:

Eigenvalues: λ1=1,λ2=1\lambda_1 = 1, \lambda_2 = -1

eA=e1P1+e1P2=e12(1111)+1e12(1111)e^A = e^1 P_1 + e^{-1} P_2 = e \cdot \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & 1 \end{pmatrix} + \frac{1}{e} \cdot \frac{1}{2}\begin{pmatrix} 1 & -1 \\ -1 & 1 \end{pmatrix}
=(cosh1sinh1sinh1cosh1)= \begin{pmatrix} \cosh 1 & \sinh 1 \\ \sinh 1 & \cosh 1 \end{pmatrix}
Example 7.38b: Matrix Square Root

Compute A\sqrt{A} for positive definite A=(4009)A = \begin{pmatrix} 4 & 0 \\ 0 & 9 \end{pmatrix}:

A=(2003)\sqrt{A} = \begin{pmatrix} 2 & 0 \\ 0 & 3 \end{pmatrix}

For non-diagonal: A=QDQT\sqrt{A} = Q \sqrt{D} Q^T

Remark 7.23b: Resolvent

The resolvent (λIT)1(\lambda I - T)^{-1} for λ\lambda not an eigenvalue is:

(λIT)1=i1λλiPi(\lambda I - T)^{-1} = \sum_i \frac{1}{\lambda - \lambda_i} P_i
Definition 7.19: Spectral Radius

The spectral radius of TT is:

ρ(T)=maxiλi\rho(T) = \max_i |\lambda_i|

For normal operators: T=ρ(T)\|T\| = \rho(T).

5. Applications

Quantum Mechanics

Observables are self-adjoint. Real eigenvalues = measurement outcomes. Eigenstates are orthogonal.

Principal Component Analysis

Covariance matrix is symmetric. Eigenvectors give principal components for dimensionality reduction.

Vibration Analysis

Stiffness/mass matrices are symmetric. Eigenvalues give natural frequencies, eigenvectors give mode shapes.

Quadratic Forms

Classify xTAxx^T A x using eigenvalue signs. Positive definite iff all eigenvalues positive.

Example 7.39: Quantum Mechanics Application

In quantum mechanics, the energy observable is represented by the Hamiltonian HH (self-adjoint).

  • Eigenvalues EnE_n = possible energy measurements (real!)
  • Eigenvectors ψn\psi_n = stationary states (orthogonal!)
  • Any state = superposition: ψ=cnψn\psi = \sum c_n \psi_n
Example 7.39a: Principal Component Analysis

Given data matrix XX, the covariance matrix C=XTX/nC = X^T X / n is symmetric positive semi-definite.

Spectral decomposition C=QDQTC = QDQ^T:

  • Columns of QQ = principal components (orthonormal)
  • Diagonal of DD = variance in each direction
  • Project data onto top kk principal components for dimensionality reduction
Example 7.39b: Coupled Oscillators

Two masses connected by springs: Mx¨=KxM\ddot{x} = -Kx where KK is symmetric.

Natural frequencies: ωi=λi\omega_i = \sqrt{\lambda_i} (eigenvalues of M1KM^{-1}K)

Mode shapes: eigenvectors describe how masses move together.

Remark 7.24: Why Self-Adjoint Matters Physically

Self-adjoint operators guarantee:

  • Real measurements: Eigenvalues are measurable quantities
  • Complete set of states: Eigenvectors span the space
  • Orthogonal decomposition: Different modes don't interfere
  • Conservation laws: Related to symmetries via Noether's theorem

6. Normal Operators

Theorem 7.34: Characterization of Normal

An operator TT is normal iff:

Tv=Tvfor all v\|Tv\| = \|T^* v\| \quad \text{for all } v
Proof:

TT normal ⟹ Tv2=Tv,Tv=TTv,v=TTv,v=Tv2\|Tv\|^2 = \langle Tv, Tv \rangle = \langle T^* Tv, v \rangle = \langle TT^* v, v \rangle = \|T^* v\|^2

Example 7.40: Unitary Operators

Unitary operators (like rotations) satisfy UU=UU=IUU^* = U^* U = I.

They are normal with λ=1|\lambda| = 1 (eigenvalues on unit circle).

U=(cosθsinθsinθcosθ)U = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}

Eigenvalues: e±iθe^{\pm i\theta}

Remark 7.25: Classes of Normal Operators
TypeDefinitionEigenvalues
Self-adjointT=TT = T^*Real
Skew-adjointT=TT = -T^*Purely imaginary
UnitaryTT=IT^* T = IOn unit circle
ProjectionT2=T=TT^2 = T = T^*0 or 1
Theorem 7.35: Simultaneous Diagonalization

Two self-adjoint operators A,BA, B can be simultaneously diagonalized (share orthonormal eigenbasis) iff they commute: AB=BAAB = BA.

Example 7.40a: Commuting Observables

In quantum mechanics, position and momentum don't commute ([x^,p^]=i[\hat{x}, \hat{p}] = i\hbar), so they can't be simultaneously measured precisely (Heisenberg uncertainty).

7. Common Mistakes

Assuming all matrices are orthogonally diagonalizable

Only symmetric/Hermitian matrices are! Non-symmetric matrices may not even be diagonalizable.

Forgetting to normalize eigenvectors

For A = QDQᵀ, Q must be orthogonal. This requires normalizing each eigenvector to unit length.

Confusing symmetric with Hermitian

Symmetric: A = Aᵀ (real). Hermitian: A = Āᵀ (complex, requires conjugate).

Forgetting to orthogonalize within eigenspaces

If eigenvalue has multiplicity >1, apply Gram-Schmidt to get orthonormal basis for that eigenspace.

Thinking normal = self-adjoint

Normal means TT* = T*T. Self-adjoint is the special case T = T*. Unitary and skew-adjoint are also normal.

Remark 7.26: Debugging Tips

When diagonalizing symmetric matrices:

  • Check A = Aᵀ first (symmetry)
  • Verify eigenvalues are real
  • Check eigenvectors from different eigenvalues are orthogonal
  • Normalize all eigenvectors
  • Verify Q is orthogonal: QᵀQ = I
  • Confirm A = QDQᵀ by matrix multiplication

8. Key Formulas Summary

Self-Adjoint

  • T=TT = T^*
  • • Real: A=ATA = A^T
  • • Complex: A=AˉTA = \bar{A}^T

Spectral Theorem

  • A=QDQTA = QDQ^T
  • A=λiPiA = \sum \lambda_i P_i
  • • Eigenvalues real, eigenvectors orthogonal

Orthogonal Diagonalization Algorithm

  1. Verify: Check A = Aᵀ (symmetric)
  2. Eigenvalues: Solve det(A − λI) = 0
  3. Eigenvectors: Find null(A − λI) for each λ
  4. Orthogonalize: Apply Gram-Schmidt within each eigenspace
  5. Normalize: Make each eigenvector unit length
  6. Assemble: Q = [v₁ | v₂ | ... | vₙ], D = diag(λ₁, ..., λₙ)
  7. Verify: Check A = QDQᵀ

9. More Worked Examples

Example 7.41: Complete 2×2 Diagonalization

Orthogonally diagonalize A=(5445)A = \begin{pmatrix} 5 & 4 \\ 4 & 5 \end{pmatrix}:

Step 1: Check symmetry: A=ATA = A^T

Step 2: Eigenvalues:

det(AλI)=(5λ)216=λ210λ+9=0\det(A - \lambda I) = (5-\lambda)^2 - 16 = \lambda^2 - 10\lambda + 9 = 0
λ1=9,λ2=1\lambda_1 = 9, \quad \lambda_2 = 1

Step 3: Eigenvectors:

λ=9:(A9I)v=0v1=(1,1)T/2\lambda = 9: (A - 9I)v = 0 \Rightarrow v_1 = (1, 1)^T / \sqrt{2}
λ=1:(AI)v=0v2=(1,1)T/2\lambda = 1: (A - I)v = 0 \Rightarrow v_2 = (1, -1)^T / \sqrt{2}

Step 4: Result:

A=12(1111)(9001)(1111)A = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix} \begin{pmatrix} 9 & 0 \\ 0 & 1 \end{pmatrix} \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}
Example 7.42: Repeated Eigenvalue

Diagonalize A=(200031013)A = \begin{pmatrix} 2 & 0 & 0 \\ 0 & 3 & 1 \\ 0 & 1 & 3 \end{pmatrix}:

Eigenvalues: λ=2\lambda = 2 (multiplicity 1), λ=4,2\lambda = 4, 2 (from lower-right block)

Note: λ=2\lambda = 2 has multiplicity 2 but the eigenspace is 2-dimensional.

Eigenvectors for λ=2\lambda = 2: (1,0,0)T(1, 0, 0)^T and (0,1,1)T/2(0, 1, -1)^T / \sqrt{2}

Example 7.43: Quadratic Form Classification

Classify Q(x,y)=3x2+4xy+3y2Q(x, y) = 3x^2 + 4xy + 3y^2:

Matrix form: A=(3223)A = \begin{pmatrix} 3 & 2 \\ 2 & 3 \end{pmatrix}

Eigenvalues: λ=5,1\lambda = 5, 1 (both positive)

Conclusion: QQ is positive definite.

In principal axes: Q=5u2+v2Q = 5u^2 + v^2 where (u,v)(u, v) are rotated coordinates.

Example 7.44: Matrix Logarithm

Compute logA\log A for positive definite A=(2003)A = \begin{pmatrix} 2 & 0 \\ 0 & 3 \end{pmatrix}:

logA=(ln200ln3)\log A = \begin{pmatrix} \ln 2 & 0 \\ 0 & \ln 3 \end{pmatrix}

For non-diagonal A: logA=Q(logD)QT\log A = Q (\log D) Q^T

10. Connections and Extensions

Building on the Spectral Theorem

SVD (LA-8.1)

SVD generalizes spectral theorem to non-square matrices. A = UΣVᵀ uses orthonormal bases.

Quadratic Forms (LA-8.2)

Spectral theorem classifies quadratic forms. Sign of eigenvalues determines definiteness.

Polar Decomposition

Every matrix A = UP where U is unitary and P is positive semi-definite.

Infinite Dimensions

Spectral theorem extends to compact self-adjoint operators on Hilbert spaces.

Theorem 7.36: Polar Decomposition

Any invertible operator AA can be written as:

A=UPA = UP

where UU is unitary and P=AAP = \sqrt{A^* A} is positive definite.

Remark 7.27: Physical Significance

The spectral theorem is one of the most important results in mathematics because:

  • Quantum mechanics: observables are self-adjoint, eigenvalues are measurable
  • Vibrations: normal modes are eigenvectors of symmetric operators
  • Statistics: PCA uses eigendecomposition of covariance
  • Graph theory: Laplacian spectral theory describes graph structure

11. Quick Reference

Key Definitions

  • Adjoint: ⟨Tx, y⟩ = ⟨x, T*y⟩
  • Self-adjoint: T = T*
  • Normal: TT* = T*T
  • Positive definite: ⟨Tv, v⟩ > 0

Key Results

  • • Eigenvalues of self-adjoint are real
  • • Eigenvectors for distinct λ are orthogonal
  • • Self-adjoint ⟹ orthonormal eigenbasis
  • • A = QDQᵀ (orthogonal diagonalization)

Applications

  • • Powers: Aⁿ = QDⁿQᵀ
  • • Functions: f(A) = Qf(D)Qᵀ
  • • Quadratic forms: xᵀAx = Σλᵢyᵢ²
  • • PCA, vibration analysis, quantum mechanics

Verification Checklist

  • ✓ A = Aᵀ (symmetric)
  • ✓ λᵢ ∈ ℝ (real eigenvalues)
  • ✓ QᵀQ = I (orthogonal Q)
  • ✓ A = QDQᵀ (decomposition)
Remark 7.28: Historical Note

The spectral theorem has roots in the work of Cauchy, Hermite, and Hilbert. The term "spectrum" was introduced by David Hilbert around 1900, inspired by the discrete frequencies (spectral lines) observed in atomic spectra.

Notation Summary

T*Adjoint of T
AᵀTranspose (real)
A*Conjugate transpose (complex)
PᵢProjection onto λᵢ-eigenspace
ρ(T)Spectral radius

12. Additional Practice Problems

Example 7.45: Practice Problem 1

Orthogonally diagonalize A=(1221)A = \begin{pmatrix} 1 & 2 \\ 2 & 1 \end{pmatrix}:

Solution outline:

  1. Check symmetry: A = Aᵀ ✓
  2. Find eigenvalues: λ = 3, −1
  3. Find eigenvectors and normalize
  4. Form Q and D
Example 7.46: Practice Problem 2

Show that A=(1203)A = \begin{pmatrix} 1 & 2 \\ 0 & 3 \end{pmatrix} is NOT orthogonally diagonalizable:

Solution: A ≠ Aᵀ (not symmetric), so spectral theorem doesn't apply.

A is diagonalizable (distinct eigenvalues), but not orthogonally diagonalizable.

Example 7.47: Practice Problem 3

Determine if B=(420210003)B = \begin{pmatrix} 4 & 2 & 0 \\ 2 & 1 & 0 \\ 0 & 0 & 3 \end{pmatrix} is positive definite:

Solution: Find eigenvalues of the 2×2 block: λ = 5, 0

Since one eigenvalue is 0, B is positive semi-definite, not positive definite.

Example 7.48: Practice Problem 4

For rotation R=(0110)R = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} (90° rotation):

(a) Show R is normal but not self-adjoint

(b) Find eigenvalues (complex)

(c) Verify eigenvalues are on unit circle

Remark 7.29: Practice Strategy

For mastery of spectral theorem:

  • Practice 2×2 and 3×3 examples by hand
  • Verify every step: symmetry, eigenvalues real, eigenvectors orthogonal
  • Connect to applications: quadratic forms, PCA
  • Understand proof structure (induction on dimension)

13. Chapter Summary

Key Takeaways

Self-Adjoint Operators
  • • T = T* (equals its adjoint)
  • • Real matrices: A = Aᵀ (symmetric)
  • • Complex matrices: A = Ā ᵀ (Hermitian)
  • • All eigenvalues are real
Spectral Theorem
  • • Orthonormal basis of eigenvectors exists
  • • A = QDQᵀ with Q orthogonal
  • • A = Σλᵢ Pᵢ (spectral decomposition)
  • • Enables matrix function calculus
Applications
  • • Quantum mechanics: observables
  • • PCA: dimensionality reduction
  • • Vibrations: normal modes
  • • Quadratic forms: classification
Normal Operators
  • • TT* = T*T (commute with adjoint)
  • • Includes self-adjoint, unitary, skew-adjoint
  • • Have orthonormal eigenbases
  • • Eigenvalues may be complex
Remark 7.30: Mastery Checklist

You've mastered the spectral theorem when you can:

  • ✓ Identify self-adjoint/symmetric/Hermitian operators
  • ✓ Prove eigenvalues are real
  • ✓ Show eigenvectors for distinct eigenvalues are orthogonal
  • ✓ Orthogonally diagonalize symmetric matrices
  • ✓ Compute spectral decomposition A = ΣλᵢPᵢ
  • ✓ Apply to matrix functions and quadratic forms
  • ✓ Distinguish normal from self-adjoint operators

Pro Tips for Exams

  • • Always check symmetry first: if A ≠ Aᵀ, spectral theorem doesn't apply
  • • For eigenspaces with multiplicity > 1, apply Gram-Schmidt
  • • Verify Q is orthogonal by checking QᵀQ = I
  • • For matrix functions, diagonalize first: f(A) = Qf(D)Qᵀ
  • • Positive definite ⟺ all eigenvalues positive
  • • Normal ≠ self-adjoint (unitary is normal but not self-adjoint)

14. Additional Theoretical Results

Theorem 7.37: Schur Decomposition

Every square matrix AA is unitarily similar to an upper triangular matrix:

A=UTUA = UTU^*

where UU is unitary and TT is upper triangular with eigenvalues on diagonal.

Remark 7.31: Schur vs Spectral

Compare the two decompositions:

  • Schur: Always exists, T is triangular
  • Spectral: Only for normal, D is diagonal
  • Both use unitary/orthogonal similarity transforms
Theorem 7.38: Cayley-Hamilton Connection

For self-adjoint AA with spectral decomposition A=λiPiA = \sum \lambda_i P_i:

p(A)=p(λi)Pip(A) = \sum p(\lambda_i) P_i

This gives an alternative proof of Cayley-Hamilton for self-adjoint operators.

Example 7.49: Verifying Cayley-Hamilton

For A=(2112)A = \begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix} with λ=3,1\lambda = 3, 1:

Characteristic polynomial: p(λ)=(λ3)(λ1)=λ24λ+3p(\lambda) = (\lambda - 3)(\lambda - 1) = \lambda^2 - 4\lambda + 3

p(A)=A24A+3I=0p(A) = A^2 - 4A + 3I = 0

Via spectral: p(A)=p(3)P1+p(1)P2=0P1+0P2=0p(A) = p(3)P_1 + p(1)P_2 = 0 \cdot P_1 + 0 \cdot P_2 = 0

Definition 7.20: Rayleigh Quotient

For self-adjoint AA and nonzero xx, the Rayleigh quotient is:

R(x)=Ax,xx,x=xTAxxTxR(x) = \frac{\langle Ax, x \rangle}{\langle x, x \rangle} = \frac{x^T A x}{x^T x}
Theorem 7.39: Min-Max Characterization

For self-adjoint AA with eigenvalues λ1λ2λn\lambda_1 \leq \lambda_2 \leq \cdots \leq \lambda_n:

λ1=minx0R(x),λn=maxx0R(x)\lambda_1 = \min_{x \neq 0} R(x), \quad \lambda_n = \max_{x \neq 0} R(x)

The extrema are achieved at the corresponding eigenvectors.

Example 7.50: Finding Eigenvalue Bounds

For A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix}:

Rayleigh quotient at x=(1,0)Tx = (1, 0)^T: R=3R = 3

Rayleigh quotient at x=(1,1)Tx = (1, 1)^T: R=4R = 4 (max eigenvalue)

Rayleigh quotient at x=(1,1)Tx = (1, -1)^T: R=2R = 2 (min eigenvalue)

Theorem 7.40: Courant-Fischer Theorem

The k-th eigenvalue can be characterized as:

λk=mindimW=kmaxxW,x0R(x)\lambda_k = \min_{\dim W = k} \max_{x \in W, x \neq 0} R(x)
Remark 7.32: Applications of Min-Max

The min-max principle is used in:

  • Numerical analysis: Eigenvalue estimation
  • Physics: Ground state energy bounds
  • Graph theory: Cheeger inequality, spectral clustering

15. Glimpse: Infinite Dimensions

Remark 7.33: Hilbert Space Spectral Theory

The spectral theorem extends to infinite-dimensional Hilbert spaces:

  • Compact operators: Spectrum is countable, accumulates only at 0
  • Unbounded operators: Continuous and point spectrum
  • Spectral measure: Replaces sum with integral
Theorem 7.41: Spectral Theorem (Compact Case)

A compact self-adjoint operator TT on Hilbert space has:

T=n=1λnPnT = \sum_{n=1}^{\infty} \lambda_n P_n

where λn0\lambda_n \to 0 and PnP_n are finite-rank projections onto eigenspaces.

Example 7.51: Integral Operators

The integral operator (Kf)(x)=01k(x,y)f(y)dy(Kf)(x) = \int_0^1 k(x, y) f(y) \, dy with symmetric kernel is compact self-adjoint.

Its eigenvalues and eigenfunctions solve:

01k(x,y)ϕn(y)dy=λnϕn(x)\int_0^1 k(x, y) \phi_n(y) \, dy = \lambda_n \phi_n(x)
Remark 7.34: Looking Ahead

The infinite-dimensional spectral theorem is fundamental to:

  • Quantum mechanics: Observables as unbounded operators
  • PDE theory: Eigenfunction expansions
  • Harmonic analysis: Fourier theory
  • Functional analysis: C*-algebras, operator theory

16. Computational Aspects

Algorithms for Symmetric Eigenproblems
AlgorithmComplexityBest For
Power iterationO(n2)O(n^2) per iterLargest eigenvalue only
QR algorithmO(n3)O(n^3)All eigenvalues
Jacobi methodO(n3)O(n^3)High accuracy
Divide-and-conquerO(n3)O(n^3)Tridiagonal
LanczosO(n2k)O(n^2 k)Sparse, few eigenvalues
Example 7.52: Power Iteration

To find largest eigenvalue of symmetric AA:

  1. Start with random v(0)v^{(0)}
  2. Iterate: v(k+1)=Av(k)/Av(k)v^{(k+1)} = Av^{(k)} / \|Av^{(k)}\|
  3. Converges to eigenvector for λmax\lambda_{\max}
  4. Rayleigh quotient gives eigenvalue estimate
Remark 7.35: Symmetric Advantage

For symmetric matrices, eigenvalue algorithms are:

  • More stable: Real eigenvalues, orthogonal eigenvectors
  • Faster: Can exploit structure (tridiagonalize first)
  • Parallel: Divide-and-conquer works well
Example 7.53: Tridiagonalization

Before diagonalization, reduce to tridiagonal form:

A=QTQTwhere T is tridiagonalA = QTQ^T \quad \text{where } T \text{ is tridiagonal}

This costs O(n3)O(n^3) but makes subsequent steps much faster.

17. Spectral Graph Theory

Definition 7.21: Graph Laplacian

For graph GG with adjacency matrix AA and degree matrix DD:

L=DAL = D - A

LL is symmetric positive semi-definite with eigenvalues 0=λ1λ20 = \lambda_1 \leq \lambda_2 \leq \cdots.

Theorem 7.42: Spectral Properties of Laplacian
  • λ1=0\lambda_1 = 0 with eigenvector 1\mathbf{1} (all ones)
  • Multiplicity of 0 = number of connected components
  • λ2\lambda_2 (algebraic connectivity) measures connectivity
Example 7.54: Path Graph

For path graph P3P_3 (3 vertices in a line):

L=(110121011)L = \begin{pmatrix} 1 & -1 & 0 \\ -1 & 2 & -1 \\ 0 & -1 & 1 \end{pmatrix}

Eigenvalues: 0,1,30, 1, 3

Remark 7.36: Applications of Graph Laplacian

The graph Laplacian is used in:

  • Spectral clustering: Partition using eigenvectors
  • Graph embedding: Low-dimensional representation
  • Random walks: Transition probabilities
  • Graph drawing: Spectral layouts

18. Final Synthesis

The Big Picture

The Spectral Theorem is one of the most beautiful and useful results in linear algebra. It says that symmetric/Hermitian operators are completely determined by their eigenvalues and can be decomposed into simple, orthogonal pieces.

Structure

A = QDQᵀ reveals the "principal axes" of a linear transformation.

Computation

Powers, functions, and systems become easy in the eigenbasis.

Applications

From quantum mechanics to machine learning, the spectral theorem is everywhere.

Remark 7.37: Looking Forward

The spectral theorem leads naturally to:

  • SVD (LA-8.1): Spectral theory for rectangular matrices
  • Quadratic forms (LA-8.2): Classification using eigenvalue signs
  • Functional analysis: Operator theory on Hilbert spaces
  • Representation theory: Group actions and characters
Spectral Theorem Practice
12
Questions
0
Correct
0%
Accuracy
1
A real symmetric matrix has:
Easy
Not attempted
2
The adjoint TT^* of operator TT satisfies:
Medium
Not attempted
3
Self-adjoint means:
Easy
Not attempted
4
For real matrices, self-adjoint equals:
Easy
Not attempted
5
Eigenvectors for distinct eigenvalues of self-adjoint TT are:
Medium
Not attempted
6
The spectral theorem says self-adjoint TT has:
Medium
Not attempted
7
Orthogonal diagonalization means A=QDQTA = QDQ^T where:
Medium
Not attempted
8
Which matrix is NOT orthogonally diagonalizable?
Medium
Not attempted
9
For Hermitian matrix AA:
Medium
Not attempted
10
A normal operator satisfies:
Hard
Not attempted
11
The spectral decomposition A=λiPiA = \sum \lambda_i P_i uses:
Hard
Not attempted
12
Physical systems use symmetric matrices because:
Hard
Not attempted

Frequently Asked Questions

What does 'spectral' mean?

The 'spectrum' of an operator is its set of eigenvalues. The spectral theorem describes how the operator decomposes into eigenspaces—like white light splitting into its spectral colors.

Why are eigenvalues of symmetric matrices real?

If Av = λv with A symmetric, then λ||v||² = ⟨Av, v⟩ = ⟨v, Av⟩ = ⟨v, λv⟩ = λ̄||v||². So λ = λ̄, meaning λ is real.

What's the difference between adjoint and transpose?

For real matrices with standard inner product, they're the same: T* = Tᵀ. For complex matrices, adjoint means conjugate transpose: T* = T̄ᵀ = Tᴴ.

Why does orthogonal diagonalization matter?

It means A = QDQᵀ where Q⁻¹ = Qᵀ (easy to compute!). Powers become Aⁿ = QDⁿQᵀ. Functions like eᴬ are easy to compute.

What about non-symmetric matrices?

They may not be orthogonally diagonalizable, or even diagonalizable at all. Normal matrices (TT* = T*T) are the largest class with orthonormal eigenbases.

How is this used in quantum mechanics?

Observables are self-adjoint operators. The spectral theorem guarantees real measurement outcomes (eigenvalues) and orthogonal states (eigenvectors).

What's the connection to quadratic forms?

For xᵀAx with symmetric A, diagonalization gives Σλᵢyᵢ² in new coordinates. Sign of eigenvalues determines if form is positive/negative definite.

Can I always find an orthonormal eigenbasis?

For self-adjoint operators: yes! That's the spectral theorem. For non-normal operators: generally no.

What are projection matrices in spectral decomposition?

Pᵢ = (1/||vᵢ||²)vᵢvᵢᵀ projects onto eigenspace for λᵢ. They satisfy Pᵢ² = Pᵢ, PᵢPⱼ = 0 for i≠j, and ΣPᵢ = I.

Why is A = QDQᵀ and not QDQ⁻¹?

For orthogonal Q, Qᵀ = Q⁻¹, so they're the same! The Qᵀ notation emphasizes that Q is orthogonal and no matrix inversion is needed.