MathIsimple
LA-C10
Available

Course 10: Inner Product Spaces

Inner product spaces add geometric structure to vector spaces, enabling concepts of length, angle, and orthogonality. This course covers inner products, orthonormal bases, projections, least squares approximation, and the fundamental spectral theorem for symmetric matrices.

15-18 hours Advanced Level 10 Objectives
Learning Objectives
  • Define inner products on real and complex vector spaces.
  • Understand the Cauchy-Schwarz inequality and its consequences.
  • Define orthogonality and orthonormal bases.
  • Master the Gram-Schmidt orthogonalization process.
  • Compute orthogonal projections onto subspaces.
  • Apply least squares approximation and solve normal equations.
  • State and apply the spectral theorem for self-adjoint operators.
  • Understand orthogonal diagonalization of symmetric matrices.
  • Connect inner products to geometry (length, angle, distance).
  • Apply these concepts to function spaces and applications.
Prerequisites
  • LA-C9: Eigenvalues & Eigenvectors
  • Vector spaces and bases
  • Linear independence
  • Complex numbers (for complex inner products)
  • Matrix operations
Historical Context

The concept of inner products evolved from the dot product in Euclidean geometry. Augustin-Louis Cauchy (1789–1857) proved the Cauchy-Schwarz inequality in 1821. Jørgen Gram (1850–1916) and Erhard Schmidt (1876–1959) developed the orthogonalization process independently. The spectral theorem, one of the most important results in linear algebra, was proven by David Hilbert (1862–1943) and others in the early 20th century. Inner product spaces, especially Hilbert spaces, are fundamental in quantum mechanics, signal processing, and functional analysis.

1. Inner Products

An inner product is a function that assigns a scalar to each pair of vectors, generalizing the dot product and enabling geometric concepts like length and angle.

Definition 1.1: Inner Product

An inner product on a vector space VV over FF is a function ,:V×VF\langle \cdot, \cdot \rangle: V \times V \to F satisfying:

  1. Linearity: αu+βv,w=αu,w+βv,w\langle \alpha u + \beta v, w \rangle = \alpha \langle u, w \rangle + \beta \langle v, w \rangle
  2. Conjugate symmetry: v,w=w,v\langle v, w \rangle = \overline{\langle w, v \rangle} (for real: symmetry)
  3. Positive definiteness: v,v0\langle v, v \rangle \geq 0 with equality iff v=0v = 0
Definition 1.2: Induced Norm

The norm induced by an inner product is v=v,v||v|| = \sqrt{\langle v, v \rangle}.

Theorem 1.1: Cauchy-Schwarz Inequality

For any vectors x,yx, y in an inner product space:

x,yxy|\langle x, y \rangle| \leq ||x|| \cdot ||y||

Equality holds if and only if xx and yy are linearly dependent.

Example 1.1: Standard Inner Products
  • Rn\mathbb{R}^n: x,y=xTy=xiyi\langle x, y \rangle = x^T y = \sum x_i y_i
  • Cn\mathbb{C}^n: x,y=xy=xˉiyi\langle x, y \rangle = x^* y = \sum \bar{x}_i y_i
  • L2[a,b]L^2[a,b]: f,g=abf(t)g(t)dt\langle f, g \rangle = \int_a^b f(t) \overline{g(t)} dt

2. Orthogonality

Orthogonality generalizes perpendicularity. Orthonormal bases provide the most convenient coordinate systems for computation.

Definition 2.1: Orthogonal and Orthonormal

Vectors u,vu, v are orthogonal if u,v=0\langle u, v \rangle = 0.

A set {e1,,ek}\{e_1, \ldots, e_k\} is orthonormal if ei,ej=δij\langle e_i, e_j \rangle = \delta_{ij} (Kronecker delta).

Theorem 2.1: Independence of Orthogonal Vectors

An orthogonal set of nonzero vectors is linearly independent.

Theorem 2.2: Fourier Coefficients

If {e1,,en}\{e_1, \ldots, e_n\} is an orthonormal basis, then for any v=cieiv = \sum c_i e_i:

ci=v,eic_i = \langle v, e_i \rangle

These are called Fourier coefficients.

Theorem 2.3: Parseval's Identity

For an orthonormal basis, v2=ci2||v||^2 = \sum |c_i|^2 where ci=v,eic_i = \langle v, e_i \rangle.

3. Gram-Schmidt Process

The Gram-Schmidt process converts any basis into an orthonormal basis, proving that every finite-dimensional inner product space has an orthonormal basis.

Definition 3.1: Gram-Schmidt Algorithm

Given linearly independent vectors {v1,,vn}\{v_1, \ldots, v_n\}:

  1. e1=v1/v1e_1 = v_1 / ||v_1|| (normalize first vector)
  2. For k=2,,nk = 2, \ldots, n:
    • uk=vki=1k1vk,eieiu_k = v_k - \sum_{i=1}^{k-1} \langle v_k, e_i \rangle e_i (subtract projections)
    • ek=uk/uke_k = u_k / ||u_k|| (normalize)

Result: {e1,,en}\{e_1, \ldots, e_n\} is orthonormal with span{e1,,ek}=span{v1,,vk}\text{span}\{e_1, \ldots, e_k\} = \text{span}\{v_1, \ldots, v_k\}.

Theorem 3.1: QR Decomposition

If AA has linearly independent columns, then A=QRA = QR where QQ has orthonormal columns (from Gram-Schmidt) and RR is upper triangular.

Example 3.1: Gram-Schmidt Example

For v1=(1,1,0),v2=(1,0,1)v_1 = (1,1,0), v_2 = (1,0,1) in R3\mathbb{R}^3:

  • e1=(1/2,1/2,0)e_1 = (1/\sqrt{2}, 1/\sqrt{2}, 0)
  • u2=(1,0,1)(1,0,1),e1e1=(1/2,1/2,1)u_2 = (1,0,1) - \langle (1,0,1), e_1 \rangle e_1 = (1/2, -1/2, 1)
  • e2=u2/u2e_2 = u_2 / ||u_2||

4. Orthogonal Projections

The orthogonal projection of a vector onto a subspace is the closest point in that subspace. This is fundamental for least squares approximation.

Definition 4.1: Orthogonal Projection

The orthogonal projection of vv onto subspace UU is the unique vector projU(v)U\text{proj}_U(v) \in U such that vprojU(v)Uv - \text{proj}_U(v) \in U^\perp.

Theorem 4.1: Projection Formula

If {e1,,ek}\{e_1, \ldots, e_k\} is an orthonormal basis for UU, then:

projU(v)=i=1kv,eiei\text{proj}_U(v) = \sum_{i=1}^k \langle v, e_i \rangle e_i
Theorem 4.2: Best Approximation

projU(v)\text{proj}_U(v) minimizes vu||v - u|| over all uUu \in U.

Definition 4.2: Least Squares

For AxbAx \approx b (overdetermined), the least squares solution minimizes Axb||Ax - b||.

Theorem 4.3: Normal Equation

The least squares solution satisfies ATAx=ATbA^T A x = A^T b (the normal equation).

Example 4.1: Linear Regression

For data points (xi,yi)(x_i, y_i), fitting y=ax+by = ax + b minimizes (yiaxib)2\sum (y_i - ax_i - b)^2. This is least squares with A=[x1],b=yA = [x \, 1], b = y.

5. Spectral Theorem

The spectral theorem is one of the most important results in linear algebra, characterizing self-adjoint operators and enabling orthogonal diagonalization of symmetric matrices.

Definition 5.1: Self-Adjoint Operator

An operator TT on an inner product space is self-adjoint if T=TT = T^*, where TT^* is the adjoint satisfying Tx,y=x,Ty\langle Tx, y \rangle = \langle x, T^* y \rangle.

For real matrices with standard inner product, self-adjoint means symmetric (A=ATA = A^T).

Theorem 5.1: Eigenvalues of Self-Adjoint Operators

All eigenvalues of a self-adjoint operator are real.

Theorem 5.2: Orthogonality of Eigenvectors

Eigenvectors corresponding to distinct eigenvalues of a self-adjoint operator are orthogonal.

Theorem 5.3: Spectral Theorem

Every self-adjoint operator on a finite-dimensional inner product space has an orthonormal basis of eigenvectors. Equivalently, every real symmetric matrix is orthogonally diagonalizable: A=QDQTA = QDQ^T where QQ is orthogonal and DD is diagonal with real entries.

Corollary 5.1: Spectral Decomposition

If A=QDQTA = QDQ^T with eigenvalues λi\lambda_i and orthonormal eigenvectors qiq_i, then:

A=i=1nλiqiqiTA = \sum_{i=1}^n \lambda_i q_i q_i^T

This is the spectral decomposition.

Remark 5.1: Applications

The spectral theorem is fundamental in quantum mechanics (observables are self-adjoint), principal component analysis, and many areas of applied mathematics.

Course 10 Practice Quiz
10
Questions
0
Correct
0%
Accuracy
1
An inner product ,\langle \cdot, \cdot \rangle must satisfy:
Easy
Not attempted
2
Cauchy-Schwarz states that:
Medium
Not attempted
3
Vectors u,vu, v are orthogonal if:
Easy
Not attempted
4
If {e1,...,ek}\{e_1,...,e_k\} is orthonormal basis for UU, then projU(v)=\text{proj}_U(v) =
Medium
Not attempted
5
The normal equation for least squares AxbAx \approx b is:
Medium
Not attempted
6
A real symmetric matrix has:
Easy
Not attempted
7
The spectral theorem says self-adjoint TT has:
Medium
Not attempted
8
Gram-Schmidt takes as input:
Easy
Not attempted
9
For orthonormal basis {e1,...,en}\{e_1,...,e_n\}, coefficient ckc_k of v=cieiv = \sum c_i e_i is:
Medium
Not attempted
10
Orthogonal diagonalization means A=QDQTA = QDQ^T where:
Medium
Not attempted

Frequently Asked Questions

What's the geometric meaning of inner products?

Inner products measure 'angle' and 'length'. ||v|| = √⟨v,v⟩ gives length. cos θ = ⟨x,y⟩/(||x||·||y||) defines angle. Orthogonality (⟨x,y⟩ = 0) means perpendicular.

Why are orthonormal bases so useful?

Three main reasons: (1) Coefficients are trivial to compute as inner products, no matrix inversion needed. (2) The Gram matrix is the identity, simplifying all computations. (3) Parseval's identity gives ||v||² = Σ|cᵢ|², preserving norms.

How does Gram-Schmidt work?

Start with first vector, normalize it. For each subsequent vector, subtract its projections onto all previous orthonormal vectors, then normalize. This makes each new vector orthogonal to all previous ones.

What is least squares approximation?

When Ax = b has no solution, find x that minimizes ||Ax - b||. This is equivalent to finding the projection of b onto col(A). The solution satisfies the normal equation AᵀAx = Aᵀb.

What does the spectral theorem say?

Every self-adjoint (symmetric/Hermitian) operator has an orthonormal basis of eigenvectors with real eigenvalues. This means A = QDQᵀ where Q is orthogonal and D is diagonal with real entries.