MathIsimple
LA-C7
Available

Course 7: Matrix Inverses, Elementary Matrices & Dual Spaces

This course deepens your understanding of matrix theory by covering inverses, the building blocks of row operations (elementary matrices), the fundamental concept of rank, and the elegant theory of dual spaces. These topics connect matrix computations to abstract linear algebra.

12-15 hours Advanced Core Level 10 Objectives
Learning Objectives
  • Define and compute matrix inverses using Gaussian elimination.
  • Understand the invertible matrix theorem and its many characterizations.
  • Master elementary matrices and their role in row operations.
  • Express invertible matrices as products of elementary matrices.
  • Define matrix rank and prove row rank equals column rank.
  • Apply rank to determine solvability of linear systems.
  • Understand rank inequalities and their applications.
  • Define dual spaces and linear functionals.
  • Construct dual bases and understand the natural isomorphism V ≅ V**.
  • Compute annihilators and understand their dimension formula.
Prerequisites
  • LA-C6: Matrix Theory
  • Matrix operations and representation
  • Gaussian elimination
  • Linear maps and kernel/image
  • Basis and dimension
Historical Context

The concept of matrix inverses emerged naturally from solving systems of linear equations, with Arthur Cayley (1821–1895) formalizing matrix algebra in 1858. Elementary matrices, though implicit in Gaussian elimination, were explicitly studied in the early 20th century. The rank of a matrix, a fundamental invariant, was recognized by James Joseph Sylvester in the 1850s. The beautiful theorem that row rank equals column rank was proven by various mathematicians, with Emmy Noetherproviding elegant abstract proofs. Dual spaces, introduced by Hermann Grassmannin the 1840s and later developed by Jean Dieudonné and others, provide a powerful framework for understanding linear maps and their adjoints.

1. Matrix Inverses

The inverse of a matrix AA is the unique matrix A1A^{-1} such that AA1=A1A=IAA^{-1} = A^{-1}A = I. Invertible matrices correspond to isomorphisms and are fundamental for solving linear systems.

Definition 1.1: Matrix Inverse

An n×nn \times n matrix AA is invertible (or nonsingular) if there exists A1A^{-1} such that AA1=A1A=InAA^{-1} = A^{-1}A = I_n.

Theorem 1.1: Uniqueness of Inverse

If an inverse exists, it is unique.

Proof:

If B,CB, C are both inverses of AA, then B=BI=B(AC)=(BA)C=IC=CB = BI = B(AC) = (BA)C = IC = C.

Theorem 1.2: Invertible Matrix Theorem

For an n×nn \times n matrix AA, the following are equivalent:

  1. AA is invertible
  2. rank(A)=n\text{rank}(A) = n
  3. The columns of AA are linearly independent
  4. Ax=0Ax = 0 has only the trivial solution
  5. Ax=bAx = b has a unique solution for every bb
  6. AA is a product of elementary matrices
  7. AA is row equivalent to InI_n
Example 1.1: Computing Inverse via Gaussian Elimination

To find A1A^{-1}, row reduce [AI][A|I] to [IA1][I|A^{-1}]. For example, if A=(1234)A = \begin{pmatrix} 1 & 2 \\ 3 & 4 \end{pmatrix}, then A1=(213/21/2)A^{-1} = \begin{pmatrix} -2 & 1 \\ 3/2 & -1/2 \end{pmatrix}.

Example 1.2: Formula for 2×2 Inverse

For A=(abcd)A = \begin{pmatrix} a & b \\ c & d \end{pmatrix} with adbc0ad - bc \neq 0:

A1=1adbc(dbca)A^{-1} = \frac{1}{ad - bc} \begin{pmatrix} d & -b \\ -c & a \end{pmatrix}
Remark 1.1: Computational Complexity

Computing A1A^{-1} via Gaussian elimination is O(n3)O(n^3) operations. For large matrices, iterative methods or specialized algorithms may be more efficient.

Theorem 1.3: Properties of Inverse
  1. (A1)1=A(A^{-1})^{-1} = A
  2. (AB)1=B1A1(AB)^{-1} = B^{-1}A^{-1} (order reverses!)
  3. (AT)1=(A1)T(A^T)^{-1} = (A^{-1})^T
  4. (cA)1=1cA1(cA)^{-1} = \frac{1}{c}A^{-1} for c0c \neq 0

2. Elementary Matrices

Elementary matrices are obtained by applying one elementary row operation to the identity matrix. They are the building blocks of Gaussian elimination and provide a factorization of invertible matrices.

Definition 2.1: Elementary Matrix

An elementary matrix is obtained by applying one elementary row operation to II.

Three types:

  • Type I (swap): EijE_{ij} swaps rows ii and jj
  • Type II (scale): Ei(c)E_i(c) multiplies row ii by c0c \neq 0
  • Type III (add): Eij(c)E_{ij}(c) adds cc times row ii to row jj
Theorem 2.1: Elementary Matrices are Invertible

Every elementary matrix is invertible, and its inverse is also elementary.

Theorem 2.2: Row Operations via Multiplication

Left-multiplying by an elementary matrix performs the corresponding row operation:

  • EijAE_{ij}A swaps rows ii and jj of AA
  • Ei(c)AE_i(c)A multiplies row ii of AA by cc
  • Eij(c)AE_{ij}(c)A adds cc times row ii to row jj of AA
Theorem 2.3: Factorization Theorem

An n×nn \times n matrix AA is invertible if and only if it is a product of elementary matrices.

Proof:

If AA is invertible, row reduce AA to II: EkE1A=IE_k \cdots E_1 A = I, so A=E11Ek1A = E_1^{-1} \cdots E_k^{-1}. Conversely, products of invertible matrices are invertible.

Example 2.1: Gaussian Elimination as Matrix Multiplication

If E3E2E1A=UE_3 E_2 E_1 A = U (row echelon form), then A=E11E21E31UA = E_1^{-1} E_2^{-1} E_3^{-1} U. This factorization is the basis for LU decomposition.

Example 2.2: Elementary Matrix Examples

For 3×33 \times 3 matrices:

  • Swap rows 1 and 2: E12=(010100001)E_{12} = \begin{pmatrix} 0 & 1 & 0 \\ 1 & 0 & 0 \\ 0 & 0 & 1 \end{pmatrix}
  • Scale row 2 by 3: E2(3)=(100030001)E_2(3) = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 3 & 0 \\ 0 & 0 & 1 \end{pmatrix}
  • Add 2×row 1 to row 3: E13(2)=(100010201)E_{13}(2) = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 2 & 0 & 1 \end{pmatrix}
Remark 2.1: Inverse of Elementary Matrices

The inverse of an elementary matrix is also elementary:

  • Eij1=EijE_{ij}^{-1} = E_{ij} (swap is its own inverse)
  • Ei(c)1=Ei(1/c)E_i(c)^{-1} = E_i(1/c)
  • Eij(c)1=Eij(c)E_{ij}(c)^{-1} = E_{ij}(-c)

3. Matrix Rank

The rank of a matrix is the dimension of its row space (or column space). It is a fundamental invariant that determines solvability of linear systems and invertibility.

Definition 3.1: Matrix Rank

The rank of a matrix AA, denoted rank(A)\text{rank}(A), is:

  • The dimension of the row space of AA (row rank)
  • The dimension of the column space of AA (column rank)
  • The number of pivots after row reduction
Theorem 3.1: Row Rank Equals Column Rank

For any matrix AA, row rank = column rank.

Proof:

Both equal dim(im TA)\dim(\text{im } T_A) where TA:xAxT_A: x \mapsto Ax is the linear map. Alternatively, row operations preserve row rank and show column rank equals it.

Theorem 3.2: Rank and Invertibility

An n×nn \times n matrix AA is invertible if and only if rank(A)=n\text{rank}(A) = n.

Theorem 3.3: Rank Inequalities
  1. rank(AB)min(rank(A),rank(B))\text{rank}(AB) \leq \min(\text{rank}(A), \text{rank}(B))
  2. rank(A+B)rank(A)+rank(B)\text{rank}(A + B) \leq \text{rank}(A) + \text{rank}(B)
  3. rank(AT)=rank(A)\text{rank}(A^T) = \text{rank}(A)
  4. rank(ATA)=rank(A)\text{rank}(A^T A) = \text{rank}(A)
Example 3.1: Rank and Linear Systems

For Ax=bAx = b:

  • Solvable iff rank(A)=rank([Ab])\text{rank}(A) = \text{rank}([A|b])
  • Unique solution iff additionally rank(A)=n\text{rank}(A) = n
  • Dimension of solution space = nrank(A)n - \text{rank}(A)
Corollary 3.1: Rank-Nullity for Matrices

For AMm×n(F)A \in M_{m \times n}(F), n=rank(A)+dim(kerA)n = \text{rank}(A) + \dim(\ker A), where dim(kerA)\dim(\ker A) is the nullity of AA.

Theorem 3.4: Sylvester's Rank Inequality

For matrices AA and BB of compatible sizes:

rank(A)+rank(B)nrank(AB)min(rank(A),rank(B))\text{rank}(A) + \text{rank}(B) - n \leq \text{rank}(AB) \leq \min(\text{rank}(A), \text{rank}(B))

where nn is the number of columns of AA (or rows of BB).

4. Dual Spaces

The dual space VV^* consists of all linear functionals (linear maps from VV to FF). It has the same dimension as VV and provides a natural way to study vectors through their evaluations.

Definition 4.1: Linear Functional

A linear functional on VV is a linear map ϕ:VF\phi: V \to F.

Definition 4.2: Dual Space

The dual space of VV is V=L(V,F)V^* = \mathcal{L}(V, F), the space of all linear functionals on VV.

Theorem 4.1: Dimension of Dual Space

If dimV=n\dim V = n (finite), then dimV=n\dim V^* = n.

Proof:

V=L(V,F)V^* = \mathcal{L}(V, F), so dimV=dimVdimF=n1=n\dim V^* = \dim V \cdot \dim F = n \cdot 1 = n.

Definition 4.3: Dual Basis

If B={v1,,vn}\mathcal{B} = \{v_1, \ldots, v_n\} is a basis for VV, the dual basis B={v1,,vn}\mathcal{B}^* = \{v_1^*, \ldots, v_n^*\} is defined by vi(vj)=δijv_i^*(v_j) = \delta_{ij} (Kronecker delta).

Theorem 4.2: Dual Basis is a Basis

The dual basis B\mathcal{B}^* is indeed a basis for VV^*.

Definition 4.4: Annihilator

For a subspace WVW \subseteq V, the annihilator is W0={ϕV:ϕ(w)=0 for all wW}W^0 = \{\phi \in V^* : \phi(w) = 0 \text{ for all } w \in W\}.

Theorem 4.3: Dimension of Annihilator

If dimV=n\dim V = n and dimW=k\dim W = k, then dim(W0)=nk\dim(W^0) = n - k.

Definition 4.5: Double Dual

The double dual is V=(V)V^{**} = (V^*)^*. The evaluation map ev:VV\text{ev}: V \to V^{**} sends vv to evv\text{ev}_v where evv(ϕ)=ϕ(v)\text{ev}_v(\phi) = \phi(v).

Theorem 4.4: Natural Isomorphism

For finite-dimensional VV, the evaluation map VVV \to V^{**} is an isomorphism. This is a natural isomorphism (doesn't depend on basis choice).

Definition 4.6: Dual Map

For T:VWT: V \to W, the dual map (or transpose) is T:WVT^*: W^* \to V^* defined by T(ϕ)=ϕTT^*(\phi) = \phi \circ T.

Theorem 4.5: Properties of Dual Map
  1. (ST)=TS(ST)^* = T^* S^* (order reverses!)
  2. ker(T)=(im T)0\ker(T^*) = (\text{im } T)^0
  3. rank(T)=rank(T)\text{rank}(T^*) = \text{rank}(T)
Example 4.1: Dual Space of ℝⁿ

For V=RnV = \mathbb{R}^n, every linear functional ϕ:RnR\phi: \mathbb{R}^n \to \mathbb{R} is of the form:

ϕ(x)=a1x1+a2x2++anxn=aTx\phi(x) = a_1 x_1 + a_2 x_2 + \cdots + a_n x_n = a^T x

for some a=(a1,,an)TRna = (a_1, \ldots, a_n)^T \in \mathbb{R}^n. So (Rn)Rn(\mathbb{R}^n)^* \cong \mathbb{R}^n.

Remark 4.1: Dual Space in Coordinates

If B={v1,,vn}\mathcal{B} = \{v_1, \ldots, v_n\} is a basis of VV, then the dual basis B\mathcal{B}^* provides coordinates for VV^*. For ϕV\phi \in V^*, ϕ=ϕ(vi)vi\phi = \sum \phi(v_i) v_i^*.

5. Applications of Matrix Inverses

Matrix inverses are fundamental for solving linear systems, computing matrix functions, and many other applications in mathematics and engineering.

Theorem 5.1: Solving Linear Systems

For an invertible n×nn \times n matrix AA, the system Ax=bAx = b has the unique solution:

x=A1bx = A^{-1} b
Proof:

Multiply both sides of Ax=bAx = b by A1A^{-1}:

A1(Ax)=A1b    (A1A)x=A1b    Ix=A1b    x=A1bA^{-1}(Ax) = A^{-1}b \implies (A^{-1}A)x = A^{-1}b \implies Ix = A^{-1}b \implies x = A^{-1}b
Example 5.1: Solving a 2×2 System

Solve {x+2y=53x+4y=6\begin{cases} x + 2y = 5 \\ 3x + 4y = 6 \end{cases}.

In matrix form: (1234)(xy)=(56)\begin{pmatrix} 1 & 2 \\ 3 & 4 \end{pmatrix} \begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} 5 \\ 6 \end{pmatrix}

Using the inverse:

(xy)=(213/21/2)(56)=(49/2)\begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} -2 & 1 \\ 3/2 & -1/2 \end{pmatrix} \begin{pmatrix} 5 \\ 6 \end{pmatrix} = \begin{pmatrix} -4 \\ 9/2 \end{pmatrix}
Remark 5.1: Computational Note

While x=A1bx = A^{-1}b is theoretically correct, in practice it's usually more efficient to solve Ax=bAx = b directly via Gaussian elimination rather than computing A1A^{-1} first. However, if you need to solve Ax=biAx = b_i for many different bib_i, computing A1A^{-1} once may be more efficient.

Theorem 5.2: Matrix Inversion and Linear Independence

The columns of an invertible matrix AA form a basis for FnF^n. Conversely, if the columns of AA form a basis, then AA is invertible.

Example 5.2: Change of Coordinates

If B={v1,,vn}\mathcal{B} = \{v_1, \ldots, v_n\} is a basis, the change of basis matrix PP from B\mathcal{B} to standard basis has columns v1,,vnv_1, \ldots, v_n.

Then P1P^{-1} converts standard coordinates to B\mathcal{B}-coordinates: [v]B=P1[v]E[v]_\mathcal{B} = P^{-1}[v]_\mathcal{E}.

Theorem 5.3: Inverse and Matrix Equations

For invertible AA and BB, the matrix equation AX=BAX = B has unique solution X=A1BX = A^{-1}B, and XA=BXA = B has unique solution X=BA1X = BA^{-1}.

Example 5.3: Finding Matrix X

If AX=BAX = B where AA is invertible, then X=A1BX = A^{-1}B.

This is useful for solving systems of matrix equations and finding matrix square roots (when B=AB = A).

Course 7 Practice Quiz
10
Questions
0
Correct
0%
Accuracy
1
If A1A^{-1} exists, what is AA1AA^{-1}?
Easy
Not attempted
2
(AB)1=?(AB)^{-1} = ?
Medium
Not attempted
3
Elementary matrices are always:
Easy
Not attempted
4
The rank of a matrix equals:
Easy
Not attempted
5
Row rank equals column rank:
Medium
Not attempted
6
What is dim(V)\dim(V^*) if dim(V)=n\dim(V) = n?
Easy
Not attempted
7
The dual of a linear map T:VWT: V \to W is T:??T^*: ? \to ?
Medium
Not attempted
8
If dim(V)=n\dim(V) = n and dim(W)=k\dim(W) = k, what is dim(W0)\dim(W^0)?
Hard
Not attempted
9
A square matrix is invertible iff its rank equals:
Easy
Not attempted
10
Every invertible matrix is a product of:
Medium
Not attempted

Frequently Asked Questions

What are the characterizations of invertibility?

For an n×n matrix A, invertibility is equivalent to: det(A) ≠ 0, rank(A) = n, columns/rows linearly independent, Ax = 0 has only trivial solution, Ax = b has unique solution for all b, A is product of elementary matrices, A is row equivalent to I, and more.

How do elementary matrices relate to Gaussian elimination?

Each step of Gaussian elimination is a row operation, which equals left-multiplying by an elementary matrix. The entire process is E_k ⋯ E_1 A = U, so Gaussian elimination is just matrix multiplication!

Why does row rank equal column rank?

This beautiful theorem has multiple proofs. One key insight: both equal dim(im A) = dim(im A^T), viewing A as a linear map. Row operations preserve row rank but show column rank equals it.

What is the dual space and why is it important?

The dual space V* is the space of all linear functionals (linear maps from V to F). It has the same dimension as V and provides a natural way to study vectors through their 'evaluations' on functionals. The double dual V** is naturally isomorphic to V.

How do I compute the annihilator of a subspace?

The annihilator W^0 of a subspace W ⊆ V is the set of all functionals that vanish on W. If dim(V) = n and dim(W) = k, then dim(W^0) = n - k. This follows from the rank-nullity theorem applied to the evaluation map.