MathIsimple
Course 6
Available

Matrix Theory

Matrices are the computational representation of linear maps. Understanding matrix operations and special matrix types is essential for both theory and computation.

1. Matrix Representation of Linear Maps

Definition 6.1: Matrix of a Linear Map

Let T:VWT: V \to W be a linear map, B={v1,,vn}\mathcal{B} = \{v_1, \ldots, v_n\} a basis of VV, and C={w1,,wm}\mathcal{C} = \{w_1, \ldots, w_m\} a basis of WW.

The matrix of TT with respect to B\mathcal{B} and C\mathcal{C}, denoted [T]BC[T]_\mathcal{B}^\mathcal{C}, is the m×nm \times n matrix whose jj-th column is [T(vj)]C[T(v_j)]_\mathcal{C} (the coordinate vector of T(vj)T(v_j) in basis C\mathcal{C}).

Theorem 6.1: Matrix-Vector Multiplication

If [T]BC=A[T]_\mathcal{B}^\mathcal{C} = A and [v]B=x[v]_\mathcal{B} = x, then:

[T(v)]C=Ax[T(v)]_\mathcal{C} = Ax
Example 6.1: Finding Matrix Representation

Let T:R2R2T: \mathbb{R}^2 \to \mathbb{R}^2 be defined by T(x,y)=(x+y,xy)T(x, y) = (x + y, x - y).

Using the standard basis {e1,e2}\{e_1, e_2\}:

  • T(e1)=T(1,0)=(1,1)T(e_1) = T(1, 0) = (1, 1)
  • T(e2)=T(0,1)=(1,1)T(e_2) = T(0, 1) = (1, -1)

So the matrix is:

[T]=(1111)[T] = \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}
Theorem 6.2: Composition and Matrix Multiplication

If T:VWT: V \to W and S:WUS: W \to U are linear maps, then:

[ST]BD=[S]CD[T]BC[S \circ T]_\mathcal{B}^\mathcal{D} = [S]_\mathcal{C}^\mathcal{D} [T]_\mathcal{B}^\mathcal{C}

where B,C,D\mathcal{B}, \mathcal{C}, \mathcal{D} are bases of V,W,UV, W, U respectively.

Remark 1.1: Matrix Representation is Basis-Dependent

The matrix representation of a linear map depends on the choice of bases. Changing bases changes the matrix, but the underlying linear map remains the same.

Example 1.2: Matrix of Zero and Identity Maps
  • The zero map 0:VW0: V \to W has matrix representation 00 (zero matrix) in any bases
  • The identity map IV:VVI_V: V \to V has matrix representation InI_n (identity matrix) in any basis

2. Matrix Operations

Definition 6.2: Matrix Addition and Scalar Multiplication

For m×nm \times n matrices AA and BB over field FF:

  • Addition: (A+B)ij=Aij+Bij(A + B)_{ij} = A_{ij} + B_{ij}
  • Scalar multiplication: (αA)ij=αAij(\alpha A)_{ij} = \alpha A_{ij}
Definition 6.3: Matrix Multiplication

If AA is m×nm \times n and BB is n×pn \times p, then ABAB is m×pm \times p with:

(AB)ij=k=1nAikBkj(AB)_{ij} = \sum_{k=1}^n A_{ik} B_{kj}

This is the dot product of row ii of AA with column jj of BB.

Theorem 6.3: Properties of Matrix Multiplication

Matrix multiplication satisfies:

  1. Associativity: (AB)C=A(BC)(AB)C = A(BC)
  2. Distributivity: A(B+C)=AB+ACA(B + C) = AB + AC and (A+B)C=AC+BC(A + B)C = AC + BC
  3. Scalar compatibility: α(AB)=(αA)B=A(αB)\alpha(AB) = (\alpha A)B = A(\alpha B)
  4. Identity: AI=IA=AAI = IA = A where II is the identity matrix

Note: Matrix multiplication is NOT commutative in general.

Example 6.2: Non-Commutativity

Let A=(1101)A = \begin{pmatrix} 1 & 1 \\ 0 & 1 \end{pmatrix} and B=(1011)B = \begin{pmatrix} 1 & 0 \\ 1 & 1 \end{pmatrix}.

Then:

AB=(2111),BA=(1112)AB = \begin{pmatrix} 2 & 1 \\ 1 & 1 \end{pmatrix}, \quad BA = \begin{pmatrix} 1 & 1 \\ 1 & 2 \end{pmatrix}

So ABBAAB \neq BA.

Definition 6.4: Transpose

The transpose of an m×nm \times n matrix AA, denoted ATA^T, is the n×mn \times m matrix with:

(AT)ij=Aji(A^T)_{ij} = A_{ji}
Theorem 6.4: Properties of Transpose

For matrices AA and BB and scalar α\alpha:

  1. (AT)T=A(A^T)^T = A
  2. (A+B)T=AT+BT(A + B)^T = A^T + B^T
  3. (αA)T=αAT(\alpha A)^T = \alpha A^T
  4. (AB)T=BTAT(AB)^T = B^T A^T (note the order reversal)
Theorem 2.1: Matrix Multiplication and Linear Combinations

For AMm×n(F)A \in M_{m \times n}(F) and x=(x1,,xn)Tx = (x_1, \ldots, x_n)^T:

Ax=x1A1+x2A2++xnAnAx = x_1 A_1 + x_2 A_2 + \cdots + x_n A_n

where AiA_i is the ii-th column of AA. So AxAx is a linear combination of the columns of AA.

Example 2.1: Column Space

The column space of AA is span{A1,,An}\text{span}\{A_1, \ldots, A_n\}, which equals im(TA)\text{im}(T_A) where TA(x)=AxT_A(x) = Ax.

3. Special Matrices

Definition 6.5: Diagonal Matrix

A square matrix DD is diagonal if Dij=0D_{ij} = 0 for all iji \neq j.

We write D=diag(d1,d2,,dn)D = \text{diag}(d_1, d_2, \ldots, d_n).

Definition 6.6: Triangular Matrices

A square matrix AA is:

  • Upper triangular if Aij=0A_{ij} = 0 for all i>ji > j
  • Lower triangular if Aij=0A_{ij} = 0 for all i<ji < j
Definition 6.7: Symmetric and Skew-Symmetric

A square matrix AA is:

  • Symmetric if A=ATA = A^T
  • Skew-symmetric if A=ATA = -A^T
Theorem 6.5: Symmetric-Skew Decomposition

Every square matrix AA can be uniquely written as:

A=S+KA = S + K

where S=12(A+AT)S = \frac{1}{2}(A + A^T) is symmetric and K=12(AAT)K = \frac{1}{2}(A - A^T) is skew-symmetric.

Definition 6.8: Orthogonal Matrix

A square matrix QQ is orthogonal if:

QTQ=IQ^T Q = I

Equivalently, Q1=QTQ^{-1} = Q^T.

Theorem 6.6: Properties of Orthogonal Matrices

For an orthogonal matrix QQ:

  1. The columns (and rows) form an orthonormal set
  2. det(Q)=±1\det(Q) = \pm 1
  3. QQ preserves lengths: Qx=x\|Qx\| = \|x\|
  4. QQ preserves angles: Qx,Qy=x,y\langle Qx, Qy \rangle = \langle x, y \rangle
Definition 6.9: Idempotent and Nilpotent

A square matrix PP is:

  • Idempotent if P2=PP^2 = P
  • Nilpotent if Pk=0P^k = 0 for some positive integer kk
Example 6.3: Special Matrix Examples
  • Diagonal: diag(2,3,5)\text{diag}(2, 3, 5)
  • Symmetric: (1223)\begin{pmatrix} 1 & 2 \\ 2 & 3 \end{pmatrix}
  • Orthogonal: (cosθsinθsinθcosθ)\begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} (rotation)
  • Idempotent: (1000)\begin{pmatrix} 1 & 0 \\ 0 & 0 \end{pmatrix} (projection onto x-axis)
Definition 3.1: Unitary and Hermitian Matrices

For complex matrices:

  • Unitary: QQ=IQ^* Q = I where Q=QTQ^* = \overline{Q^T} (conjugate transpose)
  • Hermitian: A=AA = A^* (self-adjoint)
Theorem 3.1: Properties of Special Matrices

  • Product of orthogonal matrices is orthogonal
  • Product of upper (lower) triangular matrices is upper (lower) triangular
  • Inverse of orthogonal matrix is orthogonal
  • Transpose of symmetric matrix is symmetric

4. Change of Basis for Matrices

When we change bases, the matrix representation of a linear map changes. Understanding this relationship is crucial for diagonalization, similarity, and many other applications.

Theorem 4.1: Change of Basis Formula

Let T:VVT: V \to V be a linear operator, and let B\mathcal{B} and C\mathcal{C} be two bases of VV. If PP is the change of basis matrix from B\mathcal{B} to C\mathcal{C}, then:

[T]C=P1[T]BP[T]_\mathcal{C} = P^{-1} [T]_\mathcal{B} P
Proof:

For vVv \in V, let [v]B[v]_\mathcal{B} and [v]C[v]_\mathcal{C} be coordinate vectors. Then:

[T(v)]C=[T]C[v]C=[T]CP[v]B[T(v)]_\mathcal{C} = [T]_\mathcal{C} [v]_\mathcal{C} = [T]_\mathcal{C} P [v]_\mathcal{B}

Also:

[T(v)]C=P[T(v)]B=P[T]B[v]B[T(v)]_\mathcal{C} = P [T(v)]_\mathcal{B} = P [T]_\mathcal{B} [v]_\mathcal{B}

Since this holds for all vv, we get [T]CP=P[T]B[T]_\mathcal{C} P = P [T]_\mathcal{B}, so [T]C=P[T]BP1[T]_\mathcal{C} = P [T]_\mathcal{B} P^{-1}.

Definition 4.1: Similar Matrices

Two n×nn \times n matrices AA and BB are similar if there exists an invertible matrix PP such that:

B=P1APB = P^{-1} A P

Similar matrices represent the same linear operator in different bases.

Theorem 4.2: Properties of Similarity

Similarity is an equivalence relation:

  1. Reflexive: AAA \sim A (take P=IP = I)
  2. Symmetric: If B=P1APB = P^{-1} A P, then A=(P1)1BP1A = (P^{-1})^{-1} B P^{-1}
  3. Transitive: If B=P1APB = P^{-1} A P and C=Q1BQC = Q^{-1} B Q, then C=(PQ)1A(PQ)C = (PQ)^{-1} A (PQ)
Theorem 4.3: Similar Matrices Share Properties

If ABA \sim B, then:

  • det(A)=det(B)\det(A) = \det(B)
  • rank(A)=rank(B)\text{rank}(A) = \text{rank}(B)
  • tr(A)=tr(B)\text{tr}(A) = \text{tr}(B) (trace)
  • AA and BB have the same characteristic polynomial and eigenvalues
Example 4.1: Change of Basis Example

Let T:R2R2T: \mathbb{R}^2 \to \mathbb{R}^2 be T(x,y)=(x+y,xy)T(x, y) = (x + y, x - y).

In standard basis E\mathcal{E}: [T]E=(1111)[T]_\mathcal{E} = \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}

In basis B={(1,1),(1,1)}\mathcal{B} = \{(1, 1), (1, -1)\}:

Change of basis matrix: P=(1111)P = \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}

Then [T]B=P1[T]EP=(2002)[T]_\mathcal{B} = P^{-1} [T]_\mathcal{E} P = \begin{pmatrix} 2 & 0 \\ 0 & -2 \end{pmatrix} (diagonal!)

5. Matrix Powers and Polynomials

We can raise matrices to powers and evaluate polynomials at matrices. These operations are fundamental for solving matrix equations and understanding matrix functions.

Definition 5.1: Matrix Powers

For a square matrix AA and positive integer kk:

Ak=AAAk timesA^k = \underbrace{A \cdot A \cdots A}_{k \text{ times}}

We define A0=IA^0 = I (identity matrix).

Theorem 5.1: Properties of Matrix Powers

For square matrices AA and integers m,n0m, n \geq 0:

  1. AmAn=Am+nA^m A^n = A^{m+n}
  2. (Am)n=Amn(A^m)^n = A^{mn}
  3. If AA is invertible, (A1)k=(Ak)1=Ak(A^{-1})^k = (A^k)^{-1} = A^{-k}
Definition 5.2: Matrix Polynomial

For a polynomial p(x)=a0+a1x++anxnp(x) = a_0 + a_1 x + \cdots + a_n x^n and a square matrix AA, define:

p(A)=a0I+a1A+a2A2++anAnp(A) = a_0 I + a_1 A + a_2 A^2 + \cdots + a_n A^n
Example 5.1: Matrix Polynomial Example

If p(x)=x23x+2p(x) = x^2 - 3x + 2 and A=(1201)A = \begin{pmatrix} 1 & 2 \\ 0 & 1 \end{pmatrix}, then:

p(A)=A23A+2I=(1401)3(1201)+2(1001)=(0200)p(A) = A^2 - 3A + 2I = \begin{pmatrix} 1 & 4 \\ 0 & 1 \end{pmatrix} - 3\begin{pmatrix} 1 & 2 \\ 0 & 1 \end{pmatrix} + 2\begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} = \begin{pmatrix} 0 & -2 \\ 0 & 0 \end{pmatrix}
Theorem 5.2: Polynomial Evaluation Preserves Operations

For polynomials pp and qq, and square matrix AA:

  • (p+q)(A)=p(A)+q(A)(p + q)(A) = p(A) + q(A)
  • (pq)(A)=p(A)q(A)(pq)(A) = p(A) q(A)
  • If A=P1BPA = P^{-1} B P, then p(A)=P1p(B)Pp(A) = P^{-1} p(B) P
Example 5.2: Computing High Powers

If A=P1DPA = P^{-1} D P with DD diagonal, then:

Ak=(P1DP)k=P1DkPA^k = (P^{-1} D P)^k = P^{-1} D^k P

Since DkD^k is just the diagonal matrix with entries raised to the kk-th power, this makes computing AkA^k much easier.

6. Block Matrices

Block matrices partition a matrix into submatrices (blocks). This perspective simplifies many computations and reveals structural properties.

Definition 6.1: Block Matrix

A block matrix (or partitioned matrix) is a matrix written as:

A=(A11A12A1nA21A22A2nAm1Am2Amn)A = \begin{pmatrix} A_{11} & A_{12} & \cdots & A_{1n} \\ A_{21} & A_{22} & \cdots & A_{2n} \\ \vdots & \vdots & \ddots & \vdots \\ A_{m1} & A_{m2} & \cdots & A_{mn} \end{pmatrix}

where each AijA_{ij} is a submatrix (block) of appropriate size.

Theorem 6.1: Block Matrix Multiplication

If AA and BB are block matrices with compatible block sizes, then ABAB can be computed block-wise:

(AB)ij=kAikBkj(AB)_{ij} = \sum_k A_{ik} B_{kj}

provided the block dimensions are compatible for multiplication.

Example 6.1: Block Multiplication

For A=(A1A2A3A4)A = \begin{pmatrix} A_1 & A_2 \\ A_3 & A_4 \end{pmatrix} and B=(B1B2B3B4)B = \begin{pmatrix} B_1 & B_2 \\ B_3 & B_4 \end{pmatrix}:

AB=(A1B1+A2B3A1B2+A2B4A3B1+A4B3A3B2+A4B4)AB = \begin{pmatrix} A_1 B_1 + A_2 B_3 & A_1 B_2 + A_2 B_4 \\ A_3 B_1 + A_4 B_3 & A_3 B_2 + A_4 B_4 \end{pmatrix}
Definition 6.2: Block Diagonal Matrix

A block diagonal matrix has the form:

diag(A1,A2,,Ak)=(A1000A2000Ak)\text{diag}(A_1, A_2, \ldots, A_k) = \begin{pmatrix} A_1 & 0 & \cdots & 0 \\ 0 & A_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & A_k \end{pmatrix}

where AiA_i are square matrices and off-diagonal blocks are zero.

Theorem 6.2: Determinant of Block Diagonal Matrix

If A=diag(A1,,Ak)A = \text{diag}(A_1, \ldots, A_k) is block diagonal, then:

det(A)=i=1kdet(Ai)\det(A) = \prod_{i=1}^k \det(A_i)
Example 6.2: Block Upper Triangular

A block upper triangular matrix has the form:

(AB0C)\begin{pmatrix} A & B \\ 0 & C \end{pmatrix}

where AA and CC are square blocks and the lower-left block is zero.

Its determinant is det(A)det(C)\det(A) \det(C).

Remark 6.1: Applications of Block Matrices

Block matrices are useful for:

  • Partitioning large matrices for parallel computation
  • Understanding direct sum decompositions
  • Simplifying determinant and inverse computations
  • Analyzing structured matrices (e.g., block tridiagonal)

Frequently Asked Questions

Why does matrix multiplication work the way it does?

It's designed so that [S ∘ T] = [S][T]—the matrix of a composition is the product of matrices. The row-column dot product formula follows directly from how coordinates transform under composition of linear maps.

Why isn't matrix multiplication commutative?

Geometrically: rotate then reflect ≠ reflect then rotate. Algebraically: the composition of linear maps isn't commutative, and matrix multiplication represents composition. Even when both AB and BA are defined, they usually give different results.

What is the relationship between similar matrices?

Similar matrices represent the same linear operator in different bases. They share many properties: determinant, trace, eigenvalues, characteristic polynomial, and rank. If B = P⁻¹AP, then A and B are similar, connected by the change of basis matrix P.

Why are special matrices important?

Special matrices have structure that simplifies computations and reveals properties. Diagonal matrices are easy to invert and power, triangular matrices make solving systems efficient, and symmetric matrices have real eigenvalues with orthogonal eigenvectors.

What's the geometric meaning of an orthogonal matrix?

Orthogonal matrices represent isometries—transformations that preserve lengths and angles. In 2D/3D, these are rotations (det = 1) and reflections (det = -1). They preserve the dot product: ⟨Ax, Ay⟩ = ⟨x, y⟩.

Matrix Theory Practice
10
Questions
0
Correct
0%
Accuracy
1
If AA is 3×43 \times 4 and BB is 4×24 \times 2, what size is ABAB?
Easy
Not attempted
2
Is matrix multiplication commutative?
Easy
Not attempted
3
(AB)T=?(AB)^T = ?
Medium
Not attempted
4
If AA is m×nm \times n, what size is ATA^T?
Easy
Not attempted
5
A matrix AA is symmetric if and only if:
Easy
Not attempted
6
If AA is orthogonal, then A1=?A^{-1} = ?
Easy
Not attempted
7
The product of two upper triangular matrices is:
Medium
Not attempted
8
If T:R2R2T: \mathbb{R}^2 \to \mathbb{R}^2 is represented by AA in the standard basis, what represents TT in a new basis BB?
Medium
Not attempted
9
What is the (i,j)(i,j) entry of ABAB?
Medium
Not attempted
10
The matrix of the identity map I:VVI: V \to V in any basis is:
Easy
Not attempted