MathIsimple
LA-4.6
Available
3-4 hours

Special Matrices

Certain matrices possess special structure that makes them particularly important in theory and applications. We study diagonal, triangular, symmetric, orthogonal, idempotent, and nilpotent matrices—each with distinctive properties that simplify computations and illuminate the underlying linear transformations.

9 sections15+ definitions12+ examples
Learning Objectives
  • Define and recognize diagonal, triangular, symmetric, and skew-symmetric matrices
  • Prove and apply properties of symmetric and skew-symmetric matrices
  • Understand orthogonal matrices and their geometric meaning
  • Characterize idempotent matrices as projection operators
  • Study nilpotent matrices and their index of nilpotency
  • Understand the relationship between special matrices and linear transformations
  • Decompose matrices into symmetric and skew-symmetric parts
  • Recognize when matrices of special types are invertible
  • Apply special matrix properties to simplify computations
  • Connect special matrices to eigenvalue theory
Prerequisites
  • Matrix operations and transpose (LA-4.2)
  • Matrix inverse (LA-4.3)
  • Elementary matrices (LA-4.4)
  • Basic properties of linear maps

1. Diagonal Matrices

Definition 4.6.1: Diagonal Matrix

A square matrix D=(dij)Mn(F)D = (d_{ij}) \in M_n(F) is diagonal if dij=0d_{ij} = 0 whenever iji \neq j.

D=(d1000d2000dn)=diag(d1,d2,,dn)D = \begin{pmatrix} d_1 & 0 & \cdots & 0 \\ 0 & d_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & d_n \end{pmatrix} = \text{diag}(d_1, d_2, \ldots, d_n)
Theorem 4.6.1: Properties of Diagonal Matrices

Let D=diag(d1,,dn)D = \text{diag}(d_1, \ldots, d_n) and E=diag(e1,,en)E = \text{diag}(e_1, \ldots, e_n). Then:

  1. D+E=diag(d1+e1,,dn+en)D + E = \text{diag}(d_1 + e_1, \ldots, d_n + e_n)
  2. DE=diag(d1e1,,dnen)DE = \text{diag}(d_1 e_1, \ldots, d_n e_n)
  3. Dk=diag(d1k,,dnk)D^k = \text{diag}(d_1^k, \ldots, d_n^k) for any k0k \geq 0
  4. DD is invertible iff all di0d_i \neq 0, with D1=diag(d11,,dn1)D^{-1} = \text{diag}(d_1^{-1}, \ldots, d_n^{-1})
  5. det(D)=d1d2dn\det(D) = d_1 d_2 \cdots d_n
  6. The eigenvalues of DD are d1,d2,,dnd_1, d_2, \ldots, d_n
Proof:

All properties follow from direct computation. For (2): (DE)ij=kDikEkj(DE)_{ij} = \sum_k D_{ik} E_{kj}. Since both are diagonal, only k=i=jk = i = j contributes, giving dieid_i e_i.

For (6): Dei=dieiDe_i = d_i e_i where eie_i is the ii-th standard basis vector.

Example 4.6.1: Diagonal Matrix Operations

Let D=diag(2,1,3)D = \text{diag}(2, -1, 3) and E=diag(4,2,1)E = \text{diag}(4, 2, -1).

DE=diag(8,2,3),D2=diag(4,1,9)DE = \text{diag}(8, -2, -3), \quad D^2 = \text{diag}(4, 1, 9)
D1=diag(1/2,1,1/3),det(D)=2(1)3=6D^{-1} = \text{diag}(1/2, -1, 1/3), \quad \det(D) = 2 \cdot (-1) \cdot 3 = -6
Remark 4.6.1: Scalar Matrices

A scalar matrix λI\lambda I is diagonal with all entries equal. Scalar matrices commute with all matrices:

(λI)A=λA=A(λI)(\lambda I) A = \lambda A = A(\lambda I)

In fact, scalar matrices are the only matrices that commute with all ntimesnn \\times n matrices.

Example 4.6.1b: Matrix Functions of Diagonal Matrices

For D=diag(1,2,3)D = \text{diag}(1, 2, 3), compute eDe^D:

eD=diag(e1,e2,e3)=diag(e,e2,e3)e^D = \text{diag}(e^1, e^2, e^3) = \text{diag}(e, e^2, e^3)

In general, f(D)=diag(f(d1),,f(dn))f(D) = \text{diag}(f(d_1), \ldots, f(d_n)) for any function ff.

Example 4.6.1c: Diagonal Matrix Power

For D=diag(2,3,1)D = \text{diag}(2, 3, -1), compute D100D^{100}:

D100=diag(2100,3100,(1)100)=diag(2100,3100,1)D^{100} = \text{diag}(2^{100}, 3^{100}, (-1)^{100}) = \text{diag}(2^{100}, 3^{100}, 1)

This is trivial for diagonal matrices but would be hard to compute directly for non-diagonal matrices.

Theorem 4.6.1b: Diagonal Matrices Commute

Any two diagonal matrices commute: if D1,D2D_1, D_2 are diagonal, then D1D2=D2D1D_1 D_2 = D_2 D_1.

Proof:

(D1D2)ij=(d1)i(d2)jδij=(d2)j(d1)iδij=(D2D1)ij(D_1 D_2)_{ij} = (d_1)_i (d_2)_j \delta_{ij} = (d_2)_j (d_1)_i \delta_{ij} = (D_2 D_1)_{ij}.

Remark 4.6.1b: Block Diagonal Matrices

A block diagonal matrix has square blocks along the diagonal:

diag(A1,A2,,Ak)=(A1000A2000Ak)\text{diag}(A_1, A_2, \ldots, A_k) = \begin{pmatrix} A_1 & 0 & \cdots & 0 \\ 0 & A_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & A_k \end{pmatrix}

Properties: det=idet(Ai)\det = \prod_i \det(A_i), inverse exists iff each block is invertible.

Example 4.6.1d: Block Diagonal Operations

For M=diag(A,B)M = \text{diag}(A, B) where AA is 2×22 \times 2 and BB is 3×33 \times 3:

M1=diag(A1,B1),Mn=diag(An,Bn)M^{-1} = \text{diag}(A^{-1}, B^{-1}), \quad M^n = \text{diag}(A^n, B^n)

2. Triangular Matrices

Definition 4.6.2: Upper and Lower Triangular Matrices

A matrix A=(aij)A = (a_{ij}) is:

  • Upper triangular if aij=0a_{ij} = 0 for i>ji > j (zeros below diagonal)
  • Lower triangular if aij=0a_{ij} = 0 for i<ji < j (zeros above diagonal)
  • Strictly triangular if also aii=0a_{ii} = 0 for all ii
Upper: (000),Lower: (000)\text{Upper: } \begin{pmatrix} * & * & * \\ 0 & * & * \\ 0 & 0 & * \end{pmatrix}, \quad \text{Lower: } \begin{pmatrix} * & 0 & 0 \\ * & * & 0 \\ * & * & * \end{pmatrix}
Theorem 4.6.2: Properties of Triangular Matrices

Let UU, U1U_1, U2U_2 be upper triangular. Then:

  1. U1+U2U_1 + U_2 is upper triangular
  2. U1U2U_1 U_2 is upper triangular
  3. If UU is invertible, U1U^{-1} is upper triangular
  4. det(U)=u11u22unn\det(U) = u_{11} u_{22} \cdots u_{nn}
  5. The eigenvalues of UU are its diagonal entries
Proof:

For (2): Let W=U1U2W = U_1 U_2. For i>ji > j:

wij=k=1n(u1)ik(u2)kjw_{ij} = \sum_{k=1}^n (u_1)_{ik} (u_2)_{kj}

When k<ik < i, (u1)ik=0(u_1)_{ik} = 0. When ki>jk \geq i > j, (u2)kj=0(u_2)_{kj} = 0. So wij=0w_{ij} = 0.

Example 4.6.2: Triangular Matrix Product
(213012004)(101031002)=(237035008)\begin{pmatrix} 2 & 1 & 3 \\ 0 & -1 & 2 \\ 0 & 0 & 4 \end{pmatrix} \begin{pmatrix} 1 & 0 & 1 \\ 0 & 3 & -1 \\ 0 & 0 & 2 \end{pmatrix} = \begin{pmatrix} 2 & 3 & 7 \\ 0 & -3 & 5 \\ 0 & 0 & 8 \end{pmatrix}

The product is upper triangular with diagonal entries (2,3,8)(2, -3, 8).

Corollary 4.6.1: Invertibility of Triangular Matrices

A triangular matrix is invertible if and only if all diagonal entries are nonzero.

Remark 4.6.2: LU Decomposition

Many matrices can be factored as A=LUA = LU where LL is lower triangular and UU is upper triangular. This is fundamental for solving linear systems efficiently.

Example 4.6.2b: Product of Diagonal Entries

For U=(312025001)U = \begin{pmatrix} 3 & 1 & 2 \\ 0 & 2 & 5 \\ 0 & 0 & -1 \end{pmatrix}:

det(U)=32(1)=6\det(U) = 3 \cdot 2 \cdot (-1) = -6

Eigenvalues are 3, 2, -1 (the diagonal entries).

Example 4.6.2c: LU Factorization
(2145)=(1021)(2103)\begin{pmatrix} 2 & 1 \\ 4 & 5 \end{pmatrix} = \begin{pmatrix} 1 & 0 \\ 2 & 1 \end{pmatrix} \begin{pmatrix} 2 & 1 \\ 0 & 3 \end{pmatrix}

Here LL has 1s on diagonal (unit lower triangular), UU is upper triangular.

Theorem 4.6.2b: Inverse of Triangular Matrix

If UU is upper triangular and invertible, then U1U^{-1} can be computed in O(n2)O(n^2) operations using back-substitution, rather than O(n3)O(n^3) for general matrices.

Remark 4.6.2b: Block Triangular Matrices

A block upper triangular matrix has the form:

(A11A12A130A22A2300A33)\begin{pmatrix} A_{11} & A_{12} & A_{13} \\ 0 & A_{22} & A_{23} \\ 0 & 0 & A_{33} \end{pmatrix}

Its determinant is det(A11)det(A22)det(A33)\det(A_{11}) \det(A_{22}) \det(A_{33}).

3. Symmetric and Skew-Symmetric Matrices

Definition 4.6.3: Symmetric Matrix

A matrix AMn(F)A \in M_n(F) is symmetric if A=ATA = A^T, i.e., aij=ajia_{ij} = a_{ji} for all i,ji, j.

Definition 4.6.4: Skew-Symmetric Matrix

A matrix AMn(F)A \in M_n(F) is skew-symmetric if A=ATA = -A^T, i.e., aij=ajia_{ij} = -a_{ji}.

Theorem 4.6.3: Properties of Symmetric Matrices

Let AA, BB be symmetric. Then:

  1. A+BA + B is symmetric
  2. λA\lambda A is symmetric for any scalar λ\lambda
  3. AkA^k is symmetric for any k0k \geq 0
  4. If AA is invertible, A1A^{-1} is symmetric
  5. ABAB is symmetric iff AB=BAAB = BA
Proof:

For (4): (A1)T=(AT)1=A1(A^{-1})^T = (A^T)^{-1} = A^{-1}.

For (5): (AB)T=BTAT=BA(AB)^T = B^T A^T = BA. This equals ABAB iff they commute.

Theorem 4.6.4: Properties of Skew-Symmetric Matrices

Let AA be skew-symmetric. Then:

  1. All diagonal entries of AA are zero
  2. A2A^2 is symmetric
  3. If nn is odd, then det(A)=0\det(A) = 0
  4. xTAx=0x^T A x = 0 for all xRnx \in \mathbb{R}^n
Proof:

For (1): aii=aiia_{ii} = -a_{ii} implies aii=0a_{ii} = 0.

For (3): det(A)=det(AT)=(1)ndet(A)\det(A) = \det(-A^T) = (-1)^n \det(A). If nn is odd, det(A)=det(A)\det(A) = -\det(A).

Theorem 4.6.5: Symmetric-Skew Decomposition

Every matrix AA can be uniquely written as A=S+KA = S + K where:

S=12(A+AT) (symmetric),K=12(AAT) (skew-symmetric)S = \frac{1}{2}(A + A^T) \text{ (symmetric)}, \quad K = \frac{1}{2}(A - A^T) \text{ (skew-symmetric)}
Example 4.6.3: Symmetric-Skew Decomposition

For A=(1234)A = \begin{pmatrix} 1 & 2 \\ 3 & 4 \end{pmatrix}:

S=(15/25/24),K=(01/21/20)S = \begin{pmatrix} 1 & 5/2 \\ 5/2 & 4 \end{pmatrix}, \quad K = \begin{pmatrix} 0 & -1/2 \\ 1/2 & 0 \end{pmatrix}
Example 4.6.3b: 3×3 Symmetric Matrix
A=(123254346)A = \begin{pmatrix} 1 & 2 & 3 \\ 2 & 5 & 4 \\ 3 & 4 & 6 \end{pmatrix}

This is symmetric: each entry equals its reflection across the diagonal. All eigenvalues will be real.

Example 4.6.3c: 3×3 Skew-Symmetric Matrix
B=(023204340)B = \begin{pmatrix} 0 & 2 & -3 \\ -2 & 0 & 4 \\ 3 & -4 & 0 \end{pmatrix}

Note the zero diagonal and bij=bjib_{ij} = -b_{ji}. Since n=3n = 3 is odd, det(B)=0\det(B) = 0.

Theorem 4.6.5b: Quadratic Forms and Symmetric Matrices

Every quadratic form Q(x)=xTAxQ(x) = x^T A x can be written using a symmetric matrix:

Q(x)=xT(A+AT2)xQ(x) = x^T \left(\frac{A + A^T}{2}\right) x

The skew-symmetric part contributes nothing since xTKx=0x^T K x = 0 for skew-symmetric KK.

Remark 4.6.3c: Spectral Theorem Preview

The Spectral Theorem states: every real symmetric matrix AA is orthogonally diagonalizable:

A=QΛQTA = Q \Lambda Q^T

where QQ is orthogonal and Λ\Lambda is diagonal. This is one of the most important theorems in linear algebra.

Example 4.6.3d: Spectral Decomposition

For A=(3113)A = \begin{pmatrix} 3 & 1 \\ 1 & 3 \end{pmatrix}:

Eigenvalues: λ=4,2\lambda = 4, 2. Eigenvectors: (1,1)T,(1,1)T(1,1)^T, (1,-1)^T.

A=12(1111)(4002)(1111)A = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix} \begin{pmatrix} 4 & 0 \\ 0 & 2 \end{pmatrix} \begin{pmatrix} 1 & 1 \\ 1 & -1 \end{pmatrix}

4. Orthogonal Matrices

Definition 4.6.5: Orthogonal Matrix

A real matrix QMn(R)Q \in M_n(\mathbb{R}) is orthogonal if QTQ=QQT=IQ^T Q = Q Q^T = I, equivalently Q1=QTQ^{-1} = Q^T.

Theorem 4.6.6: Characterizations of Orthogonal Matrices

The following are equivalent for QMn(R)Q \in M_n(\mathbb{R}):

  1. QQ is orthogonal
  2. The columns of QQ form an orthonormal basis
  3. The rows of QQ form an orthonormal basis
  4. QQ preserves dot products: Qx,Qy=x,y\langle Qx, Qy \rangle = \langle x, y \rangle
  5. QQ preserves lengths: Qx=x\|Qx\| = \|x\|
Theorem 4.6.7: Properties of Orthogonal Matrices
  1. det(Q)=±1\det(Q) = \pm 1
  2. Products of orthogonal matrices are orthogonal
  3. Q1=QTQ^{-1} = Q^T is orthogonal
  4. All eigenvalues λ\lambda satisfy λ=1|\lambda| = 1
Proof:

For (1): (detQ)2=det(QTQ)=det(I)=1(\det Q)^2 = \det(Q^T Q) = \det(I) = 1.

For (4): If Qv=λvQv = \lambda v, then v=Qv=λv\|v\| = \|Qv\| = |\lambda| \|v\|.

Example 4.6.4: 2D Rotation Matrix

Rotation by angle θ\theta counterclockwise:

Rθ=(cosθsinθsinθcosθ)R_\theta = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}

det(Rθ)=cos2θ+sin2θ=1\det(R_\theta) = \cos^2\theta + \sin^2\theta = 1 (proper orthogonal = rotation).

Example 4.6.5: Reflection Matrix

Reflection across the line y=xy = x:

S=(0110)S = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}

S2=IS^2 = I and det(S)=1\det(S) = -1 (improper orthogonal = reflection).

Example 4.6.5b: 3D Rotation Matrix

Rotation by θ\theta around the zz-axis:

Rz(θ)=(cosθsinθ0sinθcosθ0001)R_z(\theta) = \begin{pmatrix} \cos\theta & -\sin\theta & 0 \\ \sin\theta & \cos\theta & 0 \\ 0 & 0 & 1 \end{pmatrix}

This is orthogonal with det=1\det = 1. The axis of rotation is the zz-axis (eigenvector with eigenvalue 1).

Example 4.6.5c: Householder Reflection

Reflection through a hyperplane perpendicular to unit vector uu:

H=I2uuTH = I - 2uu^T

Verify: HTH=(I2uuT)2=I4uuT+4uuTuuT=IH^T H = (I - 2uu^T)^2 = I - 4uu^T + 4uu^T uu^T = I (using uTu=1u^T u = 1).

Also det(H)=1\det(H) = -1 and H2=IH^2 = I.

Theorem 4.6.7b: Structure of 2×2 Orthogonal Matrices

Every 2×22 \times 2 orthogonal matrix is either:

Rotation: (cosθsinθsinθcosθ)orReflection: (cosθsinθsinθcosθ)\text{Rotation: } \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} \quad \text{or} \quad \text{Reflection: } \begin{pmatrix} \cos\theta & \sin\theta \\ \sin\theta & -\cos\theta \end{pmatrix}
Remark 4.6.4b: Orthogonal Group O(n)

The set of ntimesnn \\times n orthogonal matrices forms a group under multiplication:

  • Closed: (Q1Q2)T(Q1Q2)=Q2TQ1TQ1Q2=I(Q_1 Q_2)^T (Q_1 Q_2) = Q_2^T Q_1^T Q_1 Q_2 = I
  • Identity: II is orthogonal
  • Inverses: Q1=QTQ^{-1} = Q^T is orthogonal

The subgroup SO(n)SO(n) with det=1\det = 1 consists of rotations only.

Proof:

For characterization (4) ⇒ (5): Set y=xy = x to get Qx2=Qx,Qx=x,x=x2\|Qx\|^2 = \langle Qx, Qx \rangle = \langle x, x \rangle = \|x\|^2.

For (5) ⇒ (4): Use polarization identity:

x,y=14(x+y2xy2)\langle x, y \rangle = \frac{1}{4}(\|x+y\|^2 - \|x-y\|^2)

5. Idempotent Matrices

Definition 4.6.6: Idempotent Matrix

A matrix PP is idempotent if P2=PP^2 = P.

Theorem 4.6.8: Properties of Idempotent Matrices
  1. Pk=PP^k = P for all k1k \geq 1
  2. Eigenvalues are only 0 and 1
  3. IPI - P is also idempotent
  4. rank(P)=tr(P)\text{rank}(P) = \text{tr}(P)
  5. Rn=Im(P)ker(P)\mathbb{R}^n = \text{Im}(P) \oplus \ker(P)
Proof:

For (2): If Pv=λvPv = \lambda v, then λv=Pv=P2v=λ2v\lambda v = Pv = P^2 v = \lambda^2 v, so λ=λ2\lambda = \lambda^2.

For (3): (IP)2=I2P+P2=I2P+P=IP(I-P)^2 = I - 2P + P^2 = I - 2P + P = I - P.

Example 4.6.6: Projection onto a Line

Project onto span{(1,1)T}\text{span}\{(1,1)^T\}:

P=12(1111)P = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & 1 \end{pmatrix}

Verify: P2=PP^2 = P, eigenvalues are 0 and 1, trace = 1 = rank.

Remark 4.6.3: Orthogonal Projection

An idempotent PP is an orthogonal projection iff P=PTP = P^T.

Theorem 4.6.8b: Complementary Projections

If PP is idempotent, then:

  1. Im(P)=ker(IP)\text{Im}(P) = \ker(I - P) and ker(P)=Im(IP)\ker(P) = \text{Im}(I - P)
  2. PP and IPI - P are complementary projections
  3. P(IP)=(IP)P=0P(I - P) = (I - P)P = 0
Example 4.6.6b: Projection onto a Plane

Project onto the xyxy-plane in R3\mathbb{R}^3:

P=(100010000)P = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 0 \end{pmatrix}

Then IP=diag(0,0,1)I - P = \text{diag}(0, 0, 1) projects onto the zz-axis. Both are orthogonal projections.

Example 4.6.6c: Oblique Projection
P=(1200)P = \begin{pmatrix} 1 & 2 \\ 0 & 0 \end{pmatrix}

Verify P2=PP^2 = P. This projects onto span{(1,0)T}\text{span}\{(1,0)^T\} along span{(2,1)T}\text{span}\{(-2,1)^T\}. Note PPTP \neq P^T, so this is not an orthogonal projection.

Remark 4.6.3b: Applications of Projections

Idempotent matrices appear throughout mathematics:

  • Statistics: Hat matrix H=X(XTX)1XTH = X(X^T X)^{-1} X^T in linear regression
  • Quantum mechanics: Projection operators represent measurements
  • Signal processing: Filtering operations
  • Optimization: Projecting onto constraint sets

6. Nilpotent Matrices

Definition 4.6.7: Nilpotent Matrix

A matrix NN is nilpotent if Nk=0N^k = 0 for some k1k \geq 1. The smallest such kk is the index of nilpotency.

Theorem 4.6.9: Properties of Nilpotent Matrices
  1. All eigenvalues are zero
  2. det(N)=0\det(N) = 0 and tr(N)=0\text{tr}(N) = 0
  3. INI - N is invertible: (IN)1=I+N+N2++Nk1(I-N)^{-1} = I + N + N^2 + \cdots + N^{k-1}
  4. For ntimesnn \\times n matrix: index n\leq n
Example 4.6.7: Nilpotent Matrix
N=(010001000),N2=(001000000),N3=0N = \begin{pmatrix} 0 & 1 & 0 \\ 0 & 0 & 1 \\ 0 & 0 & 0 \end{pmatrix}, \quad N^2 = \begin{pmatrix} 0 & 0 & 1 \\ 0 & 0 & 0 \\ 0 & 0 & 0 \end{pmatrix}, \quad N^3 = 0

Index of nilpotency is 3.

Proof:

For property (1): If Nv=λvNv = \lambda v with v0v \neq 0, then:

0=Nkv=λkv    λk=0    λ=00 = N^k v = \lambda^k v \implies \lambda^k = 0 \implies \lambda = 0

For property (3), verify by expanding:

(IN)(I+N+N2++Nk1)=INk=I(I - N)(I + N + N^2 + \cdots + N^{k-1}) = I - N^k = I
Example 4.6.7b: 2×2 Nilpotent Matrices

All 2×22 \times 2 nilpotent matrices have the form:

N=(abca) where a2+bc=0N = \begin{pmatrix} a & b \\ c & -a \end{pmatrix} \text{ where } a^2 + bc = 0

Example: (1111)\begin{pmatrix} 1 & -1 \\ 1 & -1 \end{pmatrix} has N2=0N^2 = 0.

Example 4.6.7c: Computing (I-N)^{-1}

For N=(0100)N = \begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix} with N2=0N^2 = 0:

(IN)1=I+N=(1101)(I - N)^{-1} = I + N = \begin{pmatrix} 1 & 1 \\ 0 & 1 \end{pmatrix}

Verify: (IN)(I+N)=IN2=I(I - N)(I + N) = I - N^2 = I.

Remark 4.6.5b: Nilpotent and Jordan Form

Nilpotent matrices are key building blocks in Jordan canonical form. Every nilpotent matrix is similar to a direct sum of nilpotent Jordan blocks:

Jk(0)=(0100001000010000)k×kJ_k(0) = \begin{pmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & & \ddots & \ddots & \vdots \\ 0 & 0 & \cdots & 0 & 1 \\ 0 & 0 & \cdots & 0 & 0 \end{pmatrix}_{k \times k}
Theorem 4.6.9b: Sum of Nilpotent Matrices

If N1N_1 and N2N_2 are nilpotent and N1N2=N2N1N_1 N_2 = N_2 N_1 (they commute), then N1+N2N_1 + N_2 is nilpotent.

Proof:

If N1p=0N_1^p = 0 and N2q=0N_2^q = 0, use binomial theorem (valid since they commute):

(N1+N2)p+q1=k=0p+q1(p+q1k)N1kN2p+q1k(N_1 + N_2)^{p+q-1} = \sum_{k=0}^{p+q-1} \binom{p+q-1}{k} N_1^k N_2^{p+q-1-k}

Each term has either kpk \geq p (so N1k=0N_1^k = 0) or p+q1kqp+q-1-k \geq q (so N2p+q1k=0N_2^{p+q-1-k} = 0).

7. Involutions and Other Special Types

Definition 4.6.8: Involution

A matrix AA is an involution if A2=IA^2 = I.

Theorem 4.6.10: Properties of Involutions
  1. A1=AA^{-1} = A
  2. Eigenvalues are +1+1 or 1-1
  3. det(A)=±1\det(A) = \pm 1
Example 4.6.8: Involutions
(1001),(0110),I\begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix}, \quad \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}, \quad -I
Remark 4.6.4: Normal Matrices

A complex matrix is normal if AA=AAA^* A = A A^*. Examples include symmetric, skew-symmetric, orthogonal, Hermitian, and unitary matrices. Normal matrices are unitarily diagonalizable.

Proof:

For involution eigenvalues: If Av=λvAv = \lambda v, then v=A2v=λ2vv = A^2 v = \lambda^2 v, so λ2=1\lambda^2 = 1.

Definition 4.6.9: Unitary Matrix

A complex matrix UU is unitary if UU=UU=IU^* U = U U^* = I, where U=UTU^* = \overline{U}^T.

Unitary matrices are the complex analogue of orthogonal matrices.

Definition 4.6.10: Hermitian Matrix

A complex matrix HH is Hermitian if H=HH = H^*.

Hermitian matrices are the complex analogue of symmetric matrices. All eigenvalues are real.

Theorem 4.6.11: Properties of Hermitian Matrices
  1. All eigenvalues of a Hermitian matrix are real
  2. Eigenvectors for distinct eigenvalues are orthogonal
  3. Hermitian matrices are unitarily diagonalizable
  4. xHxRx^* H x \in \mathbb{R} for all xx (Rayleigh quotient is real)
Example 4.6.9: Hermitian Matrix
H=(21+i1i3)H = \begin{pmatrix} 2 & 1+i \\ 1-i & 3 \end{pmatrix}

Note H=HH = H^*. The eigenvalues are 5±52\frac{5 \pm \sqrt{5}}{2}, both real.

Remark 4.6.6: Positive Definite Matrices

A symmetric (or Hermitian) matrix AA is positive definite if xTAx>0x^T A x > 0 for all x0x \neq 0. Equivalently:

  • All eigenvalues are positive
  • All leading principal minors are positive (Sylvester criterion)
  • A=BTBA = B^T B for some invertible BB (Cholesky)
Example 4.6.10: Positive Definite Matrix
A=(2113)A = \begin{pmatrix} 2 & 1 \\ 1 & 3 \end{pmatrix}

Check: det(2)=2>0\det(2) = 2 > 0 and det(A)=5>0\det(A) = 5 > 0. Or eigenvalues 5±52\frac{5 \pm \sqrt{5}}{2} are both positive.

8. Common Mistakes

Assuming products of symmetric matrices are symmetric

If AA, BB symmetric, ABAB is symmetric only if they commute.

Confusing orthogonal with symmetric

Orthogonal: QTQ=IQ^T Q = I. Symmetric: A=ATA = A^T. Different conditions!

Thinking nilpotent means zero

Nilpotent means Nk=0N^k = 0, not N=0N = 0.

Forgetting skew-symmetric diagonals are zero

aii=aiia_{ii} = -a_{ii} forces aii=0a_{ii} = 0.

9. Historical Notes

Symmetric Matrices

The study of symmetric matrices goes back to Lagrange and Laplace in the 18th century. The spectral theorem for symmetric matrices was developed by Cauchy (1829) and later refined by Sylvester and others.

Orthogonal Matrices

Orthogonal transformations were studied by Euler in relation to rigid body motion. The term "orthogonal" comes from Greek "orthos" (right) and "gonia" (angle), reflecting the right-angle preservation property.

Jordan Normal Form

Camille Jordan (1870) developed the canonical form that bears his name, giving a complete classification of linear transformations via nilpotent matrices and diagonal blocks.

Hermitian Matrices

Named after Charles Hermite (1822-1901), who studied quadratic forms over the complex numbers. Hermitian matrices are fundamental in quantum mechanics, where observables are represented by Hermitian operators.

10. Key Takeaways

Diagonal & Triangular

Eigenvalues on diagonal. Easy to invert, multiply, and solve systems.

Symmetric

Real eigenvalues, orthogonal eigenvectors. Diagonalizable by orthogonal matrix.

Orthogonal

Preserve lengths and angles. Q1=QTQ^{-1} = Q^T. Rotations and reflections.

Idempotent & Nilpotent

Projections (P2=PP^2=P) and nilpotent (Nk=0N^k=0) key in Jordan form.

11. Study Tips

Memorize Key Properties

Create flashcards for: symmetric (A=ATA = A^T), orthogonal (QTQ=IQ^T Q = I), idempotent (P2=PP^2 = P), nilpotent (Nk=0N^k = 0).

Practice Verification

Given a matrix, practice quickly checking which special properties it has. Start with 2×2 matrices before moving to larger ones.

Connect to Geometry

Orthogonal = rotations/reflections. Projections = shadows. Visualizing helps remember properties.

Understand Eigenvalue Constraints

Many special matrices restrict possible eigenvalues: symmetric → real, orthogonal → |λ| = 1, idempotent → {0,1}, nilpotent → all 0.

12. What's Next?

In the next chapter on Determinants, you will see how special matrix structure simplifies determinant computation:

  • Determinants of triangular matrices are products of diagonal entries
  • Orthogonal matrices have det=±1\det = \pm 1
  • Block diagonal determinants factor as products
  • Nilpotent matrices always have det=0\det = 0

Later, in Eigenvalues, the spectral theorem will show that symmetric matrices are always diagonalizable with orthogonal eigenvectors—one of the most important results in linear algebra.

13. Quick Reference Table

TypeDefinitionKey PropertyEigenvalues
Diagonalaij=0a_{ij}=0 for iji \neq jPowers/inverse easyDiagonal entries
TriangularZeros above/below diagonalProducts stay triangularDiagonal entries
SymmetricA=ATA = A^TOrthogonal eigenvectorsAll real
Skew-symmetricA=ATA = -A^TZero diagonalPurely imaginary
OrthogonalQTQ=IQ^T Q = IPreserves lengthλ=1|\lambda| = 1
IdempotentP2=PP^2 = PProjection0 or 1
NilpotentNk=0N^k = 0INI-N invertibleAll 0
InvolutionA2=IA^2 = IA1=AA^{-1} = A±1

14. Worked Examples

Example 1: Classify a Matrix

Given A=(011102120)A = \begin{pmatrix} 0 & 1 & -1 \\ -1 & 0 & 2 \\ 1 & -2 & 0 \end{pmatrix}, classify this matrix.

Solution:

  • Check symmetric: AT=(011102120)=AA^T = \begin{pmatrix} 0 & -1 & 1 \\ 1 & 0 & -2 \\ -1 & 2 & 0 \end{pmatrix} = -A. So AA is skew-symmetric.
  • Since n=3n = 3 is odd, det(A)=0\det(A) = 0 (all skew-symmetric matrices of odd order are singular).
  • All diagonal entries are 0 (as expected for skew-symmetric).
  • Not orthogonal, idempotent, or nilpotent (can verify).

Example 2: Find a Projection Matrix

Find the orthogonal projection onto span{(1,2,2)T}\text{span}\{(1, 2, 2)^T\}.

Solution:

For a line spanned by unit vector uu, projection is P=uuTP = uu^T.

First normalize: v=1+4+4=3\|v\| = \sqrt{1+4+4} = 3, so u=13(1,2,2)Tu = \frac{1}{3}(1, 2, 2)^T.

P=uuT=19(122)(122)=19(122244244)P = uu^T = \frac{1}{9}\begin{pmatrix} 1 \\ 2 \\ 2 \end{pmatrix}\begin{pmatrix} 1 & 2 & 2 \end{pmatrix} = \frac{1}{9}\begin{pmatrix} 1 & 2 & 2 \\ 2 & 4 & 4 \\ 2 & 4 & 4 \end{pmatrix}

Verify: P2=PP^2 = P, P=PTP = P^T, tr(P)=1\text{tr}(P) = 1, eigenvalues are 0, 0, 1.

Example 3: Verify Orthogonality

Show that Q=13(122221212)Q = \frac{1}{3}\begin{pmatrix} 1 & -2 & 2 \\ 2 & 2 & 1 \\ 2 & -1 & -2 \end{pmatrix} is orthogonal.

Solution:

Check that columns are orthonormal:

  • Column 1: 19(1+4+4)=1\frac{1}{9}(1+4+4) = 1
  • Column 2: 19(4+4+1)=1\frac{1}{9}(4+4+1) = 1
  • Column 3: 19(4+1+4)=1\frac{1}{9}(4+1+4) = 1
  • Col 1 · Col 2: 19(2+42)=0\frac{1}{9}(-2+4-2) = 0
  • Col 1 · Col 3: 19(2+24)=0\frac{1}{9}(2+2-4) = 0
  • Col 2 · Col 3: 19(4+2+2)=0\frac{1}{9}(-4+2+2) = 0

Thus QTQ=IQ^T Q = I. Also det(Q)=127(1(3)(2)(6)+2(6))=1\det(Q) = \frac{1}{27}(1 \cdot (-3) - (-2) \cdot (-6) + 2 \cdot (-6)) = 1, so QQ is a rotation.

Example 4: Decompose into Symmetric + Skew

Decompose A=(123045006)A = \begin{pmatrix} 1 & 2 & 3 \\ 0 & 4 & 5 \\ 0 & 0 & 6 \end{pmatrix} into symmetric + skew-symmetric.

Solution:

S=12(A+AT)=12(2232853512)=(113/2145/23/25/26)S = \frac{1}{2}(A + A^T) = \frac{1}{2}\begin{pmatrix} 2 & 2 & 3 \\ 2 & 8 & 5 \\ 3 & 5 & 12 \end{pmatrix} = \begin{pmatrix} 1 & 1 & 3/2 \\ 1 & 4 & 5/2 \\ 3/2 & 5/2 & 6 \end{pmatrix}
K=12(AAT)=12(023205350)=(013/2105/23/25/20)K = \frac{1}{2}(A - A^T) = \frac{1}{2}\begin{pmatrix} 0 & 2 & 3 \\ -2 & 0 & 5 \\ -3 & -5 & 0 \end{pmatrix} = \begin{pmatrix} 0 & 1 & 3/2 \\ -1 & 0 & 5/2 \\ -3/2 & -5/2 & 0 \end{pmatrix}

Verify: S+K=AS + K = A, S=STS = S^T, K=KTK = -K^T.

Example 5: Nilpotent Matrix Computation

For N=(1111)N = \begin{pmatrix} 1 & 1 \\ -1 & -1 \end{pmatrix}, show NN is nilpotent and find (IN)1(I - N)^{-1}.

Solution:

N2=(1111)(1111)=(0000)N^2 = \begin{pmatrix} 1 & 1 \\ -1 & -1 \end{pmatrix}\begin{pmatrix} 1 & 1 \\ -1 & -1 \end{pmatrix} = \begin{pmatrix} 0 & 0 \\ 0 & 0 \end{pmatrix}

So NN is nilpotent with index 2. Thus:

(IN)1=I+N=(2110)(I - N)^{-1} = I + N = \begin{pmatrix} 2 & 1 \\ -1 & 0 \end{pmatrix}

Verify: (IN)(I+N)=IN2=I(I-N)(I+N) = I - N^2 = I.

15. Applications Overview

Computer Graphics

Orthogonal matrices represent rotations and reflections in 3D graphics. Transformation matrices preserve shape when rendering objects.

Quantum Mechanics

Hermitian operators represent observables with real eigenvalues. Unitary operators represent time evolution preserving probability.

Statistics

Projection matrices compute least squares fits. Covariance matrices are symmetric positive semi-definite.

Numerical Analysis

LU decomposition uses triangular matrices for efficient solving. QR decomposition uses orthogonal matrices for stability.

Control Theory

Nilpotent matrices describe systems that "stop" after finite time. Stability analysis uses eigenvalue locations.

Machine Learning

Symmetric matrices appear in kernel methods. PCA uses eigenvectors of covariance (symmetric) matrices.

16. Relationships Between Special Matrix Types

Venn Diagram of Special Matrices

Understanding how special matrix types relate to each other:

  • Diagonal ⊂ Triangular: Every diagonal matrix is both upper and lower triangular.
  • Diagonal ⊂ Symmetric: Every diagonal matrix is symmetric.
  • Orthogonal ∩ Symmetric = Involutions: QT=Q1Q^T = Q^{-1} and QT=QQ^T = Q implies Q2=IQ^2 = I.
  • Symmetric ∩ Skew-symmetric = {0}: Only the zero matrix is both.
  • Nilpotent ⊂ Singular: All nilpotent matrices have det = 0.
  • Idempotent ≠ Nilpotent: Except P=0P = 0, these are distinct (eigenvalues 0,1 vs all 0).
Theorem 4.6.12: Orthogonal + Symmetric = Involution

A matrix is both orthogonal and symmetric if and only if it is an involution with eigenvalues ±1\pm 1.

Proof:

If QQ is orthogonal (QTQ=IQ^T Q = I) and symmetric (Q=QTQ = Q^T), then Q2=QQT=IQ^2 = Q \cdot Q^T = I.

Conversely, if Q2=IQ^2 = I and Q=QTQ = Q^T, then QTQ=QQ=IQ^T Q = Q \cdot Q = I, so QQ is orthogonal.

Remark 4.6.7: Decomposition Theorems

Several important decomposition theorems relate special matrix types:

  • Schur: Every matrix is unitarily similar to upper triangular.
  • Spectral: Normal matrices are unitarily diagonalizable.
  • Polar: A=QPA = QP with QQ orthogonal, PP symmetric positive semi-definite.
  • SVD: A=UΣVTA = U \Sigma V^T with U,VU, V orthogonal, Σ\Sigma diagonal.

17. Computational Considerations

Complexity Benefits

Matrix TypeOperationGeneralSpecial
DiagonalInverseO(n3)O(n^3)O(n)O(n)
TriangularSolve Ax=bAx=bO(n3)O(n^3)O(n2)O(n^2)
SymmetricEigenvaluesO(n3)O(n^3)O(n3)O(n^3) (faster const)
OrthogonalInverseO(n3)O(n^3)O(n2)O(n^2) (transpose)
SparseMultiplyO(n2)O(n^2)O(nnz)O(nnz)
Remark 4.6.8: Numerical Stability

Special structure often improves numerical stability:

  • Symmetric positive definite: Cholesky is stable without pivoting.
  • Orthogonal: Condition number is 1 (perfectly conditioned).
  • Triangular: Back-substitution is stable for well-conditioned systems.
  • Diagonal: Operations are exact (no accumulation of errors).
Example 4.6.11: Exploiting Structure

To solve Ax=bAx = b where A=LDLTA = LDL^T (symmetric factorization):

  1. Solve Ly=bLy = b (lower triangular) - O(n2)O(n^2)
  2. Solve Dz=yDz = y (diagonal) - O(n)O(n)
  3. Solve LTx=zL^T x = z (upper triangular) - O(n2)O(n^2)

Total: O(n2)O(n^2) after factorization, vs O(n3)O(n^3) for general Gaussian elimination.

Remark 4.6.9: Storage Optimization

Special matrices can be stored more efficiently:

  • Diagonal: Store only nn entries instead of n2n^2.
  • Triangular: Store n(n+1)/2n(n+1)/2 entries.
  • Symmetric: Store only upper (or lower) triangle - n(n+1)/2n(n+1)/2 entries.
  • Banded: For bandwidth kk, store O(nk)O(nk) entries.
  • Sparse: Use compressed formats (CSR, CSC) storing only nonzeros.
Special Matrices Practice
18
Questions
0
Correct
0%
Accuracy
1
A matrix AA is symmetric if and only if:
Easy
Not attempted
2
If AA is orthogonal, then A1=?A^{-1} = ?
Easy
Not attempted
3
The product of two upper triangular matrices is:
Medium
Not attempted
4
If AA is skew-symmetric, all diagonal entries are:
Medium
Not attempted
5
A matrix PP is idempotent if:
Easy
Not attempted
6
If NN is nilpotent with N3=0N^3 = 0 but N20N^2 \neq 0, the index is:
Medium
Not attempted
7
The determinant of an orthogonal matrix is:
Medium
Not attempted
8
Every matrix can be written as A=S+KA = S + K where:
Medium
Not attempted
9
All eigenvalues of a nilpotent matrix are:
Hard
Not attempted
10
The eigenvalues of an idempotent matrix are:
Hard
Not attempted
11
An upper triangular matrix is invertible iff:
Medium
Not attempted
12
If AA is symmetric and BB is any matrix, then BTABB^T A B is:
Medium
Not attempted
13
The columns of an orthogonal matrix form:
Medium
Not attempted
14
If AA is both symmetric and skew-symmetric, then:
Easy
Not attempted
15
Rotation matrices in R2\mathbb{R}^2 are:
Easy
Not attempted
16
A Householder matrix H=I2uuTH = I - 2uu^T with u=1\|u\|=1 is:
Hard
Not attempted
17
If AA is skew-symmetric, then eAe^A is:
Hard
Not attempted
18
The rank of an n×nn \times n idempotent matrix equals its:
Hard
Not attempted

Frequently Asked Questions

Why are special matrices important?

Special matrices have structure that simplifies computations and reveals properties. Diagonal matrices are easy to invert and power, triangular matrices make solving systems efficient, and symmetric matrices have real eigenvalues with orthogonal eigenvectors.

What's the geometric meaning of an orthogonal matrix?

Orthogonal matrices represent isometries—transformations that preserve lengths and angles. In 2D/3D, these are rotations (det = 1) and reflections (det = -1). They preserve the dot product: ⟨Ax, Ay⟩ = ⟨x, y⟩.

How do I check if a matrix is symmetric?

Check if A = A^T, i.e., whether a_{ij} = a_{ji} for all i, j. Visually, check if the matrix equals its reflection across the main diagonal.

What are idempotent matrices used for?

Idempotent matrices represent projections. If P² = P, then P projects vectors onto its image. Applying the projection twice gives the same result—hence P² = P. They're central in statistics and quantum mechanics.

What is a nilpotent matrix?

A matrix N is nilpotent if N^k = 0 for some positive integer k. The smallest such k is the index of nilpotency. Nilpotent matrices have all eigenvalues equal to zero and appear in Jordan forms.

Why do triangular matrices have eigenvalues on the diagonal?

For triangular A, det(A - λI) is the product of (a_{ii} - λ) terms. This equals zero exactly when λ equals some diagonal entry.

What's the symmetric-skew decomposition?

Every matrix A can be uniquely written as A = S + K where S = (A + A^T)/2 is symmetric and K = (A - A^T)/2 is skew-symmetric. This is analogous to decomposing a function into even and odd parts.

Are products of symmetric matrices symmetric?

Generally no! If A and B are symmetric, (AB)^T = B^T A^T = BA, which equals AB only if A and B commute. However, B^T A B is always symmetric when A is symmetric.

What matrices are both orthogonal and symmetric?

If A is orthogonal and symmetric, then A² = A·A^T = I. So A is an involution. The only such matrices have eigenvalues ±1: they're reflections across subspaces.

How do special matrices simplify eigenvalue problems?

Diagonal: eigenvalues are diagonal entries. Triangular: same. Symmetric: always diagonalizable with orthogonal eigenvectors. Orthogonal: eigenvalues have |λ| = 1. This structure makes analysis much easier.

What is the difference between orthogonal and unitary matrices?

Orthogonal matrices are for real vector spaces: Q^T Q = I. Unitary matrices are for complex spaces: U* U = I where U* is conjugate transpose. Both preserve inner products in their respective spaces.

How do I construct a projection matrix onto a subspace?

If V has orthonormal basis {v₁,...,vₖ}, then P = Σᵢ vᵢvᵢᵀ = VVᵀ where V = [v₁|...|vₖ]. If the basis isn't orthonormal, use P = V(VᵀV)⁻¹Vᵀ.

What is the Cholesky decomposition?

Every positive definite matrix A can be written as A = LLᵀ where L is lower triangular with positive diagonal. This is more efficient than LU for symmetric positive definite systems.

When is a matrix diagonalizable?

A matrix is diagonalizable iff it has n linearly independent eigenvectors. Sufficient conditions: distinct eigenvalues, or being normal (including symmetric, orthogonal, Hermitian, unitary).

What is a circulant matrix?

Each row is a cyclic shift of the row above. Circulants are diagonalized by the DFT matrix and appear in signal processing. They commute with each other.