Eigenvalue Problems
Compute eigenvalues and eigenvectors numerically. From the simple power method to sophisticated algorithms, learn to solve the fundamental problem .
- Implement the power method for dominant eigenvalues
- Use inverse power method with shifts for specific eigenvalues
- Apply Gershgorin circles for eigenvalue localization
- Understand Jacobi rotations for symmetric matrices
- Learn the principles of the QR algorithm
- Apply deflation techniques for multiple eigenvalues
1. Power Method
The power method finds the dominant eigenvalue (largest in absolute value) and its eigenvector.
Algorithm Power Method
Input: Matrix , initial vector , tolerance
- Normalize:
- For :
- Compute
- Normalize
- Estimate eigenvalue: (Rayleigh quotient)
- Stop if
Theorem 8.1: Convergence of Power Method
If has a unique dominant eigenvalue with, then:
Convergence rate:
Example: Power Method
For with :
| 1 | (0.894, 0.447) | 2.8 |
| 2 | (0.745, 0.667) | 2.96 |
| 5 | (0.707, 0.707) | 3.00 |
Converges to , .
Note:
Special cases: If the dominant eigenvalue has multiplicity > 1, or if there are complex conjugate dominant eigenvalues, modified approaches are needed.
2. Inverse Power Method
Definition 8.1: Inverse Power Method
Apply power method to : since , the smallest eigenvalue of becomes dominant for .
Each iteration solves instead of computing explicitly.
Definition 8.2: Shifted Inverse Power Method
To find eigenvalue closest to shift , apply inverse power to :
The eigenvalue of closest to becomes dominant.
Remark:
Rayleigh quotient iteration: Use the current eigenvalue estimate as the shift:. This achieves cubic convergence near the eigenvalue.
3. Eigenvalue Localization
Theorem 8.2: Gershgorin Circle Theorem
Every eigenvalue of lies in at least one Gershgorin disk:
Furthermore, if disks form a connected region disjoint from other disks, exactly eigenvalues lie in that region.
Example: Gershgorin Disks
For :
- : center 4, radius 1 →
- : center 5, radius 2 →
- : center 2, radius 0 → (isolated)
Thus exactly, and other eigenvalues lie in .
Remark:
Gershgorin circles provide initial estimates for shifts in shifted inverse iteration, making convergence faster.
4. Jacobi Method for Symmetric Matrices
Definition 8.3: Jacobi Rotation
A Jacobi rotation (Givens rotation) is an orthogonal matrix that differs from identity only in positions :
Algorithm Jacobi Method
- Find largest off-diagonal element
- Compute rotation angle to zero out
- Apply rotation:
- Repeat until off-diagonal elements are negligible
Final diagonal contains eigenvalues; product of rotations gives eigenvectors.
Theorem 8.3: Jacobi Convergence
For symmetric matrices, the Jacobi method converges. The sum of squares of off-diagonal elements decreases at each step. With proper ordering, convergence is quadratic.
Note:
The Jacobi method is simple and robust, but requires operations per sweep and many sweeps. Modern software uses the more efficient QR algorithm.
5. Deflation and the QR Algorithm
Definition 8.4: Deflation
After finding eigenvalue with eigenvector , modify to remove this eigenvalue:
Then apply power method to to find remaining eigenvalues.
Definition 8.5: QR Algorithm
The basic QR iteration:
- Start with
- For :
- QR factorization:
- Reverse multiply:
For real matrices, converges to upper quasi-triangular form (real Schur form).
Theorem 8.4: QR Convergence
If eigenvalues have distinct absolute values, QR iteration converges to upper triangular form, with eigenvalues on the diagonal.
Convergence rate for (below diagonal):
Remark:
Practical QR: Uses shifts (Wilkinson shift), Hessenberg reduction, and implicit techniques for complexity. This is the standard algorithm in LAPACK and similar libraries.
Practice Quiz
Frequently Asked Questions
When should I use power method vs. QR algorithm?
Power method: When you only need the dominant eigenvalue (and eigenvector), especially for large sparse matrices where matrix-vector products are cheap.
QR algorithm: When you need all eigenvalues of a moderate-size dense matrix. It's the standard for complete eigendecomposition.
How do I find eigenvalues of a non-symmetric matrix?
Non-symmetric matrices may have complex eigenvalues. The QR algorithm (with Hessenberg reduction) handles this, producing a quasi-triangular (real Schur) form with 1×1 and 2×2 diagonal blocks corresponding to real and complex conjugate eigenvalue pairs.
What if the power method doesn't converge?
Check for these issues:
- Equal magnitude eigenvalues: Use techniques for repeated or complex dominant eigenvalues
- Poor initial guess: Try a random starting vector
- Slow convergence: Use Rayleigh quotient iteration or shifted inverse power
How accurate are computed eigenvalues?
For well-conditioned problems, eigenvalues are computed to near machine precision. For ill-conditioned problems (nearly repeated eigenvalues, non-normal matrices), accuracy may be limited. Eigenvectors of close eigenvalues may be poorly determined.
What is the spectral decomposition?
For a symmetric matrix: where is orthogonal (columns are eigenvectors) and is diagonal (eigenvalues).
This is useful for computing functions of matrices (e.g., ) and understanding matrix behavior.