Learn how to recover missing elements in matrices using low-rank assumptions. Master the conversion from rank minimization to nuclear norm minimization, understand semidefinite programming solutions, and discover recovery conditions for perfect matrix reconstruction.
Matrix completion is the problem of recovering a matrix from a subset of its entries. Given partial observations, we want to infer the missing values under the assumption that the matrix is low-rank.
Given a matrix with observed entries at positions , find the complete matrix such that:
and has low rank (or is as low-rank as possible).
The ideal approach is to minimize the rank of the completed matrix:
where is the rank of matrix .
The nuclear norm is the tightest convex relaxation of the rank function, analogous to how L1 norm relaxes L0 norm.
The nuclear norm (trace norm) of matrix is the sum of its singular values:
where are the singular values of .
Replace rank with nuclear norm:
This is a convex optimization problem that can be solved efficiently!
Nuclear norm minimization can be reformulated as a semidefinite program:
Nuclear norm minimization is equivalent to:
subject to:
where means positive semidefinite, and are auxiliary variables.
Under what conditions can we perfectly recover a low-rank matrix from partial observations?
If matrix has rank and observations are sampled uniformly at random, then with high probability, nuclear norm minimization perfectly recovers if:
where is the number of observed entries, and is a constant (typically ).
Matrix completion is the foundation of collaborative filtering in recommendation systems.
User-item rating matrix: 10,000 users × 5,000 movies. Only 5% of ratings are observed (sparse). Matrix is low-rank (users have similar preferences, movies have similar features).
Apply nuclear norm minimization:
Enables accurate movie recommendations for users who haven't rated many movies. Powers Netflix, Amazon, and other recommendation systems.
Matrix completion recovers missing matrix entries under the low-rank assumption, enabling applications in recommendation systems and data recovery.
Rank minimization is NP-hard, but nuclear norm (sum of singular values) provides the tightest convex relaxation.
Nuclear norm minimization can be solved via semidefinite programming (SDP), providing efficient polynomial-time algorithms.
Perfect recovery is possible with observations, much less than the full entries.
Matrix completion is fundamental to collaborative filtering, enabling accurate recommendations from sparse user-item interactions.
The nuclear norm approach extends the L1 sparsity concept from vectors to matrices, completing the sparse learning framework.