Orthogonal Projections

Definition

Let be an -dimensional vector space and a -dimensional subspace with orthonormal basis . The orthogonal projection of onto is the unique vector in closest to .

Intuition

Orthogonal projection drops a perpendicular from onto the subspace : the foot of that perpendicular is the projection. The residual is orthogonal to every vector in , which is exactly what “closest point” means. When , projecting finds the nearest achievable output to , which is the least-squares solution.

Formal Description

Any can be written using a full orthonormal basis as

The orthogonal projection onto retains only the -components:

where the coefficients are computed via inner products: . Therefore:

Minimality. is the closest point in to . For any with expansion :

Projection matrix. When and the columns of are linearly independent, the projection of onto is given by

The matrix is idempotent: .

Applications

  • Least-squares regression projects the target onto the column space of the design matrix.
  • Gram-Schmidt uses successive projections to build an orthonormal basis.
  • Signal processing: projecting a signal onto a subspace removes noise components outside the subspace.

Trade-offs

The formula requires to be invertible (full column rank). If is rank-deficient, use the pseudoinverse instead. Explicitly forming is and memory-intensive; in practice, solve the normal equations directly rather than materialising .