Orthogonal Projections
Definition
Let be an -dimensional vector space and a -dimensional subspace with orthonormal basis . The orthogonal projection of onto is the unique vector in closest to .
Intuition
Orthogonal projection drops a perpendicular from onto the subspace : the foot of that perpendicular is the projection. The residual is orthogonal to every vector in , which is exactly what “closest point” means. When , projecting finds the nearest achievable output to , which is the least-squares solution.
Formal Description
Any can be written using a full orthonormal basis as
The orthogonal projection onto retains only the -components:
where the coefficients are computed via inner products: . Therefore:
Minimality. is the closest point in to . For any with expansion :
Projection matrix. When and the columns of are linearly independent, the projection of onto is given by
The matrix is idempotent: .
Applications
- Least-squares regression projects the target onto the column space of the design matrix.
- Gram-Schmidt uses successive projections to build an orthonormal basis.
- Signal processing: projecting a signal onto a subspace removes noise components outside the subspace.
Trade-offs
The formula requires to be invertible (full column rank). If is rank-deficient, use the pseudoinverse instead. Explicitly forming is and memory-intensive; in practice, solve the normal equations directly rather than materialising .