Matrix Diagonalization
Definition
A square matrix is diagonalisable if it can be written as
where is an invertible matrix whose columns are eigenvectors of , and is the diagonal matrix of corresponding eigenvalues. Equivalently, .
Intuition
Diagonalisation decomposes a transformation into independent “modes”: change to the eigenvector basis (via ), apply a pure scaling along each axis (via ), then return to the original basis (via ). In the eigenvector basis the transformation is trivially simple — each coordinate is just scaled by its eigenvalue with no mixing between dimensions. This separation of independent modes makes repeated application, inversion, and analysis dramatically easier.
Formal Description
Derivation. Collecting the eigenrelations into a single matrix equation gives
where and . If is invertible (i.e.\ the eigenvectors are linearly independent), both sides can be multiplied by to obtain .
Condition for diagonalisability. An matrix is diagonalisable if and only if it has linearly independent eigenvectors. In particular, a matrix with distinct eigenvalues is always diagonalisable.
Eigendecomposition. More generally, the eigendecomposition of is
where the columns of are eigenvectors and is the vector of eigenvalues. Not every matrix admits such a decomposition over , but every real symmetric matrix can be decomposed as
where is orthogonal () and its columns are eigenvectors of . By convention eigenvalues in are sorted in descending order; the decomposition is unique (up to sign of eigenvectors) when all eigenvalues are distinct.
Matrix powers. Diagonalisation makes computing powers straightforward. For ,
where . Each repeated application of simply raises the diagonal entries of to the next power.
Properties from eigenvalues.
- is singular if and only if at least one eigenvalue is zero.
- is positive definite if all eigenvalues are positive; positive semidefinite if all eigenvalues are non-negative. Positive semidefinite matrices satisfy for all ; positive definite matrices additionally have .
- The eigendecomposition of a real symmetric matrix can be used to optimise quadratic forms subject to .
Applications
- Computing matrix powers and matrix exponentials efficiently (e.g., solving linear ODEs).
- PCA and spectral methods in machine learning rely on eigendecomposition of covariance or kernel matrices.
- Decoupling coupled differential equations into independent scalar equations.
- Checking positive definiteness of matrices (e.g., for valid covariance matrices or convex quadratic forms).
Trade-offs
- Not all matrices are diagonalisable: a defective matrix has a repeated eigenvalue with fewer independent eigenvectors than its algebraic multiplicity. Jordan normal form is the generalisation, but it is more complex to work with.
- Even when diagonalisable, computing the full eigendecomposition costs ; for very large sparse matrices only a few dominant eigenvalues are typically computed via iterative methods (e.g., Lanczos, Arnoldi).
- Symmetric eigendecomposition () is numerically better conditioned and always exists over ; the general (non-symmetric) case can produce complex eigenvalues and an ill-conditioned .
Links
- Eigenvalues and Eigenvectors — eigenvectors and eigenvalues are the building blocks of diagonalisation
- Special Matrices — identity matrix used in the eigenvalue problem; orthogonal matrices arise in symmetric eigendecomposition