Fiveable

🧚🏽‍♀️Abstract Linear Algebra I Unit 8 Review

QR code for Abstract Linear Algebra I practice questions

8.3 Orthogonal Projections and Complementary Subspaces

8.3 Orthogonal Projections and Complementary Subspaces

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🧚🏽‍♀️Abstract Linear Algebra I
Unit & Topic Study Guides

Orthogonal projections and complementary subspaces are key concepts in inner product spaces. They allow us to break down vectors into components that are perpendicular to each other, making it easier to analyze and manipulate them.

These ideas are crucial for understanding how vectors interact in higher dimensions. By projecting vectors onto subspaces and finding their orthogonal complements, we can solve complex problems in linear algebra and its applications.

Orthogonal Projections

Definition and Properties

  • An orthogonal projection is a linear transformation PP from a vector space VV to itself such that P2=PP^2 = P and the range of PP is orthogonal to the null space of PP
  • The matrix representation of an orthogonal projection is symmetric (PT=PP^T = P) and idempotent (P2=PP^2 = P)
  • Orthogonal projections are characterized by the property that they minimize the distance between a vector and its projection onto a subspace
    • For any vector vVv \in V and subspace WVW \subseteq V, the orthogonal projection PW(v)P_W(v) is the closest point in WW to vv
    • Mathematically, vPW(v)vw\|v - P_W(v)\| \leq \|v - w\| for all wWw \in W
  • The eigenvalues of an orthogonal projection matrix are either 0 or 1
    • Eigenvectors corresponding to eigenvalue 1 span the range of the projection
    • Eigenvectors corresponding to eigenvalue 0 span the null space of the projection
  • Orthogonal projections preserve the length of vectors in the subspace they project onto
    • For any vector vWv \in W, PW(v)=v\|P_W(v)\| = \|v\|
    • This follows from the fact that PWP_W is an identity transformation on WW

Computing Orthogonal Projections

  • Given a basis for a subspace WW of a vector space VV, the orthogonal projection of a vector vv onto WW can be computed using the Gram-Schmidt process to find an orthonormal basis for WW
    • The Gram-Schmidt process takes a basis {w1,,wk}\{w_1, \ldots, w_k\} for WW and produces an orthonormal basis {u1,,uk}\{u_1, \ldots, u_k\}
    • The orthogonal projection of vv onto WW is then given by PW(v)=i=1kv,uiuiP_W(v) = \sum_{i=1}^k \langle v, u_i \rangle u_i
  • The orthogonal projection matrix onto a subspace WW is given by P=A(ATA)1ATP = A(A^T A)^{-1} A^T, where AA is a matrix whose columns form a basis for WW
    • This formula can be derived using the properties of inner products and linear algebra
    • If the columns of AA are orthonormal, then ATA=IA^T A = I and the formula simplifies to P=AATP = AA^T
  • The orthogonal projection of a vector vv onto a subspace WW can be computed as the sum of the inner products of vv with each basis vector of WW, multiplied by the corresponding basis vector
    • If {w1,,wk}\{w_1, \ldots, w_k\} is a basis for WW, then PW(v)=i=1kv,wiwiP_W(v) = \sum_{i=1}^k \langle v, w_i \rangle w_i
    • This formula follows from the linearity of inner products and the properties of bases
  • In Rn\mathbb{R}^n, the orthogonal projection of a vector vv onto a subspace WW can be found by solving the linear system Ax=vAx = v for xx, where AA is a matrix whose columns form an orthonormal basis for WW
    • The solution xx gives the coordinates of PW(v)P_W(v) with respect to the orthonormal basis
    • This approach is computationally efficient and numerically stable

Orthogonal Complements

Definition and Properties, linear algebra - Orthogonal Projection Problem - Mathematics Stack Exchange

Definition and Properties

  • The orthogonal complement of a subspace WW in a vector space VV is the set of all vectors in VV that are orthogonal to every vector in WW
    • Mathematically, W={vV:v,w=0 for all wW}W^{\perp} = \{v \in V : \langle v, w \rangle = 0 \text{ for all } w \in W\}
    • Orthogonality is defined in terms of the inner product on VV
  • The orthogonal complement of WW is denoted by WW^{\perp} and is itself a subspace of VV
    • This follows from the linearity of inner products and the subspace properties
    • If v1,v2Wv_1, v_2 \in W^{\perp} and cRc \in \mathbb{R}, then cv1+v2,w=cv1,w+v2,w=0\langle cv_1 + v_2, w \rangle = c\langle v_1, w \rangle + \langle v_2, w \rangle = 0 for all wWw \in W
  • For any subspace WW of a finite-dimensional vector space VV, the dimension of WW^{\perp} is equal to the codimension of WW, i.e., dim(W)=dim(V)dim(W)\dim(W^{\perp}) = \dim(V) - \dim(W)
    • This result follows from the rank-nullity theorem applied to the orthogonal projection map onto WW
    • The range of the projection has dimension dim(W)\dim(W), while the null space has dimension dim(V)dim(W)\dim(V) - \dim(W)
  • The orthogonal complement of the orthogonal complement of a subspace WW is WW itself, i.e., (W)=W(W^{\perp})^{\perp} = W
    • This "double complement" property follows from the definition of orthogonal complements and linear algebra
    • It shows that the orthogonal complement operation is an involution on subspaces
  • The existence and uniqueness of orthogonal complements can be proved using the rank-nullity theorem and the properties of inner products
    • For any subspace WVW \subseteq V, the orthogonal projection map onto WW is a surjective linear transformation
    • By the rank-nullity theorem, the null space of this map (which is WW^{\perp}) has dimension dim(V)dim(W)\dim(V) - \dim(W)
    • The uniqueness of WW^{\perp} follows from the fact that orthogonality is a well-defined relation on vectors

Relationship to Orthogonal Projections

  • The orthogonal complement of a subspace WW is closely related to the orthogonal projection onto WW
    • The null space of the orthogonal projection map PWP_W is precisely WW^{\perp}
    • This means that a vector vVv \in V is in WW^{\perp} if and only if PW(v)=0P_W(v) = 0
  • The orthogonal projection onto WW^{\perp} is complementary to the orthogonal projection onto WW, i.e., PW=IPWP_{W^{\perp}} = I - P_W
    • This follows from the properties of orthogonal projections and the definition of orthogonal complements
    • For any vector vVv \in V, we have v=PW(v)+PW(v)v = P_W(v) + P_{W^{\perp}}(v), where PW(v)WP_W(v) \in W and PW(v)WP_{W^{\perp}}(v) \in W^{\perp}
  • The matrix representation of the orthogonal projection onto WW^{\perp} can be computed from the matrix representation of the orthogonal projection onto WW
    • If PP is the matrix representing PWP_W, then IPI - P is the matrix representing PWP_{W^{\perp}}
    • This follows from the complementary property of orthogonal projections and matrix algebra
  • Orthogonal complements and orthogonal projections are fundamental tools in the study of inner product spaces and their applications
    • They allow us to decompose vectors and subspaces into orthogonal components
    • They provide a way to find the best approximation of a vector in a given subspace
    • They are used in various fields such as quantum mechanics, signal processing, and data analysis

Orthogonal Subspace Decompositions

Definition and Properties, Orthogonality (mathematics) - Wikipedia

Decomposing Vectors

  • Any vector vv in a vector space VV can be uniquely decomposed into a sum of two orthogonal vectors: one in a subspace WW and the other in its orthogonal complement WW^{\perp}
    • Mathematically, v=PW(v)+PW(v)v = P_W(v) + P_{W^{\perp}}(v), where PW(v)WP_W(v) \in W and PW(v)WP_{W^{\perp}}(v) \in W^{\perp}
    • This decomposition follows from the properties of orthogonal projections and complements
  • The orthogonal decomposition of a vector vv with respect to a subspace WW is given by v=PW(v)+(IPW)(v)v = P_W(v) + (I - P_W)(v), where PWP_W is the orthogonal projection matrix onto WW and II is the identity matrix
    • The first term PW(v)P_W(v) is the orthogonal projection of vv onto WW, representing the component of vv in WW
    • The second term (IPW)(v)(I - P_W)(v) is the orthogonal projection of vv onto WW^{\perp}, representing the component of vv orthogonal to WW
  • The orthogonal decomposition of a vector is unique and independent of the choice of basis for WW and WW^{\perp}
    • This follows from the uniqueness of orthogonal projections and complements
    • Different bases may lead to different expressions for the components, but the resulting decomposition is the same
  • Orthogonal decompositions are useful in solving various problems in linear algebra and its applications
    • They allow us to separate a vector into its relevant and irrelevant components with respect to a given subspace
    • They provide a way to find the closest vector in a subspace to a given vector, which is important in least-squares approximations and regression analysis
    • They are used in quantum mechanics to decompose state vectors into orthogonal subspaces corresponding to different observables or symmetries

Decomposing Vector Spaces

  • The orthogonal decomposition theorem states that a finite-dimensional vector space VV is the direct sum of a subspace WW and its orthogonal complement WW^{\perp}, i.e., V=WWV = W \oplus W^{\perp}
    • This means that every vector vVv \in V can be uniquely written as v=w+uv = w + u, where wWw \in W and uWu \in W^{\perp}
    • The direct sum notation \oplus emphasizes that the decomposition is unique and that WW={0}W \cap W^{\perp} = \{0\}
  • The orthogonal decomposition theorem is a consequence of the properties of orthogonal projections and complements
    • The orthogonal projection onto WW maps VV onto WW, while the orthogonal projection onto WW^{\perp} maps VV onto WW^{\perp}
    • The sum of these two projections is the identity map on VV, which implies that V=W+WV = W + W^{\perp}
    • The uniqueness of the decomposition follows from the fact that WW={0}W \cap W^{\perp} = \{0\}, which is a consequence of the definition of orthogonal complements
  • The orthogonal decomposition theorem can be generalized to more than two subspaces
    • If W1,,WkW_1, \ldots, W_k are mutually orthogonal subspaces of VV (i.e., WiWjW_i \perp W_j for all iji \neq j), then V=W1Wk(W1++Wk)V = W_1 \oplus \cdots \oplus W_k \oplus (W_1 + \cdots + W_k)^{\perp}
    • This decomposition is called an orthogonal direct sum and is unique
    • It allows us to decompose a vector space into orthogonal components corresponding to different properties or behaviors
  • Orthogonal decompositions of vector spaces have numerous applications in mathematics, physics, and engineering
    • In quantum mechanics, the state space of a system is decomposed into orthogonal subspaces corresponding to different eigenvalues of an observable
    • In signal processing, a signal is decomposed into orthogonal components corresponding to different frequencies or time scales (e.g., Fourier or wavelet decompositions)
    • In data analysis, a dataset is decomposed into orthogonal components corresponding to different sources of variation or latent factors (e.g., principal component analysis)

Matrix Decompositions

  • The orthogonal decomposition of a matrix AA can be used to find its best low-rank approximation, which has applications in data compression, signal processing, and machine learning
    • The best rank-kk approximation of AA is given by Ak=UkΣkVkTA_k = U_k \Sigma_k V_k^T, where UkU_k and VkV_k contain the first kk columns of UU and VV, respectively, and Σk\Sigma_k contains the first kk singular values of AA
    • This approximation is optimal in the sense that it minimizes the Frobenius norm of the difference between AA and any rank-kk matrix
    • The matrices UkU_k and VkV_k can be interpreted as the principal components of the row and column spaces of AA, respectively, while the singular values in Σk\Sigma_k represent the importance of each component
  • The orthogonal decomposition of a symmetric matrix AA is given by its eigendecomposition A=QΛQTA = Q \Lambda Q^T, where QQ is an orthogonal matrix and Λ\Lambda is a diagonal matrix containing the eigenvalues of AA
    • The columns of QQ are eigenvectors of AA and form an orthonormal basis for the underlying vector space
    • The eigenvalues in Λ\Lambda represent the variances of the data along each eigenvector direction
    • This decomposition is used in principal component analysis, spectral clustering, and other dimensionality reduction techniques
  • The orthogonal decomposition of a matrix can also be used to solve linear systems and least-squares problems
    • If A=QRA = QR is the QR decomposition of AA, where QQ is orthogonal and RR is upper triangular, then the linear system Ax=bAx = b can be solved by first solving Ry=QTbRy = Q^T b for yy and then setting x=Qyx = Qy
    • If A=UΣVTA = U \Sigma V^T is the singular value decomposition of AA, then the least-squares solution to Ax=bAx = b is given by x=VΣ+UTbx = V \Sigma^+ U^T b, where Σ+\Sigma^+ is the pseudoinverse of Σ\Sigma
    • These methods are numerically stable and efficient, especially when AA is ill-conditioned or has a high condition number
  • Matrix decompositions based on orthogonal subspaces are a fundamental tool in numerical linear algebra and have numerous applications in science and engineering
    • They provide a way to reveal the underlying structure and properties of a matrix, such as its rank, range, null space, and singular values
    • They allow us to compress, denoise, or regularize large datasets by focusing on the most important or informative components
    • They enable us to solve various optimization and approximation problems by reducing them to simpler subproblems in orthogonal subspaces