Fiveable

Abstract Linear Algebra II Unit 4 Review

QR code for Abstract Linear Algebra II practice questions

4.5 Orthogonal complements and projections

4.5 Orthogonal complements and projections

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
Abstract Linear Algebra II
Unit & Topic Study Guides

Inner product spaces let us measure angles and distances between vectors. This section dives into orthogonal complements - subspaces containing vectors perpendicular to a given subspace. We'll see how these complements relate to the original space's structure.

We'll also explore orthogonal projections, which find the closest vector in a subspace to a given vector. This concept is crucial for solving least-squares problems and decomposing vectors into orthogonal components.

Orthogonal Complements and Properties

Definition and Basic Properties

  • Orthogonal complement of subspace W in inner product space V denoted W⊥ contains all vectors in V orthogonal to every vector in W
  • For finite-dimensional inner product space V and subspace W, (W⊥)⊥ = W
  • Dimension relationship dim(W) + dim(W⊥) = dim(V) holds for finite-dimensional inner product spaces
  • Orthogonal complement of a subspace forms a subspace of the inner product space
  • Zero subspace {0} has orthogonal complement V, and V has orthogonal complement {0}

Properties Involving Multiple Subspaces

  • For subspaces U and W in V, (U + W)⊥ = U⊥ ∩ W⊥
  • Intersection and sum relationship (U ∩ W)⊥ = U⊥ + W⊥ holds for subspaces U and W
  • Linear transformation T : V → W between inner product spaces yields ker(T*) = (range(T))⊥
  • Adjoint T* of linear transformation T satisfies range(T*) = (ker(T))⊥

Orthogonal Projections onto Subspaces

Definition and Basic Properties, linear algebra - Relation between Interior Product, Inner Product, Exterior Product, Outer ...

Computation and Properties

  • Orthogonal projection of vector v onto subspace W represents closest vector in W to v
  • Formula for orthogonal projection onto W with orthonormal basis {u₁, ..., uₖ} projW(v)=i=1kv,uiuiproj_W(v) = \sum_{i=1}^k \langle v, u_i \rangle u_i
  • Projection matrix P for orthogonal projection onto W P=U(UU)1UP = U(U^*U)^{-1}U^* where U forms basis for W
  • Orthogonal projection P onto W satisfies P² = P (idempotent) and P* = P (self-adjoint)
  • Orthogonal projection onto W⊥ calculated as v - proj_W(v)
  • Error vector e = v - proj_W(v) orthogonal to all vectors in W

Orthogonal Decomposition

  • Vector v in V decomposes as v = proj_W(v) + proj_W⊥(v)
  • Decomposition represents v as sum of components in W and W⊥
  • Orthogonal decomposition unique for each vector v in V

Orthogonal Decomposition Theorem

Definition and Basic Properties, linear algebra - $\mathbb{C}^3$: Orthogonal Complement - Mathematics Stack Exchange

Statement and Proof

  • Theorem states every vector v in V uniquely expressed as v = w + w⊥, w in W and w⊥ in W⊥
  • Existence proof constructs w = proj_W(v) and w⊥ = v - proj_W(v)
  • Uniqueness proof assumes two decompositions v = w₁ + w₁⊥ = w₂ + w₂⊥ and shows w₁ = w₂ and w₁⊥ = w₂⊥
  • Proof utilizes inner product properties and orthogonal complement definition

Implications and Applications

  • Theorem implies V = W ⊕ W⊥ (direct sum) for any subspace W of V
  • Unique decomposition of vectors into projections onto W and W⊥
  • Fundamental theorem in understanding inner product space structure
  • Applications in linear algebra, functional analysis, and quantum mechanics (state vector decomposition)

Least-Squares Problems with Projections

Problem Formulation and Solution

  • Least-squares problem minimizes ||Ax - b||² for matrix A and vector b
  • Normal equations AAx = Ab characterize least-squares solution
  • Solution x̂ given by x^=(AA)1Ab\hat{x} = (A^*A)^{-1}A^*b when A*A invertible
  • Geometrically, solution finds closest vector Ax̂ in column space of A to b
  • Orthogonal projection of b onto column space of A projcol(A)(b)=A(AA)1Abproj_{col(A)}(b) = A(A^*A)^{-1}A^*b
  • Residual vector r = b - Ax̂ orthogonal to column space of A

Applications and Significance

  • Data fitting applications (linear regression, polynomial fitting)
  • Model matrix A represents predictor variables, b represents observed data
  • Signal processing uses (noise reduction, signal approximation)
  • Parameter estimation in scientific and engineering fields (system identification)
  • Statistical analysis applications (ANOVA, multiple regression)
Pep mascot
Upgrade your Fiveable account to print any study guide

Download study guides as beautiful PDFs See example

Print or share PDFs with your students

Always prints our latest, updated content

Mark up and annotate as you study

Click below to go to billing portal → update your plan → choose Yearly → and select "Fiveable Share Plan". Only pay the difference

Plan is open to all students, teachers, parents, etc
Pep mascot
Upgrade your Fiveable account to export vocabulary

Download study guides as beautiful PDFs See example

Print or share PDFs with your students

Always prints our latest, updated content

Mark up and annotate as you study

Plan is open to all students, teachers, parents, etc
report an error
description

screenshots help us find and fix the issue faster (optional)

add screenshot

2,589 studying →