All Study Guides Abstract Linear Algebra I Unit 8
🧚🏽♀️ Abstract Linear Algebra I Unit 8 – Inner Products and OrthogonalityInner products and orthogonality are fundamental concepts in linear algebra, extending geometric ideas to abstract vector spaces. These tools allow us to define angles, lengths, and perpendicularity, enabling powerful techniques like orthogonal projections and the Gram-Schmidt process.
These concepts have wide-ranging applications, from least-squares approximations to quantum mechanics. Understanding inner products and orthogonality provides a solid foundation for advanced topics in linear algebra and its applications in various fields of mathematics and science.
Key Concepts and Definitions
Inner product is a generalization of the dot product that allows us to define angles and lengths in abstract vector spaces
Orthogonality refers to two vectors being perpendicular or at right angles to each other
Orthogonal vectors have an inner product of zero
Norm of a vector is a measure of its length or magnitude in a vector space
Induced by the inner product as ⟨ v , v ⟩ \sqrt{\langle v, v \rangle} ⟨ v , v ⟩
Orthonormal set is a collection of vectors that are both orthogonal to each other and have unit norm (length 1)
Orthogonal projection is the process of finding the closest point in a subspace to a given vector
Useful for approximating solutions and minimizing errors
Gram-Schmidt process is an algorithm for constructing an orthonormal basis from a linearly independent set of vectors
Cauchy-Schwarz inequality states that the absolute value of the inner product of two vectors is less than or equal to the product of their norms
∣ ⟨ u , v ⟩ ∣ ≤ ∥ u ∥ ∥ v ∥ |\langle u, v \rangle| \leq \|u\| \|v\| ∣ ⟨ u , v ⟩ ∣ ≤ ∥ u ∥∥ v ∥
Inner Product Spaces
An inner product space is a vector space equipped with an inner product operation
Inner product is a function that takes two vectors and returns a scalar value
Denoted as ⟨ u , v ⟩ \langle u, v \rangle ⟨ u , v ⟩ for vectors u u u and v v v
Inner product spaces allow us to define geometric concepts like angles, lengths, and orthogonality in abstract vector spaces
Examples of inner product spaces include:
Euclidean space R n \mathbb{R}^n R n with the dot product
Space of continuous functions C [ a , b ] C[a, b] C [ a , b ] with the integral inner product ⟨ f , g ⟩ = ∫ a b f ( x ) g ( x ) d x \langle f, g \rangle = \int_a^b f(x)g(x) dx ⟨ f , g ⟩ = ∫ a b f ( x ) g ( x ) d x
Inner product spaces have a rich structure and satisfy several important properties
Symmetry, linearity, and positive-definiteness
Many concepts from Euclidean geometry can be generalized to inner product spaces
Orthogonal projections, Gram-Schmidt process, and least-squares approximations
Properties of Inner Products
Symmetry: ⟨ u , v ⟩ = ⟨ v , u ⟩ \langle u, v \rangle = \langle v, u \rangle ⟨ u , v ⟩ = ⟨ v , u ⟩ for all vectors u u u and v v v
Linearity in the first argument:
⟨ a u , v ⟩ = a ⟨ u , v ⟩ \langle au, v \rangle = a\langle u, v \rangle ⟨ a u , v ⟩ = a ⟨ u , v ⟩ for any scalar a a a
⟨ u 1 + u 2 , v ⟩ = ⟨ u 1 , v ⟩ + ⟨ u 2 , v ⟩ \langle u_1 + u_2, v \rangle = \langle u_1, v \rangle + \langle u_2, v \rangle ⟨ u 1 + u 2 , v ⟩ = ⟨ u 1 , v ⟩ + ⟨ u 2 , v ⟩ for any vectors u 1 u_1 u 1 , u 2 u_2 u 2 , and v v v
Positive-definiteness: ⟨ v , v ⟩ ≥ 0 \langle v, v \rangle \geq 0 ⟨ v , v ⟩ ≥ 0 for all vectors v v v , with equality if and only if v = 0 v = 0 v = 0
Cauchy-Schwarz inequality: ∣ ⟨ u , v ⟩ ∣ ≤ ∥ u ∥ ∥ v ∥ |\langle u, v \rangle| \leq \|u\| \|v\| ∣ ⟨ u , v ⟩ ∣ ≤ ∥ u ∥∥ v ∥ for all vectors u u u and v v v
Equality holds if and only if u u u and v v v are linearly dependent
Norm induced by the inner product: ∥ v ∥ = ⟨ v , v ⟩ \|v\| = \sqrt{\langle v, v \rangle} ∥ v ∥ = ⟨ v , v ⟩
Satisfies the properties of a norm (non-negativity, homogeneity, and triangle inequality)
Parallelogram law: ∥ u + v ∥ 2 + ∥ u − v ∥ 2 = 2 ( ∥ u ∥ 2 + ∥ v ∥ 2 ) \|u + v\|^2 + \|u - v\|^2 = 2(\|u\|^2 + \|v\|^2) ∥ u + v ∥ 2 + ∥ u − v ∥ 2 = 2 ( ∥ u ∥ 2 + ∥ v ∥ 2 ) for all vectors u u u and v v v
Orthogonality and Orthonormal Sets
Two vectors u u u and v v v are orthogonal if their inner product is zero: ⟨ u , v ⟩ = 0 \langle u, v \rangle = 0 ⟨ u , v ⟩ = 0
Orthogonal vectors are perpendicular or at right angles to each other
An orthogonal set is a collection of non-zero vectors that are pairwise orthogonal
⟨ u i , u j ⟩ = 0 \langle u_i, u_j \rangle = 0 ⟨ u i , u j ⟩ = 0 for all i ≠ j i \neq j i = j
Orthonormal set is an orthogonal set where each vector has unit norm (length 1)
∥ u i ∥ = 1 \|u_i\| = 1 ∥ u i ∥ = 1 for all vectors u i u_i u i in the set
Orthonormal sets are particularly useful as bases for inner product spaces
Coefficients of a vector with respect to an orthonormal basis are easily computed using inner products
Orthonormal bases simplify many computations and have desirable numerical properties
Minimize roundoff errors and provide a natural coordinate system
Examples of orthonormal sets include:
Standard basis vectors { e 1 , e 2 , … , e n } \{e_1, e_2, \ldots, e_n\} { e 1 , e 2 , … , e n } in R n \mathbb{R}^n R n
Trigonometric functions { 1 2 π , 1 π cos ( n x ) , 1 π sin ( n x ) } \{\frac{1}{\sqrt{2\pi}}, \frac{1}{\sqrt{\pi}}\cos(nx), \frac{1}{\sqrt{\pi}}\sin(nx)\} { 2 π 1 , π 1 cos ( n x ) , π 1 sin ( n x )} in L 2 [ − π , π ] L^2[-\pi, \pi] L 2 [ − π , π ]
Gram-Schmidt Process
Gram-Schmidt process is an algorithm for constructing an orthonormal basis from a linearly independent set of vectors
Takes a linearly independent set { v 1 , v 2 , … , v n } \{v_1, v_2, \ldots, v_n\} { v 1 , v 2 , … , v n } and produces an orthonormal set { u 1 , u 2 , … , u n } \{u_1, u_2, \ldots, u_n\} { u 1 , u 2 , … , u n }
The process works by iteratively orthogonalizing and normalizing the vectors
Set u 1 = v 1 ∥ v 1 ∥ u_1 = \frac{v_1}{\|v_1\|} u 1 = ∥ v 1 ∥ v 1
For i = 2 , … , n i = 2, \ldots, n i = 2 , … , n :
Compute the projection of v i v_i v i onto the subspace spanned by { u 1 , … , u i − 1 } \{u_1, \ldots, u_{i-1}\} { u 1 , … , u i − 1 } :
p r o j i = ∑ j = 1 i − 1 ⟨ v i , u j ⟩ u j proj_i = \sum_{j=1}^{i-1} \langle v_i, u_j \rangle u_j p ro j i = ∑ j = 1 i − 1 ⟨ v i , u j ⟩ u j
Subtract the projection from v i v_i v i to obtain the orthogonal component:
u i ′ = v i − p r o j i u_i' = v_i - proj_i u i ′ = v i − p ro j i
Normalize u i ′ u_i' u i ′ to obtain the orthonormal vector:
u i = u i ′ ∥ u i ′ ∥ u_i = \frac{u_i'}{\|u_i'\|} u i = ∥ u i ′ ∥ u i ′
The resulting set { u 1 , u 2 , … , u n } \{u_1, u_2, \ldots, u_n\} { u 1 , u 2 , … , u n } is an orthonormal basis for the subspace spanned by the original vectors
Gram-Schmidt process is widely used in numerical linear algebra and has applications in:
Least-squares approximations
QR factorization
Solving systems of linear equations
Orthogonal Projections
Orthogonal projection is the process of finding the closest point in a subspace to a given vector
Given a subspace W W W and a vector v v v , the orthogonal projection of v v v onto W W W is the unique vector p r o j W ( v ) proj_W(v) p ro j W ( v ) in W W W that minimizes the distance to v v v
p r o j W ( v ) = arg min w ∈ W ∥ v − w ∥ proj_W(v) = \arg\min_{w \in W} \|v - w\| p ro j W ( v ) = arg min w ∈ W ∥ v − w ∥
Orthogonal projection can be computed using an orthonormal basis { u 1 , … , u k } \{u_1, \ldots, u_k\} { u 1 , … , u k } for the subspace W W W :
p r o j W ( v ) = ∑ i = 1 k ⟨ v , u i ⟩ u i proj_W(v) = \sum_{i=1}^k \langle v, u_i \rangle u_i p ro j W ( v ) = ∑ i = 1 k ⟨ v , u i ⟩ u i
Properties of orthogonal projections:
p r o j W ( v ) proj_W(v) p ro j W ( v ) is the unique vector in W W W such that v − p r o j W ( v ) v - proj_W(v) v − p ro j W ( v ) is orthogonal to every vector in W W W
p r o j W proj_W p ro j W is a linear transformation
p r o j W ( v ) = v proj_W(v) = v p ro j W ( v ) = v if and only if v ∈ W v \in W v ∈ W
∥ v − p r o j W ( v ) ∥ ≤ ∥ v − w ∥ \|v - proj_W(v)\| \leq \|v - w\| ∥ v − p ro j W ( v ) ∥ ≤ ∥ v − w ∥ for all w ∈ W w \in W w ∈ W
Orthogonal projections have numerous applications, including:
Least-squares approximations and regression analysis
Signal and image processing (denoising, compression)
Solving systems of linear equations and optimization problems
Applications in Linear Algebra
Inner products and orthogonality have a wide range of applications in linear algebra and related fields
Least-squares approximations:
Finding the best approximation of a vector in a subspace
Minimizing the sum of squared errors between data points and a model
Orthogonal diagonalization of symmetric matrices:
Eigenvectors of a symmetric matrix form an orthonormal basis
Allows for efficient computation of matrix powers and exponentials
Principal component analysis (PCA):
Identifying the directions of maximum variance in a dataset
Useful for dimensionality reduction and data visualization
Quantum mechanics:
State vectors in a Hilbert space (an infinite-dimensional inner product space)
Observables represented by Hermitian operators with orthogonal eigenvectors
Fourier analysis and signal processing:
Representing functions as linear combinations of orthogonal basis functions (e.g., trigonometric functions, wavelets)
Analyzing and filtering signals in the frequency domain
Practice Problems and Examples
Verify that the following functions define an inner product on the vector space of continuous functions C [ 0 , 1 ] C[0, 1] C [ 0 , 1 ] :
⟨ f , g ⟩ = ∫ 0 1 f ( x ) g ( x ) d x \langle f, g \rangle = \int_0^1 f(x)g(x) dx ⟨ f , g ⟩ = ∫ 0 1 f ( x ) g ( x ) d x
⟨ f , g ⟩ = f ( 0 ) g ( 0 ) + ∫ 0 1 f ′ ( x ) g ′ ( x ) d x \langle f, g \rangle = f(0)g(0) + \int_0^1 f'(x)g'(x) dx ⟨ f , g ⟩ = f ( 0 ) g ( 0 ) + ∫ 0 1 f ′ ( x ) g ′ ( x ) d x
Compute the orthogonal projection of the vector v = ( 1 , 2 , 3 ) v = (1, 2, 3) v = ( 1 , 2 , 3 ) onto the subspace W = span { ( 1 , 1 , 1 ) , ( 1 , 0 , − 1 ) } W = \text{span}\{(1, 1, 1), (1, 0, -1)\} W = span {( 1 , 1 , 1 ) , ( 1 , 0 , − 1 )} in R 3 \mathbb{R}^3 R 3 .
Apply the Gram-Schmidt process to the following set of vectors in R 4 \mathbb{R}^4 R 4 :
v 1 = ( 1 , 0 , 0 , 0 ) v_1 = (1, 0, 0, 0) v 1 = ( 1 , 0 , 0 , 0 ) , v 2 = ( 1 , 1 , 0 , 0 ) v_2 = (1, 1, 0, 0) v 2 = ( 1 , 1 , 0 , 0 ) , v 3 = ( 1 , 1 , 1 , 0 ) v_3 = (1, 1, 1, 0) v 3 = ( 1 , 1 , 1 , 0 ) , v 4 = ( 1 , 1 , 1 , 1 ) v_4 = (1, 1, 1, 1) v 4 = ( 1 , 1 , 1 , 1 )
Prove that if { u 1 , … , u n } \{u_1, \ldots, u_n\} { u 1 , … , u n } is an orthonormal set in an inner product space V V V , then for any vector v ∈ V v \in V v ∈ V :
∥ v ∥ 2 = ∑ i = 1 n ∣ ⟨ v , u i ⟩ ∣ 2 + ∥ v − ∑ i = 1 n ⟨ v , u i ⟩ u i ∥ 2 \|v\|^2 = \sum_{i=1}^n |\langle v, u_i \rangle|^2 + \|v - \sum_{i=1}^n \langle v, u_i \rangle u_i\|^2 ∥ v ∥ 2 = ∑ i = 1 n ∣ ⟨ v , u i ⟩ ∣ 2 + ∥ v − ∑ i = 1 n ⟨ v , u i ⟩ u i ∥ 2
Find the closest point in the plane x + y + z = 1 x + y + z = 1 x + y + z = 1 to the point ( 2 , 3 , 4 ) (2, 3, 4) ( 2 , 3 , 4 ) in R 3 \mathbb{R}^3 R 3 .
Determine whether the following sets of vectors are orthogonal, orthonormal, or neither:
{ ( 1 , 1 , 0 ) , ( 1 , − 1 , 0 ) , ( 0 , 0 , 1 ) } \{(1, 1, 0), (1, -1, 0), (0, 0, 1)\} {( 1 , 1 , 0 ) , ( 1 , − 1 , 0 ) , ( 0 , 0 , 1 )} in R 3 \mathbb{R}^3 R 3
{ ( 1 , 0 , 1 ) , ( 0 , 1 , 1 ) , ( − 1 , 1 , 0 ) } \{(1, 0, 1), (0, 1, 1), (-1, 1, 0)\} {( 1 , 0 , 1 ) , ( 0 , 1 , 1 ) , ( − 1 , 1 , 0 )} in R 3 \mathbb{R}^3 R 3
{ sin x , cos x } \{\sin x, \cos x\} { sin x , cos x } in L 2 [ − π , π ] L^2[-\pi, \pi] L 2 [ − π , π ]
Compute the Fourier coefficients of the function f ( x ) = x 2 f(x) = x^2 f ( x ) = x 2 on the interval [ − π , π ] [-\pi, \pi] [ − π , π ] with respect to the orthonormal basis { 1 2 π , 1 π cos ( n x ) , 1 π sin ( n x ) } \{\frac{1}{\sqrt{2\pi}}, \frac{1}{\sqrt{\pi}}\cos(nx), \frac{1}{\sqrt{\pi}}\sin(nx)\} { 2 π 1 , π 1 cos ( n x ) , π 1 sin ( n x )} .