Fiveable

🧠Thinking Like a Mathematician Unit 8 Review

QR code for Thinking Like a Mathematician practice questions

8.4 Vector spaces

8.4 Vector spaces

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🧠Thinking Like a Mathematician
Unit & Topic Study Guides

Definition of vector spaces

A vector space is a collection of objects (called vectors) that you can add together and multiply by scalars, with the results always staying in the collection. This structure shows up everywhere in mathematics, physics, and computer science because it captures the essence of "linear" behavior in a single, clean framework.

What makes vector spaces powerful is their generality. Once you prove something about vector spaces in the abstract, that result applies to arrows in 3D space, to polynomials, to matrices, and to functions. You learn one set of rules and unlock tools for dozens of different settings.

Properties of vector spaces

A vector space over a field F\mathbb{F} (usually R\mathbb{R} or C\mathbb{C}) must satisfy all of the following axioms. Missing even one means you don't have a vector space.

Addition axioms:

  • Closure under addition: If u\vec{u} and v\vec{v} are in the space, then u+v\vec{u} + \vec{v} is also in the space.
  • Associativity: (u+v)+w=u+(v+w)(\vec{u} + \vec{v}) + \vec{w} = \vec{u} + (\vec{v} + \vec{w})
  • Commutativity: u+v=v+u\vec{u} + \vec{v} = \vec{v} + \vec{u}
  • Zero vector: There exists a vector 0\vec{0} such that v+0=v\vec{v} + \vec{0} = \vec{v} for all v\vec{v}.
  • Additive inverses: For every v\vec{v}, there exists v-\vec{v} such that v+(v)=0\vec{v} + (-\vec{v}) = \vec{0}.

Scalar multiplication axioms:

  • Closure under scalar multiplication: If cFc \in \mathbb{F} and v\vec{v} is in the space, then cvc\vec{v} is in the space.
  • Distributivity over vector addition: c(u+v)=cu+cvc(\vec{u} + \vec{v}) = c\vec{u} + c\vec{v}
  • Distributivity over scalar addition: (a+b)v=av+bv(a + b)\vec{v} = a\vec{v} + b\vec{v}
  • Associativity of scalar multiplication: (ab)v=a(bv)(ab)\vec{v} = a(b\vec{v})
  • Multiplicative identity: 1v=v1\vec{v} = \vec{v}

Examples of vector spaces

  • Rn\mathbb{R}^n (real coordinate spaces): Vectors with nn real-number components. R2\mathbb{R}^2 is the familiar plane, R3\mathbb{R}^3 is 3D space.
  • Cn\mathbb{C}^n (complex coordinate spaces): Same idea, but components are complex numbers.
  • Polynomial spaces PnP_n: All polynomials of degree n\leq n. For example, P2P_2 includes vectors like 3x2x+73x^2 - x + 7. Addition and scalar multiplication work the way you'd expect.
  • Matrix spaces Mm×nM_{m \times n}: All m×nm \times n matrices. You add them entry-by-entry and scale them entry-by-entry.
  • Function spaces: Sets of functions (say, all continuous functions on [0,1][0,1]) that satisfy the axioms when you define addition and scaling pointwise.

Non-examples of vector spaces

Non-examples sharpen your understanding of why each axiom matters.

  • The set of positive real numbers (under usual addition): The additive inverse of 55 is 5-5, which isn't positive. No additive inverses means no vector space.
  • The set of integers Z\mathbb{Z} (as a space over R\mathbb{R}): Scalar multiplication fails closure. Multiplying the integer 33 by the scalar 12\frac{1}{2} gives 1.51.5, which isn't an integer.
  • A circle in R2\mathbb{R}^2: Scaling a point on a unit circle by 22 moves it off the circle. No closure under scalar multiplication.
  • A plane in R3\mathbb{R}^3 that doesn't pass through the origin: It won't contain the zero vector, so it fails that axiom. (A plane through the origin, however, is a valid subspace.)

Vector operations

The two core operations in any vector space are vector addition and scalar multiplication. Everything else builds on these.

Vector addition

Vector addition combines two vectors component by component. In R3\mathbb{R}^3, for instance: 1,2,3+4,1,0=5,1,3\langle 1, 2, 3 \rangle + \langle 4, -1, 0 \rangle = \langle 5, 1, 3 \rangle.

Geometrically, you can visualize this with the parallelogram law: place the tails of two vectors at the same point, complete the parallelogram, and the diagonal is the sum.

The key properties carry over from the axioms:

  • Commutative: a+b=b+a\vec{a} + \vec{b} = \vec{b} + \vec{a}
  • Associative: (a+b)+c=a+(b+c)(\vec{a} + \vec{b}) + \vec{c} = \vec{a} + (\vec{b} + \vec{c})
  • Identity: a+0=a\vec{a} + \vec{0} = \vec{a}

Scalar multiplication

Scalar multiplication scales a vector's magnitude and can reverse its direction. Multiplying each component by the scalar does the job: 31,2=3,63 \langle 1, -2 \rangle = \langle 3, -6 \rangle.

  • A scalar >1> 1 stretches the vector; a scalar between 00 and 11 shrinks it.
  • A negative scalar flips the vector's direction.
  • The scalar 11 leaves the vector unchanged: 1v=v1\vec{v} = \vec{v}.
  • Distributive properties connect the two operations: c(a+b)=ca+cbc(\vec{a} + \vec{b}) = c\vec{a} + c\vec{b} and (a+b)v=av+bv(a + b)\vec{v} = a\vec{v} + b\vec{v}.

Linear combinations

A linear combination takes a set of vectors, scales each one, and adds the results:

c1v1+c2v2++cnvnc_1\vec{v}_1 + c_2\vec{v}_2 + \cdots + c_n\vec{v}_n

The scalars c1,c2,,cnc_1, c_2, \ldots, c_n are called coefficients and can be any elements of the field.

Linear combinations are the central building block for nearly everything that follows. Span, linear independence, bases, and solutions to linear systems are all defined in terms of linear combinations.

Subspaces

A subspace is a subset of a vector space that is itself a vector space under the same addition and scalar multiplication. Think of it as a "smaller world" living inside a bigger one that still follows all the same rules.

Definition of subspaces

A subset WW of a vector space VV is a subspace if:

  • WW uses the same addition and scalar multiplication as VV.
  • WW satisfies all the vector space axioms on its own.

Most of the axioms (associativity, commutativity, distributivity, etc.) are automatically inherited from the parent space. So you don't need to check all ten axioms from scratch. You just need the subspace test.

Criteria for subspaces (the subspace test)

To verify that WVW \subseteq V is a subspace, check three things:

  1. Non-empty: WW contains the zero vector 0\vec{0}.
  2. Closed under addition: If u,vW\vec{u}, \vec{v} \in W, then u+vW\vec{u} + \vec{v} \in W.
  3. Closed under scalar multiplication: If vW\vec{v} \in W and cFc \in \mathbb{F}, then cvWc\vec{v} \in W.

You can combine steps 2 and 3 into a single check: verify closure under linear combinations. If u,vW\vec{u}, \vec{v} \in W and a,bFa, b \in \mathbb{F}, then au+bvWa\vec{u} + b\vec{v} \in W.

Properties of vector spaces, vector spaces - quick way to check Linear Independence - Mathematics Stack Exchange

Common subspaces

  • Null space (kernel) of a matrix AA: all vectors x\vec{x} such that Ax=0A\vec{x} = \vec{0}. This is always a subspace.
  • Column space (image) of a matrix: the span of the columns of AA. Tells you which outputs are reachable.
  • Row space: the span of the rows of AA.
  • Eigenspaces: for a given eigenvalue λ\lambda, the set of all vectors satisfying Av=λvA\vec{v} = \lambda\vec{v}.
  • Solution sets of homogeneous systems (Ax=0A\vec{x} = \vec{0}) are subspaces. Non-homogeneous solution sets (Ax=bA\vec{x} = \vec{b} with b0\vec{b} \neq \vec{0}) are not subspaces because they don't contain 0\vec{0}.

Span and linear independence

These two concepts work together to describe the "reach" and "efficiency" of a set of vectors.

Span of vectors

The span of a set of vectors is the collection of all linear combinations you can form from them:

Span{v1,v2,,vn}={c1v1+c2v2++cnvnciF}\text{Span}\{\vec{v}_1, \vec{v}_2, \ldots, \vec{v}_n\} = \{c_1\vec{v}_1 + c_2\vec{v}_2 + \cdots + c_n\vec{v}_n \mid c_i \in \mathbb{F}\}

The span is always a subspace (it's the smallest subspace containing those vectors). If the span equals the entire vector space VV, you say the vectors span VV, meaning every vector in VV can be built from them.

Linear independence vs. dependence

A set of vectors {v1,,vn}\{\vec{v}_1, \ldots, \vec{v}_n\} is linearly independent if the only way to get the zero vector from a linear combination is the trivial way:

c1v1+c2v2++cnvn=0    c1=c2==cn=0c_1\vec{v}_1 + c_2\vec{v}_2 + \cdots + c_n\vec{v}_n = \vec{0} \implies c_1 = c_2 = \cdots = c_n = 0

If there's a non-trivial combination that gives 0\vec{0}, the set is linearly dependent. That means at least one vector in the set is redundant: it can be written as a linear combination of the others.

Geometric intuition:

  • In R2\mathbb{R}^2: two vectors are independent if they point in different (non-parallel) directions.
  • In R3\mathbb{R}^3: three vectors are independent if they don't all lie in the same plane.

Basis of a vector space

A basis is a set of vectors that is both linearly independent and spans the entire space. It's the "just right" set: no redundant vectors, but enough to reach everything.

Key facts about bases:

  • Every vector in the space can be written as a unique linear combination of basis vectors.
  • A basis is a minimal spanning set (remove any vector and it no longer spans).
  • A basis is a maximal independent set (add any vector and it becomes dependent).
  • The standard basis for R3\mathbb{R}^3 is {1,0,0,0,1,0,0,0,1}\{\langle 1,0,0 \rangle, \langle 0,1,0 \rangle, \langle 0,0,1 \rangle\}.
  • Different bases give different coordinate representations of the same vectors, which is useful for simplifying problems.

Dimension of vector spaces

The dimension of a vector space is the number of vectors in any basis. This single number captures a lot about the space's structure.

Finite vs. infinite dimensions

  • Finite-dimensional: Rn\mathbb{R}^n has dimension nn. The polynomial space P3P_3 (polynomials of degree 3\leq 3) has dimension 4, because a basis is {1,x,x2,x3}\{1, x, x^2, x^3\}.
  • Infinite-dimensional: The space of all polynomials (no degree bound) is infinite-dimensional. So is the space of continuous functions on [0,1][0,1]. No finite set of vectors can span these spaces.

Dimension theorem

Every basis of a given vector space has the same number of elements. This is why dimension is well-defined: it doesn't depend on which basis you choose.

Useful consequences:

  • If dim(V)=n\dim(V) = n, then any set of more than nn vectors in VV must be linearly dependent.
  • Any linearly independent set of exactly nn vectors in VV automatically spans VV (and is therefore a basis).
  • If WW is a subspace of VV, then dim(W)dim(V)\dim(W) \leq \dim(V).

Coordinate systems

Once you fix a basis {b1,,bn}\{\vec{b}_1, \ldots, \vec{b}_n\} for a space, every vector v\vec{v} can be written uniquely as:

v=c1b1+c2b2++cnbn\vec{v} = c_1\vec{b}_1 + c_2\vec{b}_2 + \cdots + c_n\vec{b}_n

The scalars (c1,c2,,cn)(c_1, c_2, \ldots, c_n) are the coordinates of v\vec{v} relative to that basis. The standard basis in Rn\mathbb{R}^n gives the familiar coordinates you're used to. Switching to a different basis can simplify a problem dramatically, which is why change of basis formulas matter.

Vector space transformations

A transformation between vector spaces is a function that sends vectors from one space to another. The most important kind preserves the linear structure.

Properties of vector spaces, linear algebra - Prove in full detail that the set is a vector space - Mathematics Stack Exchange

Linear transformations

A function T:VWT: V \to W is a linear transformation if it satisfies two conditions for all vectors u,vV\vec{u}, \vec{v} \in V and all scalars cc:

  1. T(u+v)=T(u)+T(v)T(\vec{u} + \vec{v}) = T(\vec{u}) + T(\vec{v})
  2. T(cv)=cT(v)T(c\vec{v}) = cT(\vec{v})

These two conditions together mean TT preserves linear combinations. In finite-dimensional spaces, every linear transformation can be represented by a matrix. Familiar geometric operations like rotations, reflections, projections, and scaling are all linear transformations.

Kernel and image

Every linear transformation T:VWT: V \to W has two important subspaces associated with it:

  • Kernel (null space): ker(T)={vVT(v)=0}\ker(T) = \{\vec{v} \in V \mid T(\vec{v}) = \vec{0}\}. This is a subspace of VV. If the kernel contains only 0\vec{0}, then TT is injective (one-to-one).
  • Image (range): im(T)={T(v)vV}\text{im}(T) = \{T(\vec{v}) \mid \vec{v} \in V\}. This is a subspace of WW. If the image equals all of WW, then TT is surjective (onto).

The Rank-Nullity Theorem ties these together:

dim(ker(T))+dim(im(T))=dim(V)\dim(\ker(T)) + \dim(\text{im}(T)) = \dim(V)

This is one of the most useful results in linear algebra. It tells you that the "information lost" by TT (the kernel) plus the "information preserved" (the image) always adds up to the dimension of the domain.

Isomorphisms between spaces

An isomorphism is a linear transformation that is both injective and surjective (bijective). Two vector spaces are isomorphic if there exists an isomorphism between them.

Isomorphic spaces are structurally identical from a linear algebra perspective. The key result: two finite-dimensional vector spaces over the same field are isomorphic if and only if they have the same dimension. So R3\mathbb{R}^3, P2P_2, and M1×3M_{1 \times 3} are all isomorphic because each has dimension 3.

Inner product spaces

An inner product space is a vector space equipped with an additional operation called an inner product. This operation lets you define geometric concepts like length, distance, and angle within the algebraic framework of vector spaces.

Definition of inner products

An inner product on a vector space VV is a function ,:V×VF\langle \cdot, \cdot \rangle: V \times V \to \mathbb{F} satisfying:

  • Positive definiteness: v,v0\langle \vec{v}, \vec{v} \rangle \geq 0, with equality only when v=0\vec{v} = \vec{0}.
  • Conjugate symmetry: u,v=v,u\langle \vec{u}, \vec{v} \rangle = \overline{\langle \vec{v}, \vec{u} \rangle}. (For real spaces, this simplifies to plain symmetry: u,v=v,u\langle \vec{u}, \vec{v} \rangle = \langle \vec{v}, \vec{u} \rangle.)
  • Linearity in the first argument: au+bv,w=au,w+bv,w\langle a\vec{u} + b\vec{v}, \vec{w} \rangle = a\langle \vec{u}, \vec{w} \rangle + b\langle \vec{v}, \vec{w} \rangle.

The standard example is the dot product in Rn\mathbb{R}^n: u,v=u1v1+u2v2++unvn\langle \vec{u}, \vec{v} \rangle = u_1v_1 + u_2v_2 + \cdots + u_nv_n.

From an inner product, you can define the norm (length) of a vector: v=v,v\|\vec{v}\| = \sqrt{\langle \vec{v}, \vec{v} \rangle}.

Orthogonality and orthonormality

Two vectors are orthogonal if u,v=0\langle \vec{u}, \vec{v} \rangle = 0. In geometric terms, they're perpendicular.

An orthonormal set goes one step further: the vectors are orthogonal and each has unit length (v=1\|\vec{v}\| = 1).

Why care? Orthonormal bases make computations much cleaner. If {e1,,en}\{\vec{e}_1, \ldots, \vec{e}_n\} is an orthonormal basis, then the coordinates of any vector v\vec{v} are simply ci=v,eic_i = \langle \vec{v}, \vec{e}_i \rangle. No systems of equations needed.

Gram-Schmidt process

The Gram-Schmidt process takes any linearly independent set and produces an orthonormal basis spanning the same subspace.

Steps:

  1. Start with a linearly independent set {v1,v2,,vn}\{\vec{v}_1, \vec{v}_2, \ldots, \vec{v}_n\}.

  2. Set u1=v1\vec{u}_1 = \vec{v}_1.

  3. For each subsequent vector vk\vec{v}_k, subtract off its projections onto all previously computed u\vec{u}'s: uk=vkj=1k1vk,ujuj,ujuj\vec{u}_k = \vec{v}_k - \sum_{j=1}^{k-1} \frac{\langle \vec{v}_k, \vec{u}_j \rangle}{\langle \vec{u}_j, \vec{u}_j \rangle} \vec{u}_j

  4. Normalize each uk\vec{u}_k to get a unit vector: ek=ukuk\vec{e}_k = \frac{\vec{u}_k}{\|\vec{u}_k\|}.

The result {e1,,en}\{\vec{e}_1, \ldots, \vec{e}_n\} is orthonormal and spans the same subspace as the original set. This process is used in least squares problems, QR factorization, and quantum mechanics.

Applications of vector spaces

Vector spaces aren't just abstract theory. They provide the language and tools for solving concrete problems across many fields.

Linear algebra connections

  • Systems of linear equations translate directly into questions about spans, null spaces, and column spaces.
  • Eigenvalue problems arise in differential equations and dynamical systems, where you need to find vectors that a transformation only scales.
  • Matrix decompositions (SVD, LU, QR) break matrices into simpler pieces for data compression, numerical stability, and analysis.
  • Least squares approximation finds the "best fit" when an exact solution doesn't exist, using projections onto subspaces.
  • Fourier analysis represents signals as linear combinations of sine and cosine functions, treating them as vectors in a function space.

Physics and engineering uses

  • Quantum mechanics represents particle states as vectors in Hilbert spaces (infinite-dimensional inner product spaces).
  • Electromagnetic theory models fields as vector-valued functions across space.
  • Structural engineering uses finite element methods, which discretize continuous structures into systems of linear equations.
  • Control theory describes dynamic systems using state vectors and linear transformations.
  • Robotics relies on transformation matrices for motion planning and positioning.

Computer graphics applications

  • 3D transformations (rotation, translation, scaling) manipulate objects and cameras using matrix multiplication.
  • Texture mapping applies coordinate transformations to wrap images onto surfaces.
  • Ray tracing computes light paths using vector operations to produce realistic images.
  • Animation interpolates between positions and orientations in vector spaces to create smooth motion.
  • Color spaces treat colors as vectors, and transformations convert between different color representations (RGB, HSV, etc.).