Why This Matters
Linear transformations are the backbone of linear algebra—they're the functions that actually do something to vectors while respecting the structure of vector spaces. When you're tested on this material, you're not just being asked to recall definitions. You're being evaluated on whether you understand how transformations behave, why certain properties guarantee invertibility, and how matrix operations connect to geometric intuition. Every concept here—from kernels to eigenvalues—builds toward solving systems, analyzing stability, and working in different coordinate systems.
Don't just memorize that "the kernel is the set of vectors mapping to zero." Know why that matters: it tells you about information loss, injectivity, and the dimension of your solution space. Understand how composition corresponds to matrix multiplication, and when a transformation can be reversed. These connections between algebraic properties and geometric meaning are exactly what exam questions target—especially in proofs and applications.
Foundational Definitions and Structure
Linear transformations must satisfy two critical properties that preserve the algebraic structure of vector spaces. These properties—additivity and homogeneity—are what distinguish linear maps from arbitrary functions.
- Preserves vector addition and scalar multiplication—a function T:V→W is linear if T(x+y)=T(x)+T(y) for all vectors
- Homogeneity condition requires T(cx)=cT(x) for any scalar c, meaning scaling before or after transformation gives the same result
- Zero vector always maps to zero—this is a direct consequence of linearity and serves as a quick check for whether a function could be linear
- Every linear transformation has a matrix representation once you fix bases for the domain and codomain vector spaces
- Matrix-vector multiplication computes the transformation: if A represents T, then T(x)=Ax
- Basis choice matters—the same transformation has different matrix representations in different bases, which is why change of basis becomes important
Compare: Definition vs. Matrix Representation—the definition gives you the abstract properties to verify, while the matrix gives you a computational tool. On proofs, use the definition; for calculations, use the matrix.
Kernel, Image, and Injectivity/Surjectivity
These concepts describe what a transformation does to the vector space—what it collapses, what it reaches, and whether information is preserved or lost. The Rank-Nullity Theorem connects these ideas quantitatively.
Kernel (Null Space) and Image (Range)
- Kernel ker(T)={v∈V:T(v)=0}—represents the transformation's "information loss" and is always a subspace
- Image Im(T)={T(v):v∈V}—the transformation's "reach" in the codomain, also a subspace
- Rank-Nullity Theorem states dim(kerT)+dim(Im T)=dim(V), connecting these two fundamental subspaces
- Injective (one-to-one) means ker(T)={0}—no two different inputs produce the same output
- Surjective (onto) means Im(T)=W—every vector in the codomain is reachable
- Bijective transformations are both injective and surjective, which is precisely when an inverse exists
Compare: Kernel vs. Image—kernel measures what's "lost" (dimension = nullity), image measures what's "reached" (dimension = rank). If an FRQ asks about invertibility, check both: trivial kernel AND full image.
Composition and Inverses
When you chain transformations together or undo them, the algebra of matrices mirrors the algebra of functions. Matrix multiplication corresponds to function composition, and matrix inversion corresponds to function inversion.
- Composition preserves linearity—if T1 and T2 are linear, then (T2∘T1)(x)=T2(T1(x)) is also linear
- Matrix product represents composition—if A represents T1 and B represents T2, then BA represents T2∘T1 (note the order!)
- Order matters because matrix multiplication is not commutative; T2∘T1=T1∘T2 in general
- Inverse undoes the transformation—T−1(T(x))=x and T(T−1(y))=y
- Exists if and only if T is bijective (both injective and surjective), which for square matrices means det(A)=0
- Inverse matrix A−1 satisfies A−1A=AA−1=I, and can be computed via row reduction or the adjugate formula
Compare: Composition vs. Inverse—composition builds complexity (multiply matrices), inverse removes it (find A−1). Remember: (AB)−1=B−1A−1—the order reverses!
Eigenvalues and Eigenvectors
Eigenvectors reveal the "natural directions" of a transformation—directions that don't rotate, only stretch or compress. This spectral information is crucial for understanding long-term behavior of iterated transformations and solving differential equations.
- Eigenvector definition—a non-zero vector v satisfying T(v)=λv, meaning the transformation only scales it
- Eigenvalue λ is the scaling factor; found by solving det(A−λI)=0 (the characteristic equation)
- Applications include stability analysis, diagonalization, solving systems of differential equations, and principal component analysis
Compare: Eigenvalues vs. Eigenvectors—eigenvalues tell you how much scaling occurs, eigenvectors tell you in which directions. A transformation can have the same eigenvalue for multiple independent eigenvectors (eigenspaces).
These concrete examples illustrate how abstract linear transformations manifest geometrically. Each type has a characteristic matrix form that you should recognize.
- Rotation by angle θ in R2 uses matrix (cosθsinθ−sinθcosθ)—preserves lengths and angles
- Reflection over a line preserves distances but reverses orientation; determinant equals −1
- Both are orthogonal transformations—their matrices satisfy ATA=I, meaning A−1=AT
- Scaling matrices are diagonal—entries determine stretch factors along coordinate axes; (k100k2) scales by k1 horizontally and k2 vertically
- Shear matrices are triangular—they distort shapes by sliding layers; (10k1) is a horizontal shear
- Shears preserve area (determinant = 1) but change angles; scaling changes area by factor ∣k1k2∣
Compare: Rotation vs. Shear—both can be represented by matrices with 1s on the diagonal (for unit rotations/shears), but rotation preserves angles while shear distorts them. Check the determinant: rotation has det=1, reflection has det=−1.
Change of Basis
Different bases can dramatically simplify how a transformation looks. The goal is often to find a basis of eigenvectors, making the matrix diagonal.
- Same transformation, different matrix—if P is the change of basis matrix, the new representation is P−1AP
- Similar matrices represent the same transformation in different bases; they share eigenvalues, determinant, and trace
- Diagonalization occurs when you can find a basis of eigenvectors, yielding P−1AP=D where D is diagonal
Compare: Original basis vs. Eigenbasis—in the standard basis, a transformation might look complicated; in an eigenbasis (if one exists), it becomes diagonal and trivial to analyze. This is why eigenvalues matter so much!
Quick Reference Table
|
| Linearity Conditions | T(x+y)=T(x)+T(y), T(cx)=cT(x), T(0)=0 |
| Kernel & Injectivity | ker(T)={0} iff T is injective |
| Image & Surjectivity | Im(T)=W iff T is surjective |
| Invertibility | Requires bijection; det(A)=0 for square matrices |
| Composition | (T2∘T1)↔BA (order reverses!) |
| Eigenvalue Equation | Av=λv, solve det(A−λI)=0 |
| Orthogonal Transformations | Rotations, reflections; ATA=I |
| Change of Basis | Similar matrices: P−1AP, same eigenvalues |
Self-Check Questions
-
If T:R3→R3 has a two-dimensional kernel, what can you conclude about its image dimension and whether T is invertible?
-
Compare and contrast rotation and reflection transformations in R2: what properties do they share, and how do their determinants differ?
-
Given that (AB)−1=B−1A−1, explain why the order reverses by thinking about composition of transformations.
-
A transformation has eigenvalues λ1=2 and λ2=−1. What happens to vectors along each eigendirection after applying the transformation twice?
-
Why does changing the basis change the matrix representation but not the kernel dimension or eigenvalues? Connect this to the concept of similar matrices.