Why This Matters
Diagonalization is one of the most powerful techniques you'll encounter in linear algebra because it transforms complex matrix operations into simple arithmetic. When you can express a matrix as A=PDP−1, suddenly computing A100 becomes trivial, solving systems of differential equations becomes manageable, and understanding how a linear transformation behaves becomes intuitive. You're being tested on your ability to recognize when diagonalization is possible, execute the computational steps correctly, and understand why the process works.
The concepts here—eigenvalues, eigenvectors, characteristic polynomials, and multiplicity conditions—form the backbone of advanced applications in differential equations, quantum mechanics, data science, and stability analysis. Don't just memorize the formula A=PDP−1; know what each piece represents, why the conditions for diagonalizability exist, and how to troubleshoot when a matrix fails to diagonalize.
Foundational Definitions
Before diving into techniques, you need rock-solid definitions. These terms appear constantly in exam questions, so precision matters.
Definition of Diagonalization
- Diagonalization converts a square matrix into diagonal form—specifically, expressing A=PDP−1 where D contains eigenvalues on its diagonal
- Matrix P contains eigenvectors as columns—this invertible matrix performs the "change of basis" that reveals the diagonal structure
- Diagonal matrices simplify computations dramatically—powers become Ak=PDkP−1, reducing matrix exponentiation to scalar exponentiation
Eigenvalues and Eigenvectors
- Eigenvalues are scalars λ satisfying Av=λv—they tell you how much the transformation stretches along special directions
- Eigenvectors represent invariant directions—vectors that only get scaled (not rotated) when the matrix acts on them
- The spectrum (set of all eigenvalues) determines matrix behavior—including stability, convergence, and long-term dynamics
Eigenvalue Equation (Av=λv)
- This equation defines the eigenvalue-eigenvector relationship—the matrix A acting on v produces a scalar multiple of v
- Eigenvectors are special because they resist rotation—they emerge unchanged in direction, only scaled by λ
- Rearranging gives (A−λI)v=0—this homogeneous system is how you actually compute eigenvectors
Compare: Eigenvalues vs. Eigenvectors—eigenvalues are scalars telling you the scaling factor, while eigenvectors are vectors telling you the direction. On FRQs, you'll often need to find both: eigenvalues first (from the characteristic equation), then eigenvectors (by solving the null space).
Finding Eigenvalues
The characteristic polynomial is your primary tool for extracting eigenvalues. Understanding this process is essential for every diagonalization problem.
Characteristic Equation and Polynomial
- The characteristic polynomial comes from det(A−λI)=0—setting this determinant equal to zero yields the eigenvalues as roots
- Roots of the polynomial are the eigenvalues—for a 2×2 matrix, you'll typically use the quadratic formula; larger matrices may factor or require numerical methods
- Polynomial degree equals matrix size—an n×n matrix has a degree-n characteristic polynomial with exactly n eigenvalues (counting multiplicity)
Multiplicity and Diagonalizability Conditions
This is where students often lose points. The relationship between algebraic and geometric multiplicity determines whether diagonalization succeeds or fails.
Algebraic and Geometric Multiplicity
- Algebraic multiplicity counts repeated roots—how many times λ appears as a solution to the characteristic equation
- Geometric multiplicity counts independent eigenvectors—the dimension of the eigenspace null(A−λI)
- Key inequality: geometric ≤ algebraic always holds—diagonalization requires equality for every eigenvalue
Conditions for a Matrix to Be Diagonalizable
- Sufficient eigenvectors must exist—specifically, n linearly independent eigenvectors for an n×n matrix
- Distinct eigenvalues guarantee diagonalizability—if all n eigenvalues differ, you automatically have n independent eigenvectors
- Repeated eigenvalues require checking geometric multiplicity—this is where diagonalization can fail if the eigenspace is "deficient"
Compare: Algebraic vs. Geometric Multiplicity—algebraic multiplicity comes from the characteristic polynomial (how many times the root repeats), while geometric multiplicity comes from the null space dimension (how many independent eigenvectors exist). If an FRQ asks "is this matrix diagonalizable?", check whether these match for each eigenvalue.
The Theory Behind Diagonalization
Understanding why diagonalization works helps you apply it correctly and recognize edge cases.
Diagonalization Theorem
- A matrix is diagonalizable iff eigenvectors form a basis—this is the fundamental equivalence you need to internalize
- The theorem connects algebra to geometry—eigenvalues (algebraic) and eigenvectors (geometric) must align properly
- Non-diagonalizable matrices require Jordan form—a more complex canonical form for matrices with insufficient eigenvectors
- Matrices A and B are similar if A=PBP−1—they represent the same linear transformation in different bases
- Similar matrices share eigenvalues and characteristic polynomials—these are similarity invariants
- Diagonalization is similarity to a diagonal matrix—the "simplest" representative of the similarity class
Compare: Diagonalization vs. General Similarity—all diagonalizations are similarity transformations, but not all similar matrices are diagonal. The diagonal form is special because it makes computation trivial. When asked about matrix properties preserved under similarity, eigenvalues and trace are your go-to answers.
Computational Procedure
Exam problems frequently ask you to execute diagonalization step-by-step. Practice this algorithm until it's automatic.
Steps to Diagonalize a Matrix
- Step 1: Solve det(A−λI)=0 for eigenvalues—expand the determinant and find all roots of the characteristic polynomial
- Step 2: Find eigenvectors by solving (A−λI)v=0—for each eigenvalue, row-reduce and identify the null space basis
- Step 3: Construct P and D—columns of P are eigenvectors, diagonal entries of D are corresponding eigenvalues (in matching order!)
Real-World Applications
Understanding applications helps you contextualize abstract concepts and prepares you for applied problems.
Applications of Diagonalization
- Matrix powers become trivial—Ak=PDkP−1 reduces to exponentiating diagonal entries, crucial for Markov chains and recurrence relations
- Differential equations decouple—systems dtdx=Ax separate into independent equations when A is diagonalized
- Principal Component Analysis (PCA) relies on diagonalization—covariance matrix eigenvectors identify directions of maximum variance in data
Compare: Diagonalization in Pure vs. Applied Contexts—in theory problems, you verify conditions and execute the algorithm; in applications, you interpret what eigenvalues and eigenvectors mean (stability, principal components, growth rates). Expect exam questions to test both.
Quick Reference Table
|
| Diagonalization form | A=PDP−1 |
| Eigenvalue equation | Av=λv or (A−λI)v=0 |
| Characteristic polynomial | det(A−λI)=0 |
| Diagonalizability condition | Geometric multiplicity = Algebraic multiplicity for all λ |
| Sufficient condition | n distinct eigenvalues → always diagonalizable |
| Matrix powers | Ak=PDkP−1 |
| Similarity invariants | Eigenvalues, determinant, trace, characteristic polynomial |
| Column order in P | Must match eigenvalue order in D |
Self-Check Questions
-
If a 3×3 matrix has eigenvalues λ=2,2,5, what additional information do you need to determine if it's diagonalizable?
-
Compare and contrast algebraic multiplicity and geometric multiplicity. Why does their relationship determine diagonalizability?
-
Given A=PDP−1, explain why computing A50 is easier in this form than working with A directly.
-
Two matrices have the same eigenvalues. Are they necessarily similar? Are they necessarily equal? Justify your answer.
-
A matrix fails to diagonalize because one eigenvalue has geometric multiplicity 1 but algebraic multiplicity 2. What does this tell you about the eigenvector situation, and what alternative canonical form might you use?