Fiveable

🎛️Control Theory Unit 2 Review

QR code for Control Theory practice questions

2.5 State-space representation

2.5 State-space representation

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🎛️Control Theory
Unit & Topic Study Guides

State-space representation gives you a way to model dynamic systems using a set of first-order differential equations instead of a single higher-order transfer function. This approach is especially valuable for multi-input, multi-output (MIMO) systems and opens the door to modern control techniques like optimal control and state estimation.

Where transfer functions only show you the input-output relationship, state-space models reveal what's happening inside the system. The matrices, equations, and properties covered here form the foundation for designing controllers and observers in later units.

State-space models

State-space models describe system dynamics using a set of first-order differential equations (continuous-time) or difference equations (discrete-time), all written in matrix form. This gives you a compact, systematic way to represent even complex MIMO systems.

Difference between state-space and transfer function models

FeatureState-SpaceTransfer Function
System descriptionInternal state variablesInput-output relationship
MIMO handlingDirect, naturalRequires a matrix of transfer functions
Information providedInternal dynamics + external behaviorExternal behavior only
BasisTime domain (set of first-order ODEs)Frequency domain (polynomials in ss)
Transfer functions are great for SISO analysis, but state-space models give you a fuller picture. You can always convert between the two for linear time-invariant (LTI) systems.

Advantages of state-space representation

  • Handles MIMO systems in a single, unified framework
  • Exposes internal structure, not just input-output behavior
  • Enables state feedback controller and observer design
  • Lets you check system properties like controllability and observability directly
  • Supports modern control methods: optimal control (LQR), robust control, Kalman filtering

State variables

State variables are the smallest set of variables that fully describe a system's condition at any moment. They carry the system's "memory," capturing how past inputs have shaped the current state.

Definition of state variables

If you know the state variables at some time t0t_0 and the input u(t)u(t) for all tt0t \geq t_0, you can determine the future behavior of the system completely. That's what makes them the minimum set needed to characterize the system.

Selection of state variables

The choice of state variables isn't unique. For a mechanical system, you might pick position and velocity. For an electrical circuit, capacitor voltages and inductor currents are natural choices. You could also use more abstract mathematical quantities like error signals or their integrals.

What matters is that your selection:

  • Uses the minimum number of variables (equal to the system order)
  • Leads to a representation that is controllable and observable (when possible)

State-space equations

The state-space description consists of two equations: the state equation (how internal states evolve) and the output equation (what you can measure).

General form of state-space equations

State equation (describes the dynamics):

x˙(t)=Ax(t)+Bu(t)\dot{x}(t) = Ax(t) + Bu(t)

  • x(t)x(t): state vector (n×1n \times 1)
  • u(t)u(t): input vector (m×1m \times 1)
  • AA: state matrix (n×nn \times n)
  • BB: input matrix (n×mn \times m)

Output equation (relates states to measurements):

y(t)=Cx(t)+Du(t)y(t) = Cx(t) + Du(t)

  • y(t)y(t): output vector (p×1p \times 1)
  • CC: output matrix (p×np \times n)
  • DD: feedthrough matrix (p×mp \times m)

Together, these four matrices (A,B,C,D)(A, B, C, D) completely define the LTI system.

State equation vs output equation

The state equation governs how the internal states change over time. It captures the system's dynamics. The output equation is purely algebraic: it tells you what combination of states and inputs produces the measured output at each instant. Think of the state equation as the "engine" and the output equation as the "dashboard."

Linearization of nonlinear systems

Most real systems are nonlinear, but state-space tools are built for linear systems. To bridge this gap, you linearize around an operating point (also called an equilibrium point):

  1. Define the nonlinear system: x˙=f(x,u)\dot{x} = f(x, u), y=g(x,u)y = g(x, u)
  2. Find an equilibrium point (xe,ue)(x_e, u_e) where f(xe,ue)=0f(x_e, u_e) = 0
  3. Compute the Jacobian matrices (partial derivatives) of ff and gg evaluated at (xe,ue)(x_e, u_e)
  4. The resulting matrices become your A,B,C,DA, B, C, D for the linearized model

The linearized model is only accurate near the operating point. If the system moves far from that point, the approximation breaks down.

State-space matrices

State matrix A

The AA matrix (n×nn \times n) describes how the states interact with each other in the absence of any input. Its eigenvalues directly determine system stability and natural response:

  • All eigenvalues with negative real parts → system is stable
  • Any eigenvalue with a positive real part → system is unstable
  • Eigenvalues with zero real parts → marginally stable (needs further analysis)

Input matrix B

The BB matrix (n×mn \times m) maps external inputs to the state derivatives. Each column of BB shows how one particular input influences all the state variables. If a column is all zeros, that input has no direct effect on the states.

Output matrix C

The CC matrix (p×np \times n) determines which state variables (or combinations of them) appear in the output. Each row of CC corresponds to one output measurement.

Feedthrough matrix D

The DD matrix (p×mp \times m) represents any direct path from input to output that bypasses the state dynamics entirely. In many physical systems, D=0D = 0 because inputs must propagate through the system's dynamics before affecting the output. A nonzero DD means the output responds instantaneously to input changes.

Controllability

Controllability tells you whether you can drive the system from any initial state to any desired state in finite time using the available inputs. If a system isn't fully controllable, there are internal modes you simply cannot influence, no matter what input you apply.

Difference between state-space and transfer function models, control theory - Derivation for state equation linearization - Mathematics Stack Exchange

Definition of controllability

A system is controllable if, for any initial state x(t0)x(t_0) and any desired final state x(tf)x(t_f), there exists an input u(t)u(t) that transfers the system from x(t0)x(t_0) to x(tf)x(t_f) in finite time.

Controllability matrix

For an LTI system with nn states, the controllability matrix is:

C=[BABA2BAn1B]\mathcal{C} = [B \quad AB \quad A^2B \quad \cdots \quad A^{n-1}B]

This matrix has dimensions n×nmn \times nm (for mm inputs).

Controllability tests

  • Rank test: The system is controllable if and only if rank(C)=n\text{rank}(\mathcal{C}) = n (full row rank).
  • PBH test: The system is controllable if and only if rank([AλIB])=n\text{rank}([A - \lambda I \quad B]) = n for every eigenvalue λ\lambda of AA. This test is particularly useful for identifying which modes are uncontrollable.

Controllable canonical form

The controllable canonical form is a specific arrangement of the state-space matrices where the controllability structure is immediately visible. In this form, AA takes a companion matrix structure and BB has a single nonzero entry. This form is especially useful for designing state feedback controllers via pole placement, since the feedback gains map directly to the desired characteristic polynomial coefficients.

Observability

Observability is the dual of controllability. Instead of asking "can you control every state?", it asks "can you determine every state from the outputs?"

Definition of observability

A system is observable if the initial state x(t0)x(t_0) can be uniquely determined from knowledge of the input u(t)u(t) and output y(t)y(t) over a finite time interval.

Observability matrix

For an LTI system with nn states, the observability matrix is:

O=[CCACA2CAn1]\mathcal{O} = \begin{bmatrix} C \\ CA \\ CA^2 \\ \vdots \\ CA^{n-1} \end{bmatrix}

This matrix has dimensions np×nnp \times n (for pp outputs).

Observability tests

  • Rank test: The system is observable if and only if rank(O)=n\text{rank}(\mathcal{O}) = n (full column rank).
  • PBH test: The system is observable if and only if rank[AλIC]=n\text{rank}\begin{bmatrix} A - \lambda I \\ C \end{bmatrix} = n for every eigenvalue λ\lambda of AA.

Observable canonical form

The observable canonical form is the dual of the controllable canonical form. Here, AA and CC are structured so that the observability properties are explicit. This form is the natural starting point for designing state observers.

Duality between controllability and observability: A system (A,B,C,D)(A, B, C, D) is controllable if and only if the dual system (AT,CT,BT,DT)(A^T, C^T, B^T, D^T) is observable. This means any result you prove for controllability has a corresponding observability result, and vice versa.

Transformations of state-space models

Different choices of state variables lead to different (A,B,C,D)(A, B, C, D) matrices, but the underlying system is the same. Transformations let you switch between representations while preserving the input-output behavior.

Similarity transformations

Given a nonsingular matrix TT, define a new state vector x~(t)=Tx(t)\tilde{x}(t) = Tx(t). The transformed matrices are:

A~=TAT1,B~=TB,C~=CT1,D~=D\tilde{A} = TAT^{-1}, \quad \tilde{B} = TB, \quad \tilde{C} = CT^{-1}, \quad \tilde{D} = D

Key properties preserved under similarity transformations:

  • Eigenvalues of AA (and therefore stability)
  • Transfer function G(s)G(s)
  • Controllability and observability

Coordinate transformations

Coordinate transformations are similarity transformations chosen to simplify the representation. You might rotate, scale, or translate the state variables to decouple the dynamics or align states with physically meaningful quantities. For example, diagonalizing AA (when possible) decouples the state equations so each state evolves independently.

Canonical forms

Canonical forms are standardized representations with specific structural properties:

  • Controllable canonical form: highlights controllability, useful for pole placement
  • Observable canonical form: highlights observability, useful for observer design
  • Jordan canonical form: diagonalizes AA (or puts it in Jordan blocks for repeated eigenvalues), revealing the system's natural modes directly

Solution of state-space equations

State transition matrix

The state transition matrix Φ(t,t0)\Phi(t, t_0) maps the state at time t0t_0 to the state at time tt for the unforced (zero-input) system. It satisfies:

Φ˙(t,t0)=AΦ(t,t0),Φ(t0,t0)=I\dot{\Phi}(t, t_0) = A\Phi(t, t_0), \quad \Phi(t_0, t_0) = I

The complete solution (with input) is:

x(t)=Φ(t,t0)x(t0)+t0tΦ(t,τ)Bu(τ)dτx(t) = \Phi(t, t_0)x(t_0) + \int_{t_0}^{t} \Phi(t, \tau)Bu(\tau) \, d\tau

The first term is the zero-input response (due to initial conditions), and the integral is the zero-state response (due to the input).

Difference between state-space and transfer function models, Frontiers | Wavelet Fuzzy Brain Emotional Learning Control System Design for MIMO Uncertain ...

Matrix exponential

For LTI systems, the state transition matrix equals the matrix exponential:

Φ(t,t0)=eA(tt0)\Phi(t, t_0) = e^{A(t - t_0)}

The matrix exponential is defined by the power series:

eAt=I+At+(At)22!+(At)33!+e^{At} = I + At + \frac{(At)^2}{2!} + \frac{(At)^3}{3!} + \cdots

Computing eAte^{At} directly from the series is rarely practical. Common methods include:

  • Diagonalization: if A=PDP1A = PDP^{-1}, then eAt=PeDtP1e^{At} = Pe^{Dt}P^{-1}
  • Laplace transform: eAt=L1{(sIA)1}e^{At} = \mathcal{L}^{-1}\{(sI - A)^{-1}\}
  • Cayley-Hamilton theorem

Laplace transform approach

Taking the Laplace transform of x˙(t)=Ax(t)+Bu(t)\dot{x}(t) = Ax(t) + Bu(t):

  1. Transform: sX(s)x(0)=AX(s)+BU(s)sX(s) - x(0) = AX(s) + BU(s)

  2. Rearrange: (sIA)X(s)=x(0)+BU(s)(sI - A)X(s) = x(0) + BU(s)

  3. Solve: X(s)=(sIA)1x(0)+(sIA)1BU(s)X(s) = (sI - A)^{-1}x(0) + (sI - A)^{-1}BU(s)

  4. Inverse Laplace transform to get x(t)x(t)

This approach converts differential equations into algebraic equations, which is often easier to handle.

Numerical methods for simulation

When analytical solutions are impractical (large systems, time-varying parameters, nonlinearities), numerical methods approximate the solution by stepping through time:

  • Euler method: simplest, but can be inaccurate for stiff systems
  • Runge-Kutta methods (e.g., RK4): good balance of accuracy and computational cost
  • Dormand-Prince: adaptive step-size variant of Runge-Kutta, used by MATLAB's ode45

These methods discretize the continuous equations and compute x(tk+1)x(t_{k+1}) from x(tk)x(t_k) iteratively.

Steady-state response

Once transients decay, the system settles into its steady-state behavior. Understanding this long-term behavior is critical for evaluating system performance.

Equilibrium points

An equilibrium point is a state where the system stays put if undisturbed. You find it by setting x˙=0\dot{x} = 0:

0=Axe+Bue0 = Ax_e + Bu_e

For a given constant input ueu_e, the equilibrium state is xe=A1Buex_e = -A^{-1}Bu_e (assuming AA is invertible). If AA is singular, equilibrium points may not exist or may not be unique.

Stability of equilibrium points

Stability determines whether the system returns to equilibrium after a small perturbation:

  • Asymptotically stable: all eigenvalues of AA have strictly negative real parts. The system returns to equilibrium.
  • Marginally stable: eigenvalues on the imaginary axis (with no repeated eigenvalues there), none with positive real parts. The system neither grows nor decays.
  • Unstable: at least one eigenvalue has a positive real part. Perturbations grow over time.

For nonlinear systems, Lyapunov stability theory provides tools beyond eigenvalue analysis.

Steady-state error

Steady-state error is the persistent difference between the desired output and the actual output after transients have died out. You can analyze it using:

  • The final value theorem: ess=lims0sE(s)e_{ss} = \lim_{s \to 0} sE(s), where E(s)E(s) is the error in the Laplace domain
  • The steady-state gain of the system, which depends on the DC gain G(0)=CA1B+DG(0) = -CA^{-1}B + D

The system type (number of integrators) determines how well it tracks step, ramp, and parabolic inputs.

Relationship between state-space and transfer function models

Obtaining transfer functions from state-space models

Starting from the state-space equations, take the Laplace transform and eliminate the state vector to get:

G(s)=C(sIA)1B+DG(s) = C(sI - A)^{-1}B + D

This is the transfer function matrix. For a SISO system, G(s)G(s) is a single rational function of ss. For MIMO systems, it's a matrix where each entry Gij(s)G_{ij}(s) relates the jj-th input to the ii-th output.

Minimal realization of transfer functions

A given transfer function can be realized by infinitely many state-space models. A minimal realization uses the fewest possible state variables and is both controllable and observable. If your realization has uncontrollable or unobservable modes, those correspond to pole-zero cancellations in the transfer function that hide internal dynamics.

Techniques for finding minimal realizations include Gilbert's method and Kalman decomposition, which separates the state space into controllable/observable, controllable/unobservable, uncontrollable/observable, and uncontrollable/unobservable parts.

Applications of state-space representation

Control system design

State feedback controllers use the form u(t)=Kx(t)+r(t)u(t) = -Kx(t) + r(t), where KK is a gain matrix chosen to place the closed-loop eigenvalues of (ABK)(A - BK) at desired locations. This is pole placement. For optimal performance, the LQR (Linear Quadratic Regulator) computes KK by minimizing a cost function that balances state deviations against control effort.

Observer design

When you can't measure all state variables directly, an observer estimates them. The Luenberger observer uses the model plus output error correction:

x^˙(t)=Ax^(t)+Bu(t)+L(y(t)Cx^(t))\dot{\hat{x}}(t) = A\hat{x}(t) + Bu(t) + L(y(t) - C\hat{x}(t))

The observer gain LL is chosen so that the estimation error decays quickly. The eigenvalues of (ALC)(A - LC) determine how fast the observer converges. Observability is required for this design to work.

Kalman filtering

The Kalman filter is the optimal observer for systems with Gaussian process and measurement noise. It recursively updates the state estimate by combining the model prediction with noisy measurements, minimizing the mean square estimation error. Applications include GPS navigation, robot localization, and sensor fusion.

Optimal control

Optimal control finds the input u(t)u(t) that minimizes a cost function, typically of the form:

J=0[xTQx+uTRu]dtJ = \int_0^\infty [x^T Q x + u^T R u] \, dt

where QQ penalizes state deviations and RR penalizes control effort. The solution for LTI systems is the LQR, which yields a constant state feedback gain. Combining LQR with a Kalman filter gives LQG (Linear Quadratic Gaussian) control, a powerful framework for handling both performance optimization and noise.