Constraints, MPC, and Safe Operation

A bridge page showing why real systems must respect hard limits, how model predictive control handles them naturally, and why safe operation is often an optimization problem in the loop.
Modified

April 26, 2026

Keywords

constraints, MPC, safety, receding horizon, control

1 Application Snapshot

Many real systems are not limited by math elegance. They are limited by physics and safety:

  • motors saturate
  • batteries drain
  • voltages and torques have hard bounds
  • temperatures, pressures, and positions must stay inside safe regions

So a controller is often not allowed to ask only:

what control action would be best if nothing were limited?

It must ask:

what control action is best among the ones that keep the system feasible and safe?

That is why constrained control and MPC appear so often in robotics, vehicles, energy systems, and process control.

2 Problem Setting

Suppose the system evolves as

\[ x_{t+1} = f(x_t, u_t), \]

and we want to minimize a finite-horizon cost while respecting constraints:

\[ x_t \in X, \qquad u_t \in U. \]

Here:

  • \(X\) is the safe or allowable state set
  • \(U\) is the allowable input set

At each time step, the controller solves a planning problem over a horizon:

\[ u_0, u_1, \dots, u_{N-1} \]

but applies only the first control move, then replans after the next state or estimate arrives.

That repeated solve-act-resolve loop is the application-level meaning of model predictive control.

3 Why This Math Appears

This language reuses several math layers already on the site:

  • Optimization: constraints and cost must be handled together, not as separate afterthoughts
  • Numerical Methods: the optimization has to be solved fast enough to run inside a real loop
  • Control and Dynamics: the candidate control sequence is judged through predicted future dynamics
  • Estimation: when state is uncertain, the constrained decision may be based on an estimate rather than a direct measurement
  • Stochastic Control and Dynamic Programming: repeated replanning is a sequential decision process, not a one-shot solve

So safe operation is not a cosmetic wrapper around control. It is often the reason the control problem must be formulated as constrained optimization in the first place.

4 Math Objects In Use

  • state \(x_t\)
  • input \(u_t\)
  • prediction model \(f\)
  • state constraint set \(X\)
  • input constraint set \(U\)
  • horizon length \(N\)
  • running and terminal costs
  • feasible plan over the horizon

The application picture is:

  1. predict what future states would happen under a candidate control sequence
  2. discard plans that violate safety or actuator limits
  3. choose the best remaining feasible plan
  4. apply only the first action
  5. replan at the next step

5 A Small Worked Walkthrough

Imagine an autonomous car approaching a slower vehicle ahead.

The system may want to:

  • maintain speed
  • avoid hard braking
  • preserve a safe following distance

An unconstrained planner might propose a control sequence that reduces tracking error quickly but violates:

  • acceleration limits
  • jerk limits
  • minimum-distance safety bounds

A constrained MPC-style planner instead solves:

  • minimize tracking and effort cost
  • subject to the car dynamics
  • subject to speed, acceleration, and spacing constraints

The application point is not only that MPC “optimizes better.”

It is that the optimization explicitly excludes dangerous futures before control is chosen.

That is why safe operation often looks like prediction plus feasibility, not just gain tuning.

6 Implementation or Computation Note

Three practical design questions appear immediately:

  1. Feasibility What happens if the constraints and reference are temporarily incompatible?

  2. Model accuracy How much safety margin is needed if the predictive model is imperfect?

  3. Computation time Can the constrained problem be solved fast enough at the controller’s update rate?

Use these pages as the strongest follow-on support:

7 Failure Modes

  • designing an elegant unconstrained controller that cannot respect actuator limits
  • assuming safety constraints can be “patched on later” outside the control design
  • ignoring the fact that the optimization must finish within the control update window
  • treating a predicted feasible trajectory as safe even when the state estimate is poor
  • forgetting that repeated replanning can fail if the problem becomes infeasible

8 Paper Bridge

9 Sources and Further Reading

  • 16.323 / Model Predictive Control - First pass - official MIT MPC lecture anchor. Checked 2026-04-25.
  • 16.323 lecture notes - First pass - official MIT optimal-control notes that frame MPC in the larger planning story. Checked 2026-04-25.
  • EE364b course information - Second pass - official Stanford convex optimization course info with strong constrained-control relevance. Checked 2026-04-25.
  • EE364b lectures - Second pass - useful when you want the optimization layer under constrained control to be more explicit. Checked 2026-04-25.
  • EE364b MPC slides - Second pass - direct Stanford slide deck for MPC language and structure. Checked 2026-04-25.
  • AA203 / Optimal and Learning-Based Control - Bridge outward - official Stanford course entry where constrained and optimal control ideas scale outward into broader control design. Checked 2026-04-25.
Back to top