Linear Quadratic Regulation and Riccati Intuition
LQR, Riccati equation, optimal control, state feedback, quadratic cost
1 Role
This is the fourth page of the Control and Dynamics module.
Its job is to show what happens when we stop choosing feedback gains only by pole locations and instead ask for the best feedback law relative to a precise objective.
That is the first-pass entry point into optimal control.
2 First-Pass Promise
Read this page after Feedback, Stability, and Pole Placement.
If you stop here, you should still understand:
- why LQR is a feedback-design problem with an optimization objective
- what the matrices
QandRare penalizing - why the optimal control law is linear state feedback
- why the Riccati equation is the object that packages the whole design tradeoff
3 Why It Matters
Pole placement is powerful, but it leaves an important question open:
which stable feedback gain should we prefer among the many possible ones?
LQR answers that by introducing a cost functional that balances two goals:
- keep the state small
- avoid using unnecessarily large control effort
So instead of saying only
move the poles somewhere stable
we say
choose the feedback law that minimizes a mathematically explicit performance objective
That makes LQR one of the cleanest bridges in the whole site:
- from dynamics to control
- from feedback to optimization
- from matrices to Riccati equations
4 Prerequisite Recall
- full-state feedback has the form
u=-Kx+r - the closed-loop matrix is
A-BK - stability by itself does not say how expensive the control signal is
- quadratic forms and positive semidefinite matrices are the natural language for energy-like penalties
5 Intuition
5.1 LQR Optimizes Over Feedback Behavior
For the linear system
\[ \dot x = Ax + Bu, \]
we choose a cost such as
\[ J(u) = \int_0^\infty \left(x(t)^TQx(t) + u(t)^TRu(t)\right)\,dt. \]
The matrix Q says which state directions we care about keeping small.
The matrix R says how expensive different control actions are.
5.2 The Best Input Law Ends Up Being Linear
A first-pass surprise is that for this linear system with quadratic cost, the optimal control law still has the form
\[ u = -Kx. \]
So the optimization problem does not lead to a complicated nonlinear controller here.
It leads back to state feedback, but now with a gain chosen by a principled tradeoff.
5.3 The Riccati Equation Encodes The Tradeoff
The gain is not picked directly.
Instead, we solve for a symmetric matrix P that measures the value of being at state x.
That matrix satisfies the algebraic Riccati equation, and then the gain comes out from P.
So the picture is:
cost choice -> Riccati equation -> matrix P -> gain K -> closed-loop dynamics
5.4 LQR Is More Structured Than Hand-Tuned Pole Placement
Pole placement says where the modes should go.
LQR says how much state error and control effort are worth relative to each other.
The resulting poles are not chosen by hand. They are the consequence of solving the optimization problem.
6 Formal Core
Definition 1 (Definition: Continuous-Time Infinite-Horizon LQR) Consider the linear system
\[ \dot x = Ax + Bu. \]
The continuous-time infinite-horizon LQR problem is to choose u(t) to minimize
\[ J(u)=\int_0^\infty \left(x(t)^TQx(t)+u(t)^TRu(t)\right)\,dt, \]
where Q \succeq 0 and R \succ 0.
Definition 2 (Definition: State And Control Penalties) At a first pass:
Qpenalizes state deviationRpenalizes control effort
Larger weights in Q push the controller to suppress those state directions more strongly.
Larger weights in R make aggressive control more expensive.
Theorem 1 (Theorem Idea: Optimal Control Is Linear State Feedback) Under the standard stabilizability and detectability conditions, the optimal control law for infinite-horizon continuous-time LQR has the form
\[ u^\ast = -Kx, \]
with
\[ K = R^{-1}B^TP, \]
where P is a symmetric positive semidefinite solution of the algebraic Riccati equation.
Definition 3 (Definition: Continuous-Time Algebraic Riccati Equation) The matrix P satisfies
\[ A^TP + PA - PBR^{-1}B^TP + Q = 0. \]
Theorem 2 (Theorem Idea: The Riccati Solution Produces A Stabilizing Closed Loop) Under the standard first-pass assumptions, the gain
\[ K = R^{-1}B^TP \]
produces a stabilizing closed-loop matrix
\[ A-BK. \]
This is why LQR is not only an optimization story. It is also a stability-design story.
7 Worked Example
Consider the scalar system
\[ \dot x = ax + bu, \]
with cost
\[ J(u)=\int_0^\infty \left(x(t)^2 + \rho\,u(t)^2\right)\,dt, \]
where \rho>0.
Here
\[ Q=1, \qquad R=\rho. \]
The Riccati equation becomes
\[ 2aP - \frac{b^2}{\rho}P^2 + 1 = 0. \]
So instead of searching directly over all possible input signals, we solve a scalar quadratic equation for P.
Once we obtain the stabilizing solution P, the optimal feedback gain is
\[ K = \frac{b}{\rho}P. \]
Then the closed-loop system is
\[ \dot x = (a-bK)x. \]
This example shows the core control pattern:
- write the objective
- derive the Riccati equation
- solve for
P - recover
K - read the resulting closed-loop behavior
8 Computation Lens
When you see an LQR problem, ask:
- what are the actual state directions we care about, and how are they weighted in
Q? - what control actions are expensive, and how is that encoded in
R? - are we in continuous time or discrete time?
- is the system stabilizable, so that the Riccati solution can produce a meaningful stabilizing gain?
- are we solving one regulator problem around one operating point, or trying to control a genuinely nonlinear system far from that point?
These questions usually matter more than memorizing the Riccati equation by itself.
9 Application Lens
9.1 Balancing State Error And Control Effort
LQR is one of the cleanest ways to formalize the engineering idea that good control should be accurate without being unnecessarily aggressive.
9.2 Bridge To Estimation And Kalman Filtering
The next natural page is not just another feedback page.
It is the estimator side of the same systems story:
Kalman filtering introduces a closely related Riccati equation on the observation side.
9.3 Bridge To MPC And Modern Control
Finite-horizon optimal control, model predictive control, and many learning-based control methods all look easier once LQR is in place as the first exact benchmark.
10 Stop Here For First Pass
If you can now explain:
- why LQR is more structured than hand-chosen pole placement
- what
QandRmean - why the optimal law is still linear feedback
- why the Riccati equation is the central matrix object
- how stabilization and optimization are joined in the same design
then this page has done its job.
11 Go Deeper
The next natural step in this module is:
The strongest adjacent live pages right now are:
12 Optional Deeper Reading After First Pass
The strongest current references connected to this page are:
- MIT 16.30 Topic 18: Linear Quadratic Regulator - official lecture notes for the continuous-time LQR problem and Riccati viewpoint. Checked
2026-04-25. - MIT 16.323: Principles of Optimal Control lecture notes - official notes index for dynamic programming, continuous/discrete LQR, and optimal-control structure. Checked
2026-04-25. - MIT 16.323 syllabus - official syllabus showing where LQR sits in a broader optimal-control course. Checked
2026-04-25. - Stanford EE363 bulletin - official current course description connecting state-space models, LQR, and Kalman filtering. Checked
2026-04-25. - Stanford AA203 bulletin - official archived course description for optimal and learning-based control. Checked
2026-04-25.
13 Sources and Further Reading
- MIT 16.30 Topic 18: Linear Quadratic Regulator -
First pass- official notes for the finite/infinite-horizon LQR setup and the Riccati viewpoint. Checked2026-04-25. - MIT 16.323: Principles of Optimal Control lecture notes -
Second pass- official notes index for discrete LQR, continuous LQR, and dynamic-programming viewpoints. Checked2026-04-25. - MIT 16.323 syllabus -
Second pass- official syllabus for the broader optimal-control context around LQR. Checked2026-04-25. - Stanford EE363 bulletin -
Second pass- official current description connecting regulator design and estimation in one course arc. Checked2026-04-25. - Stanford AA203 bulletin -
Second pass- official archived description connecting optimal control with modern learning-based directions. Checked2026-04-25.