Computation Lab: Projection Geometry and Regression Residuals
An interactive lab for seeing how slope, intercept, residuals, and projection geometry fit together in least squares.
Keywords
computation, simulation, visualization, least squares
1 Lab Goal
This lab helps you see one specific fact:
least squares is the place where the residual becomes orthogonal to the model directions.
2 Math Question
For the three-point regression example, how do the slope and intercept affect:
- the fitted line
- the residual vector
- the sum of squared residuals
- the orthogonality checks behind the normal equations
3 Model or Setup
We use the line-fitting problem from the concept page:
\[ y \approx \beta_0 + \beta_1 t \]
for the points (0,1), (1,2), and (2,2).
The least-squares solution is
\[ \hat{\beta}_0 = \frac{7}{6}, \qquad \hat{\beta}_1 = \frac{1}{2}. \]
4 Parameters and Controls
Intercept: moves the line up and downSlope: tilts the line
The dashed line is the current choice from the sliders. The solid line is the true least-squares fit.
5 Code and Simulation
6 What To Observe
- Start by moving the sliders until the dashed line sits on top of the solid least-squares line.
- When the sliders are at the least-squares values, the
SSEis smallest. - At the same point, both orthogonality checks are approximately zero.
- Moving away from the least-squares line makes the red residual segments longer and breaks at least one orthogonality check.
- The residual is not trying to be “small componentwise.” It is trying to be orthogonal to the model directions.
7 Interpretation
This lab is showing the normal equations in geometric form.
The intercept column and the slope column span the model subspace. The least-squares fit is the unique point in that subspace where the residual lands in the orthogonal complement.
So the table and plot are both visual proofs of the same fact:
\[ X^\top (y - X\hat{\beta}) = 0. \]
8 Failure Modes and Numerical Cautions
- This is a tiny deterministic example, so it hides conditioning issues.
- On larger problems, explicitly forming \(X^\top X\) can be numerically worse than using
QRorSVD. - A line fit can look visually reasonable even when the modeling assumptions are poor.
- Squared loss is sensitive to outliers, so projection geometry does not automatically imply robust modeling.
9 Reproducibility Notes
- execution engine:
Observable JS - no randomness and no seed required
- built-in Quarto client-side libraries:
Inputs,Plot,md - static-site friendly: no server or notebook kernel required after render
10 Extensions
- replace the line fit with a two-feature design matrix and watch the residual stay orthogonal to each feature column
- compare the least-squares fit with a deliberately collinear feature matrix
- connect this lab to SVD and Low-Rank Approximation by asking how
QRandSVDcompute the same fit more stably
11 Sources and Further Reading
- MIT 18.06SC Linear Algebra resource index -
First pass- good official support for projection-based least-squares intuition. Checked2026-04-24. - MIT 2.086 Unit 3 notes -
Second pass- useful for turning the geometric picture into numerical workflow. Checked2026-04-24. - A Statistical Perspective on Randomized Sketching for Ordinary Least-Squares -
Paper bridge- shows that even after compression, the geometry of the least-squares problem is still the core object. Checked2026-04-24.