Approximation, Differentiation, Integration, and Error Control
interpolation, numerical differentiation, quadrature, truncation error, error control
1 Role
This is the sixth page of the Numerical Methods module.
Its job is to show that interpolation, numerical differentiation, and numerical integration all share the same numerical pattern:
replace a hard continuous object by an easier approximation, then manage the resulting error
2 First-Pass Promise
Read this page after Eigenvalue and SVD Computation.
If you stop here, you should still understand:
- why interpolation, finite differences, and quadrature are all approximation schemes
- why truncation error is different from roundoff error
- why making the step size smaller does not always improve the answer
- why adaptive error control is really a strategy for spending computation where it matters most
3 Why It Matters
A huge fraction of numerical work is not about solving a linear system exactly.
It is about replacing an intractable continuous object by a simpler discrete one:
- replace a function by an interpolating polynomial
- replace a derivative by a finite difference
- replace an integral by a weighted sum
This always creates approximation error.
So the key questions become:
- what approximation family should we use?
- how quickly does the error shrink as the mesh or step size changes?
- when does roundoff start to compete with truncation error?
- how do we choose computational effort intelligently?
This is the core error-control viewpoint behind practical scientific computing.
4 Prerequisite Recall
- Taylor expansion gives local polynomial approximations and remainder intuition
- floating-point arithmetic introduces roundoff at scale
\varepsilon_{\mathrm{mach}} - earlier pages in this module separated conditioning from algorithmic stability
5 Intuition
5.1 Interpolation
Interpolation replaces a complicated function by an easier surrogate that matches the function at selected points.
The hope is that:
- the surrogate is cheap to evaluate or integrate
- the mismatch between surrogate and true function is controlled
5.2 Numerical Differentiation
Differentiation is sensitive because it magnifies small perturbations.
Finite-difference formulas use nearby function values to approximate a derivative, but they mix two errors:
- truncation error from replacing the derivative by a formula of finite order
- roundoff error from subtracting nearly equal numbers
5.3 Numerical Integration
Quadrature replaces an integral by a weighted sum of function values.
The approximation improves when the rule matches the local function behavior well, but it still depends on step size, smoothness, and error accumulation.
5.4 Why Error Control Matters
There is almost always a tradeoff:
- coarse discretization gives large truncation error
- excessively tiny steps can amplify roundoff or waste work
Good numerical computation is about choosing a scale where the total error is acceptable for the cost.
6 Formal Core
Definition 1 (Definition: Interpolation) Interpolation chooses an approximating function p(x) so that
\[ p(x_i)=f(x_i) \]
at selected nodes x_i.
The approximation problem is then to understand how well p represents f away from the nodes.
Definition 2 (Definition: Finite-Difference Approximation) A basic forward-difference approximation to f'(x) is
\[ \frac{f(x+h)-f(x)}{h}. \]
By Taylor expansion, this has first-order truncation error in h when f is sufficiently smooth.
Definition 3 (Definition: Quadrature Rule) A quadrature rule approximates an integral by a weighted sum
\[ \int_a^b f(x)\,dx \approx \sum_{j=1}^m w_j f(x_j). \]
The nodes x_j and weights w_j encode the approximation scheme.
Theorem 1 (Theorem Idea: Total Numerical Error Often Splits Into Competing Pieces) In many approximation schemes, total error can be understood as a combination of:
- truncation or discretization error
- roundoff error
The useful step size is often the one that balances those contributions rather than minimizing either one alone.
Theorem 2 (Theorem Idea: Adaptive Error Control Uses Local Difficulty) Adaptive methods refine the grid or step size more where the function or solution is hard to approximate and less where it is easy.
This is why adaptive methods often outperform uniform refinement at the same cost.
7 A Small Worked Example
Take
\[ f(x)=e^x \]
and approximate f'(0) using the forward difference:
\[ \frac{f(h)-f(0)}{h} = \frac{e^h-1}{h}. \]
Using Taylor expansion,
\[ e^h = 1+h+\frac{h^2}{2}+O(h^3), \]
so
\[ \frac{e^h-1}{h} = 1+\frac{h}{2}+O(h^2). \]
The exact derivative is f'(0)=1, so the truncation error is about h/2 for small h.
This suggests smaller h is better.
But on a machine, if h becomes extremely small, the subtraction
\[ e^h-1 \]
can lose relative accuracy through cancellation, and roundoff begins to matter.
So the best practical h is not “as small as possible.” It is the scale where truncation and roundoff are reasonably balanced.
8 Computation Lens
When you face an approximation scheme, ask:
- what continuous object am I replacing?
- what is the leading truncation or discretization error term?
- how does roundoff enter the formula?
- should I refine uniformly, or would adaptive refinement spend work more intelligently?
Those questions usually matter more than memorizing a menu of formulas.
9 Application Lens
9.1 Scientific Computing
Interpolation, differentiation, and quadrature are the building blocks behind simulation, data fitting, and time-stepping pipelines.
9.2 Optimization
Finite-difference derivatives, local polynomial models, and error-controlled line-search or trust-region subroutines all rely on approximation language.
9.3 Engineering Systems
Practical computation in mechanics, controls, and signal-processing pipelines depends on deciding how much approximation error is acceptable and where to refine.
10 Stop Here For First Pass
If you can now explain:
- why these topics are all approximation problems
- why truncation error and roundoff error are different
- why step size should not be shrunk blindly
- why adaptive error control is about allocating computational effort intelligently
then this page has done its job.
11 Go Deeper
After this page, the next natural extension is:
The strongest adjacent pages are:
12 Optional Deeper Reading After First Pass
The strongest current references connected to this page are:
- Stanford CME108 syllabus pdf - official syllabus covering interpolation, quadrature, and adaptive numerical thinking in one scientific-computing track. Checked
2026-04-25. - Stanford CS137 syllabus - official current scientific-computing syllabus with interpolation, numerical differentiation, and composite quadrature. Checked
2026-04-25. - MIT 2.086 differentiation pdf - official MIT notes for finite-difference approximation and its error analysis. Checked
2026-04-25. - MIT 10.34 numerical integration lecture - official MIT lecture notes for quadrature from interpolation and error viewpoints. Checked
2026-04-25. - Cornell CS4210 course page - official Cornell course page for the complementary numerical-analysis half focused on error analysis, approximation, interpolation, and integration. Checked
2026-04-25. - Stanford CME108 bulletin - official course description linking these approximation topics to the broader scientific-computing curriculum. Checked
2026-04-25.
13 Sources and Further Reading
- Stanford CME108 syllabus pdf -
First pass- official syllabus covering interpolation, differentiation, quadrature, and approximation-aware computation. Checked2026-04-25. - Stanford CS137 syllabus -
First pass- official current scientific-computing syllabus with interpolation, numerical differentiation, and composite quadrature. Checked2026-04-25. - MIT 2.086 differentiation pdf -
First pass- official MIT notes for finite-difference approximation and error analysis. Checked2026-04-25. - MIT 10.34 numerical integration lecture -
First pass- official MIT lecture notes for interpolation-based quadrature and integration error. Checked2026-04-25. - Cornell CS4210 course page -
Second pass- official Cornell course page for the approximation/interpolation/integration half of the numerical-analysis sequence. Checked2026-04-25. - Stanford CME108 bulletin -
Second pass- official course description placing these tools inside a broader scientific-computing sequence. Checked2026-04-25.