- 2025年10月8日
Statistics as a Paradigm Shift: How We Handle Truth, Error, Analog & Digital, and the Bridge to Analysis
Statistics as a Paradigm Shift: How We Handle Truth, Error, Analog & Digital, and the Bridge to Analysis
1) The Paradigm Shift that Statistics Introduced
Nineteenth-century science is often portrayed as deterministic, but twentieth-century statistics reshaped the rules of scientific argument. Once we accept that we only ever see samples, the central question becomes: What can we responsibly claim about the population under uncertainty?
Modern statistics answers this with estimation (intervals and uncertainty), testing (Type I/II errors and power), experimental design (especially randomization), and—more recently—causal inference for both trials and observational data. In this regime, error is not a nuisance to be ignored; it is the precondition for making disciplined claims. For continuous distributions, probability at an exact point is 0; what carries meaning is interval probability via a density. That simple convention is one of the quiet revolutions that makes the whole framework work.
2) Measurement and Error: Accuracy, Precision, Validity, Reliability
To reason clearly about data and devices, we separate four ideas:
- Accuracy: small systematic error (low bias from the true value).
- Precision: small random error (high repeatability/low spread).
- Validity: the instrument or procedure truly measures the intended construct.
- Reliability: consistency of measurement (it gives similar results under similar conditions).
Engineering translates these into designable targets. Instead of imagining “perfect parts,” we specify tolerances (USL/LSL), monitor processes with statistical process control (SPC), and rate capability with Cp/Cpk. We also use robust design and design of experiments (DOE) to make performance insensitive to noise. The practical lesson is simple but profound: in complex systems, quality comes less from chasing a metaphysical “true value” and more from stably landing within spec, run after run. In short: aim for low bias and low variance, and design the system so that residual variance is absorbed by the specification, not amplified by the assembly.
3) Analysis’s “Arbitrarily Fine” Promise
Analysis complements statistics by offering a principled way to approach exactness procedurally. The ε–δ viewpoint says: if you request an error no larger than ε, we can produce a computation whose deviation stays below ε. We rarely “close-form everything,” but we can approximate with guaranteed error control.
Two classic tools make this concrete:
- Polynomial/series approximations (Taylor if the function is analytic at the point; more generally, by Weierstrass, continuous functions on a closed interval can be uniformly approximated by polynomials). Crucially, remainder estimates bound the truncation error, so you can pick approximation length to meet a target tolerance.
- Fourier methods approximate periodic or square-integrable signals, again with explicit control of approximation error.
A clarifying example is the circle: circumference is exactly C=πdC=\pi dC=πd and area exactly A=πr2A=\pi r^2A=πr2. The constant π\piπ is transcendental, so its decimal expansion never terminates; nevertheless, analysis guarantees we can compute π\piπ (and any quantity depending on it) to any requested precision. This is the bridge between the ideal form (symbolic truth) and practical computation (controllable approximation).
4) Analog and Digital: A Productive Irreversibility
Analog phenomena become data through sampling and quantization; digital signals become analog through reconstruction. Both directions introduce irreversible information loss unless idealized conditions hold. Modern systems succeed not by eliminating this loss but by designing around it:
- choose sampling rates and quantizers to meet performance targets,
- propagate tolerances through the pipeline, and
- ensure that downstream algorithms are robust to the remaining uncertainty.
In other words, we do not require perfection; we require predictable, bounded imperfection that the system can tolerate without catastrophic failure.
5) Conclusion: Carrying “Truth” by Procedure
“Truth” in modern science is less a single number and more a well-specified procedure that (i) acknowledges uncertainty, (ii) controls error, and (iii) remains reproducible. Statistics supplies the norms for inference under uncertainty; analysis supplies the norms for approximation with guarantees. Engineering binds the two by encoding uncertainty into specifications, processes, and safeguards. That triad—inference, approximation, and design—is how contemporary research and industry “carry” truth from concept to artifact without pretending that error can be wished away.
Quick Glossary (for readers outside your field)
- Bias / Systematic error: persistent shift from the true value (hurts accuracy).
- Variance / Random error: scatter of repeated measurements (hurts precision).
- Confidence interval: a range that, under repeated sampling, would cover the true parameter at a stated long-run rate (e.g., 95%).
- Power: probability a test will detect a real effect of a specified size.
- DOE / Robust design: structuring experiments and products so performance is strong against noise factors.
- Cp/Cpk: indices comparing process spread/centering with specification limits.
- ε–δ: the formal language for “as close as you like” in limits and approximation.