Distributions 1 | Motivation and Delta Function [dark version]
Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
The Heaviside step function H(x) has a jump at x=0, so its classical derivative fails to exist there.
Briefing
Distributions were introduced to make sense of derivatives and other operations that break down at “sharp” features—especially jumps like the Heaviside step function. In classical calculus, the derivative of the Heaviside function H(x) (0 for x<0 and 1 for x>0) fails to exist at x=0 because of the jump. Yet many applications—most notably differential equations and Fourier analysis—need a derivative-like object that still behaves correctly even when solutions have discontinuities or corners.
The motivation traces back to Paul Dirac’s 1927 work. Dirac wanted a new “delta function” δ(x) that would act as the derivative of the Heaviside function. Away from x=0, this is straightforward: H is constant on each side, so any derivative-like object should be zero there. The real challenge is the single point at x=0. Dirac also demanded that a fundamental theorem of calculus–type identity remain valid in an integral form: integrating δ(x) over a tiny interval around 0 should reproduce the jump in H. Concretely, for arbitrarily small ε, the integral of δ(x) from −ε to ε should equal H(ε)−H(−ε), which is 1.
But those requirements clash with what ordinary functions can do. If δ(x) were an ordinary function that is zero everywhere except at one point, then its integral over any interval would be zero—because changing a function at a single point does not affect integrals (in the sense of “almost everywhere,” i.e., with respect to the Lebesgue measure). That produces a contradiction: the integral would have to be 0, not 1. The situation worsens when one tries to differentiate δ again; even writing down δ′ requires a framework that can interpret such expressions without relying on pointwise values.
The response is to replace “functions” with a broader concept: distributions, also called generalized functions. Instead of treating δ as a literal function with a graph, distributions are treated as linear functionals acting on test functions. The guiding picture is that distributions behave like mass densities. For an ordinary function f(x), integrating f(x) against a test function φ(x) over a region produces a real number that can be interpreted as “mass measured” in the bump-shaped region where φ is nonzero. For the delta distribution, all the mass is concentrated at x=0. When a test function φ is used, the resulting number corresponds to φ evaluated at the point of concentration—so δ “picks out” the value at 0.
This shift—from pointwise objects to measurement-by-integration—lets δ satisfy the integral jump condition and supports further algebraic operations like differentiation in a consistent way. The next step is to formalize what counts as a test function and to define precisely how distributions act as linear maps on them.
Cornell Notes
The Heaviside step function H(x) jumps from 0 to 1 at x=0, so its classical derivative fails to exist at that point. Dirac introduced the delta function δ(x) as a generalized derivative of H, requiring that integrating δ over any small interval around 0 reproduces the jump: ∫_{-ε}^{ε} δ(x) dx = 1. That can’t happen with an ordinary function that is zero everywhere except at a single point, because such a function would integrate to 0. Distributions fix this by treating δ not as a function with a graph, but as a linear functional acting on test functions φ: the delta distribution returns φ(0), reflecting a “point mass” density at x=0. This framework also sets up meaningful operations like differentiating δ.
Why does the classical derivative of the Heaviside function break down at x=0?
What integral condition does Dirac want δ(x) to satisfy, and why does it matter?
Why can’t δ(x) be an ordinary function that is zero everywhere except at x=0?
How do distributions reinterpret δ so the integral condition becomes possible?
What are test functions, and what role do they play in measuring a distribution?
Review Questions
- What specific contradiction arises if δ is assumed to be an ordinary function that is zero almost everywhere?
- How does the “point mass” interpretation explain why δ acting on a test function yields φ(0)?
- What additional difficulty appears when trying to differentiate δ again, and why does that push toward distributions?
Key Points
- 1
The Heaviside step function H(x) has a jump at x=0, so its classical derivative fails to exist there.
- 2
Dirac’s delta function δ(x) was introduced to act like the derivative of H while preserving an integral jump condition.
- 3
The requirement ∫_{-ε}^{ε} δ(x) dx = H(ε)−H(−ε) forces the integral to equal 1 for all small ε.
- 4
An ordinary function that is zero everywhere except at one point would still integrate to 0, creating a contradiction.
- 5
Distributions replace pointwise functions with linear functionals that act on test functions.
- 6
The delta distribution behaves like a point mass at x=0: when applied to a test function φ, it returns φ(0).
- 7
The framework is designed to support operations like differentiation even when classical derivatives are undefined at sharp features.