Multivariable Calculus 18 | Local Extrema [dark version]
Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
A local maximum at x0 requires f(x0) ≥ f(x) for all x in D within some ε-neighborhood of x0; local minima reverse the inequality.
Briefing
Local extrema in multivariable calculus are defined by comparing function values only inside a small neighborhood around a specific point, not across the whole domain. A function f has a local maximum at x0 if there exists an ε-neighborhood around x0 such that f(x0) is greater than or equal to f(x) for every x in the domain within that neighborhood. The “local” part means the comparison is restricted to points near x0, while the “maximum” part means x0 beats (or ties) all nearby values. To avoid flat cases like constant functions, the notion of an isolated local maximum strengthens the condition to a strict inequality: f(x0) is strictly greater than f(x) for all nearby points except x0 itself. Local minima mirror this definition with the inequality reversed, and a local extremum means either a local maximum or a local minimum.
Once local extrema are defined, the calculus criteria start to look like the one-variable story, but with gradients and Hessians. For continuously differentiable functions on R^n, any local extremum at x0 must be a critical point: the gradient ∇f(x0) must equal the zero vector. Intuitively, the gradient points in the direction of fastest increase, so at a local extremum there can’t be any nearby direction that increases the function.
To decide whether a critical point is actually a maximum or minimum, the discussion turns to Taylor’s theorem with a quadratic approximation using the Hessian matrix Hf, built from all second partial derivatives. Near a critical point x0, the function behaves like f(x0 + h) ≈ f(x0) + (quadratic form involving Hf) + a remainder term. The sign pattern of this quadratic form determines the type of extremum: - If Hf is positive definite, the quadratic form is strictly positive for every nonzero direction h, forcing f(x0 + h) to be larger than f(x0). That guarantees an isolated local minimum at x0. - If Hf is negative definite, the quadratic form is strictly negative for every nonzero h, forcing f(x0 + h) to be smaller than f(x0). That guarantees an isolated local maximum. - If Hf is indefinite—meaning there exist directions where the quadratic form is positive and other directions where it is negative—then nearby points can produce both larger and smaller values. This rules out any local extremum and produces a saddle point (the classic “looks like a max in one direction and a min in another” geometry).
The transcript also gives one-directional necessary conditions. If f has a local maximum at x0 (not necessarily isolated), then Hf must be negative semi-definite, allowing the quadratic form to be zero in some nonzero directions. Similarly, a local minimum forces Hf to be positive semi-definite. Because semi-definiteness can occur even when the extremum isn’t isolated, the strict “definite” conditions are sufficient for isolated extrema, while the semi-definite conditions are only necessary for extrema.
In short: local extrema are neighborhood-based comparisons; critical points require ∇f(x0)=0; and the Hessian’s definiteness—positive definite, negative definite, or indefinite—determines whether the point is an isolated minimum, isolated maximum, or a saddle point, with semi-definiteness providing the weaker necessary condition when isolation isn’t guaranteed.
Cornell Notes
A function has a local maximum (or minimum) at x0 if, within some ε-neighborhood around x0, its value at x0 is respectively the largest (or smallest) among nearby points in the domain. For continuously differentiable functions, any local extremum must occur at a critical point where the gradient vanishes: ∇f(x0)=0. To classify a critical point, the quadratic (Taylor) approximation uses the Hessian matrix Hf. Positive definite Hf guarantees an isolated local minimum; negative definite Hf guarantees an isolated local maximum; indefinite Hf produces a saddle point with no local extremum. If a local maximum/minimum exists without requiring isolation, Hf must be negative/positive semi-definite (allowing zero curvature in some directions).
What exactly makes a maximum “local” in multivariable calculus?
Why must a local extremum occur at a critical point?
How does the Hessian matrix classify a critical point?
What is the difference between “isolated” and non-isolated local extrema in the Hessian test?
Why does an indefinite Hessian imply no local extremum?
Review Questions
- For a continuously differentiable function f: R^n → R, what condition must hold at any local extremum point x0?
- State the Hessian definiteness conditions that guarantee an isolated local minimum, an isolated local maximum, and a saddle point.
- If Hf is negative semi-definite at a critical point x0, what can be concluded (and what can’t be concluded) about local maxima?
Key Points
- 1
A local maximum at x0 requires f(x0) ≥ f(x) for all x in D within some ε-neighborhood of x0; local minima reverse the inequality.
- 2
Isolated local maxima/minima strengthen the neighborhood condition to strict inequality for all nearby points except x0, excluding flat cases like constant functions.
- 3
For continuously differentiable functions, any local extremum must be a critical point where the gradient vanishes: ∇f(x0)=0.
- 4
The Hessian matrix Hf determines the type of extremum via the sign of the quadratic approximation near x0.
- 5
Positive definite Hessian guarantees an isolated local minimum; negative definite Hessian guarantees an isolated local maximum.
- 6
Indefinite Hessian rules out local extrema and indicates a saddle point.
- 7
If a local maximum/minimum exists without assuming isolation, Hf must be negative/positive semi-definite, respectively.