Get AI summaries of any video or article — Sign up free
Multivariable Calculus 18 | Local Extrema [dark version] thumbnail

Multivariable Calculus 18 | Local Extrema [dark version]

5 min read

Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

A local maximum at x0 requires f(x0) ≥ f(x) for all x in D within some ε-neighborhood of x0; local minima reverse the inequality.

Briefing

Local extrema in multivariable calculus are defined by comparing function values only inside a small neighborhood around a specific point, not across the whole domain. A function f has a local maximum at x0 if there exists an ε-neighborhood around x0 such that f(x0) is greater than or equal to f(x) for every x in the domain within that neighborhood. The “local” part means the comparison is restricted to points near x0, while the “maximum” part means x0 beats (or ties) all nearby values. To avoid flat cases like constant functions, the notion of an isolated local maximum strengthens the condition to a strict inequality: f(x0) is strictly greater than f(x) for all nearby points except x0 itself. Local minima mirror this definition with the inequality reversed, and a local extremum means either a local maximum or a local minimum.

Once local extrema are defined, the calculus criteria start to look like the one-variable story, but with gradients and Hessians. For continuously differentiable functions on R^n, any local extremum at x0 must be a critical point: the gradient ∇f(x0) must equal the zero vector. Intuitively, the gradient points in the direction of fastest increase, so at a local extremum there can’t be any nearby direction that increases the function.

To decide whether a critical point is actually a maximum or minimum, the discussion turns to Taylor’s theorem with a quadratic approximation using the Hessian matrix Hf, built from all second partial derivatives. Near a critical point x0, the function behaves like f(x0 + h) ≈ f(x0) + (quadratic form involving Hf) + a remainder term. The sign pattern of this quadratic form determines the type of extremum: - If Hf is positive definite, the quadratic form is strictly positive for every nonzero direction h, forcing f(x0 + h) to be larger than f(x0). That guarantees an isolated local minimum at x0. - If Hf is negative definite, the quadratic form is strictly negative for every nonzero h, forcing f(x0 + h) to be smaller than f(x0). That guarantees an isolated local maximum. - If Hf is indefinite—meaning there exist directions where the quadratic form is positive and other directions where it is negative—then nearby points can produce both larger and smaller values. This rules out any local extremum and produces a saddle point (the classic “looks like a max in one direction and a min in another” geometry).

The transcript also gives one-directional necessary conditions. If f has a local maximum at x0 (not necessarily isolated), then Hf must be negative semi-definite, allowing the quadratic form to be zero in some nonzero directions. Similarly, a local minimum forces Hf to be positive semi-definite. Because semi-definiteness can occur even when the extremum isn’t isolated, the strict “definite” conditions are sufficient for isolated extrema, while the semi-definite conditions are only necessary for extrema.

In short: local extrema are neighborhood-based comparisons; critical points require ∇f(x0)=0; and the Hessian’s definiteness—positive definite, negative definite, or indefinite—determines whether the point is an isolated minimum, isolated maximum, or a saddle point, with semi-definiteness providing the weaker necessary condition when isolation isn’t guaranteed.

Cornell Notes

A function has a local maximum (or minimum) at x0 if, within some ε-neighborhood around x0, its value at x0 is respectively the largest (or smallest) among nearby points in the domain. For continuously differentiable functions, any local extremum must occur at a critical point where the gradient vanishes: ∇f(x0)=0. To classify a critical point, the quadratic (Taylor) approximation uses the Hessian matrix Hf. Positive definite Hf guarantees an isolated local minimum; negative definite Hf guarantees an isolated local maximum; indefinite Hf produces a saddle point with no local extremum. If a local maximum/minimum exists without requiring isolation, Hf must be negative/positive semi-definite (allowing zero curvature in some directions).

What exactly makes a maximum “local” in multivariable calculus?

Local means the comparison is restricted to points near x0. Formally, f has a local maximum at x0 if there exists ε>0 such that f(x0) ≥ f(x) for all x in D ∩ Bε(x0). The neighborhood Bε(x0) is the only region that matters; behavior far away from x0 is irrelevant to the definition.

Why must a local extremum occur at a critical point?

For continuously differentiable f, a local extremum at x0 forces the gradient to vanish: ∇f(x0)=0. The gradient represents the direction of fastest increase, so at a local maximum or minimum there can’t be any nearby direction that increases (or decreases) the function—hence no nonzero directional derivative remains.

How does the Hessian matrix classify a critical point?

Near a critical point x0, Taylor’s theorem gives a quadratic approximation of f(x0+h). The Hessian Hf forms the quadratic term from second partial derivatives. If that quadratic form is strictly positive for all nonzero h, Hf is positive definite and f has an isolated local minimum. If it’s strictly negative for all nonzero h, Hf is negative definite and f has an isolated local maximum. If it takes both positive and negative values for different directions h, Hf is indefinite and x0 is a saddle point.

What is the difference between “isolated” and non-isolated local extrema in the Hessian test?

Isolated extrema require strict inequality in the neighborhood definition, which corresponds to definiteness of the Hessian (positive definite for isolated minima, negative definite for isolated maxima). Without isolation, the necessary Hessian condition weakens to semi-definiteness: local maxima imply Hf is negative semi-definite (quadratic form ≤ 0), and local minima imply Hf is positive semi-definite (quadratic form ≥ 0). Semi-definiteness allows some nonzero directions where the quadratic term is zero.

Why does an indefinite Hessian imply no local extremum?

Indefinite means there exist directions h1 and h2 such that the quadratic approximation yields f(x0+h1) > f(x0) and f(x0+h2) < f(x0). Since nearby points can be both larger and smaller than f(x0), x0 cannot be a local maximum or minimum—this is the saddle point scenario.

Review Questions

  1. For a continuously differentiable function f: R^n → R, what condition must hold at any local extremum point x0?
  2. State the Hessian definiteness conditions that guarantee an isolated local minimum, an isolated local maximum, and a saddle point.
  3. If Hf is negative semi-definite at a critical point x0, what can be concluded (and what can’t be concluded) about local maxima?

Key Points

  1. 1

    A local maximum at x0 requires f(x0) ≥ f(x) for all x in D within some ε-neighborhood of x0; local minima reverse the inequality.

  2. 2

    Isolated local maxima/minima strengthen the neighborhood condition to strict inequality for all nearby points except x0, excluding flat cases like constant functions.

  3. 3

    For continuously differentiable functions, any local extremum must be a critical point where the gradient vanishes: ∇f(x0)=0.

  4. 4

    The Hessian matrix Hf determines the type of extremum via the sign of the quadratic approximation near x0.

  5. 5

    Positive definite Hessian guarantees an isolated local minimum; negative definite Hessian guarantees an isolated local maximum.

  6. 6

    Indefinite Hessian rules out local extrema and indicates a saddle point.

  7. 7

    If a local maximum/minimum exists without assuming isolation, Hf must be negative/positive semi-definite, respectively.

Highlights

Local extrema are defined by comparing function values only inside an ε-neighborhood around x0, not globally.
At any local extremum for a C^1 function, the gradient must be zero: ∇f(x0)=0.
Hessian definiteness gives the classification: positive definite → isolated minimum, negative definite → isolated maximum, indefinite → saddle point.
Semi-definite Hessians are necessary for (possibly non-isolated) extrema, but definiteness is what guarantees isolation.

Topics