Multivariable Calculus 16 | Taylor's Theorem [dark version]
Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
For f ∈ C^{k+1}(R^n), values near x̃ satisfy f(x̃ + h) = T_k(x̃, h) + R_k with an error that shrinks as h → 0.
Briefing
Taylor’s theorem for multivariable functions generalizes the familiar 1D idea of approximating a smooth function near an expansion point with a polynomial whose coefficients come from derivatives. The key takeaway is that for a function f: R^n → R that is sufficiently differentiable (specifically f ∈ C^{k+1}(R^n)), the value at a nearby point x̃ + h can be written as a k-th order Taylor polynomial plus a remainder term that becomes negligible faster than ‖h‖^k as h → 0. This matters because it turns local behavior of multivariable functions into an explicit polynomial approximation—central for analysis, optimization, and numerical methods.
The construction starts with the same logic as in one dimension: the first-order (linear) approximation uses differentiability to match the tangent behavior. In multiple dimensions, the linear term is expressed using the Jacobian matrix J_f evaluated at the expansion point x̃, producing a linear map applied to the increment vector h. The quality of this approximation is tracked by an error term Φ(h) that satisfies Φ(h)/‖h‖ → 0 as h → 0.
For the quadratic approximation, the multivariable version replaces “the second derivative” with the Hessian matrix. The quadratic term takes the form h^T H_f(x̃) h, where the vector h appears on both sides so that the scalar output matches the second-order curvature information encoded in the Hessian. As with the linear case, a remainder term Ψ(h) measures the error, and the condition for a true quadratic approximation is that Ψ(h)/‖h‖^2 → 0 as h → 0.
With these building blocks, the full k-th order Taylor theorem is stated using multi-index notation. For multi-indices α, the Taylor polynomial T_k(x̃, h) is a sum over all α with |α| ≤ k of the form (∂^α f(x̃)/α!) h^α. Here, ∂^α denotes the mixed partial derivative corresponding to α, α! is the product of factorials of the components of α, and h^α is the corresponding product of powers of the components of h. The theorem then adds a remainder term R_k that captures the approximation error.
The remainder term mirrors the 1D structure: it uses derivatives of order k+1 evaluated at an intermediate point c, which lies on the line segment between x̃ and x̃ + h. In the multivariable setting, “intermediate” means c is a point on that straight path in R^n, generalizing the single-variable point between x̃ and x̃ + h. The result reduces to the standard 1D Taylor theorem when n = 1, making the multivariable formula a direct extension rather than a new, unrelated concept. The practical implication is clear: once f has enough continuous derivatives, its local behavior near x̃ is well-approximated by a polynomial whose coefficients are determined by partial derivatives at x̃, with a controlled error that shrinks rapidly as h becomes small.
Cornell Notes
Multivariable Taylor’s theorem extends the 1D polynomial approximation idea to functions f: R^n → R near an expansion point x̃. If f has continuous partial derivatives up to order k+1 (f ∈ C^{k+1}(R^n)), then for a small increment h the value f(x̃ + h) equals a k-th order Taylor polynomial T_k(x̃, h) plus a remainder R_k. The polynomial uses multi-index notation: T_k sums terms (∂^α f(x̃)/α!) h^α over all multi-indices α with |α| ≤ k. The remainder depends on derivatives of order k+1 evaluated at an intermediate point c on the line segment between x̃ and x̃ + h. This matters because it provides a systematic way to approximate multivariable functions locally with a controlled error.
How does the linear approximation of a multivariable function near x̃ relate to the Jacobian matrix?
What replaces the “second derivative” in the quadratic approximation, and how does it enter the formula?
How is the k-th order Taylor polynomial written using multi-index notation?
What does the remainder term R_k depend on, and what is the role of the intermediate point c?
Why does the multivariable theorem reduce to the familiar 1D Taylor theorem when n = 1?
Review Questions
- What conditions on f (in terms of C^{k+1}(R^n)) are needed to guarantee a k-th order Taylor expansion with a controlled remainder?
- Write the general form of the k-th order Taylor polynomial T_k(x̃, h) using multi-index notation, including the roles of ∂^α, α!, and h^α.
- In the remainder term, where does the intermediate point c lie relative to x̃ and x̃ + h, and why is that geometric description important?
Key Points
- 1
For f ∈ C^{k+1}(R^n), values near x̃ satisfy f(x̃ + h) = T_k(x̃, h) + R_k with an error that shrinks as h → 0.
- 2
The linear approximation uses the Jacobian matrix J_f(x̃) applied to the increment vector h, with an error term Φ(h) satisfying Φ(h)/‖h‖ → 0.
- 3
The quadratic approximation uses the Hessian matrix H_f(x̃) in the scalar form h^T H_f(x̃) h, with a remainder Ψ(h) satisfying Ψ(h)/‖h‖^2 → 0.
- 4
The k-th order Taylor polynomial is built from mixed partial derivatives at x̃ using multi-index notation: Σ_{|α|≤k} (∂^α f(x̃)/α!) h^α.
- 5
The remainder term R_k depends on derivatives of order k+1 evaluated at an intermediate point c on the line segment between x̃ and x̃ + h.
- 6
Setting n = 1 collapses the multivariable formulas back to the standard one-dimensional Taylor theorem, including the “between points” intermediate location.