Get AI summaries of any video or article — Sign up free
Ordinary Differential Equations 22 | Properties of the Matrix Exponential thumbnail

Ordinary Differential Equations 22 | Properties of the Matrix Exponential

4 min read

Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Matrix exponentials e^(tA) are defined for every square matrix A using the power series Σ_{k=0}^∞ (t^k A^k)/k!.

Briefing

Matrix exponentials turn linear systems of ordinary differential equations into an explicit solution formula: once a square matrix A is given, the columns of e^(tA) span the solution space of the homogeneous system x' = Ax, and multiplying e^(tA) by an initial vector x0 yields the unique solution of the initial value problem. That makes it essential not only to define e^(tA), but also to understand how it behaves under differentiation and how it interacts with algebraic operations like inversion.

The matrix exponential is defined for every square matrix A via the power series e^(tA) = Σ_{k=0}^∞ (t^k A^k)/k!. Convergence is handled entry-by-entry: each of the n^2 matrix entries becomes an ordinary real power series, so the matrix exponential exists for all real t and all square matrices A. On a compact interval [a,b], the convergence is uniform, which upgrades the function from merely well-defined to differentiable in t. Differentiability matters because it allows the matrix exponential to be used as a building block for solving differential equations.

Differentiation follows from the derivative definition using a limit as h → 0 of (e^((t+h)A) − e^(tA))/h. The key technical step is exchanging limit operations, which is justified by the uniform convergence on compact intervals. After substituting the series definition and differentiating term-by-term, the derivative simplifies to a form mirroring the scalar exponential rule: d/dt e^(tA) = A e^(tA) = e^(tA) A. The equality of the “left-multiplication” and “right-multiplication” versions comes from the structure of the series, where A can be pulled through the matrix products by reindexing.

Not every familiar exponential identity survives unchanged in the matrix setting. The scalar identity e^(a+b) = e^a e^b generally fails because matrix multiplication is not commutative. A related product formula (often involving the “Cauchy product” of power series) can be proved cleanly only when the matrices commute—specifically, when AB = BA. Under that commutativity condition, the usual exponentiation identity holds, and it becomes possible to compute inverses. In particular, if A commutes with −A (which it does), then e^(tA) e^(−tA) = e^((tA) + (−tA)) = e^0 = I, so (e^(tA))^−1 = e^(−tA). This inversion property is what makes the matrix exponential a practical tool for solving and manipulating solutions to linear ODE systems.

With these properties—existence via power series, differentiability, the derivative rule, and the inverse formula under commuting conditions—the stage is set for using e^(tA) directly to solve systems of linear differential equations in subsequent lessons.

Cornell Notes

For a square matrix A, the matrix exponential e^(tA) is defined by the power series Σ_{k=0}^∞ (t^k A^k)/k!. Each entry is a convergent scalar series, and on compact t-intervals the convergence is uniform, which implies e^(tA) is differentiable in t. Differentiating using the limit definition and justified limit exchange yields the key rule d/dt e^(tA) = A e^(tA) = e^(tA) A. The familiar scalar identity e^(a+b) = e^a e^b does not hold for arbitrary matrices because multiplication is noncommutative; it works when the matrices commute (AB = BA). With that commuting condition, inverses follow: (e^(tA))^−1 = e^(−tA).

Why does the matrix exponential e^(tA) exist for every square matrix A and every real t?

Its definition uses the power series e^(tA) = Σ_{k=0}^∞ (t^k A^k)/k!. Convergence is checked entry-by-entry: each of the n^2 entries becomes a real power series in t, so each entry converges. Because all entries converge, the matrix exponential exists as an n×n matrix for any real t and any square A.

How does uniform convergence on a compact interval [a,b] help with differentiation?

Uniform convergence of the sequence of matrix-valued functions (built from partial sums of the series) allows exchanging limit operations when applying the derivative definition. That technical permission makes term-by-term differentiation valid in the needed limit process, leading to a clean closed-form derivative.

What is the derivative of e^(tA) with respect to t, and why are there two equivalent forms?

The derivative is d/dt e^(tA) = A e^(tA) = e^(tA) A. The series computation produces a sum where an extra factor of A appears; reindexing and the structure of A^k within the series show that the factor A can be placed on the left or on the right without changing the result.

Why doesn’t the scalar exponentiation identity e^(a+b) = e^a e^b automatically carry over to matrices?

The scalar identity relies on commutativity when manipulating power series products. For matrices, multiplication is generally noncommutative, so the needed rearrangements fail. The identity holds when the matrices commute, specifically when AB = BA.

How does commutativity enable the inverse formula (e^(tA))^−1 = e^(−tA)?

If A commutes with −A (true since −A is just a scalar multiple of A), then e^(tA) e^(−tA) can be combined using the commuting exponentiation identity: e^(tA) e^(−tA) = e^((tA)+(-tA)) = e^0 = I. That shows e^(−tA) is the inverse of e^(tA).

Review Questions

  1. State the power series definition of e^(tA) and explain how convergence is justified for matrix entries.
  2. Derive (or recall) the formula for d/dt e^(tA) and state both equivalent multiplication orders.
  3. Under what condition does e^(A+B) = e^A e^B hold for matrices, and how does that condition lead to an inverse formula?

Key Points

  1. 1

    Matrix exponentials e^(tA) are defined for every square matrix A using the power series Σ_{k=0}^∞ (t^k A^k)/k!.

  2. 2

    Each matrix entry converges because the series can be treated as n^2 scalar power series in t.

  3. 3

    Uniform convergence on compact intervals implies e^(tA) is differentiable with respect to t.

  4. 4

    The derivative satisfies d/dt e^(tA) = A e^(tA) = e^(tA) A.

  5. 5

    The scalar rule e^(a+b) = e^a e^b fails for general matrices due to noncommutativity.

  6. 6

    The exponentiation identity works when matrices commute (AB = BA).

  7. 7

    When the commuting condition applies, inverses follow: (e^(tA))^−1 = e^(−tA).

Highlights

Uniform convergence on compact t-intervals is what makes the derivative computation legitimate for matrix exponentials.
The clean derivative rule mirrors the scalar case: d/dt e^(tA) = A e^(tA), and it also equals e^(tA) A.
The exponentiation identity e^(A+B) = e^A e^B requires commutativity (AB = BA); without it, the algebra breaks.
Once commutativity is available, inversion becomes immediate: (e^(tA))^−1 = e^(−tA).