Get AI summaries of any video or article — Sign up free
How (and why) to raise e to the power of a matrix | DE6 thumbnail

How (and why) to raise e to the power of a matrix | DE6

3Blue1Brown·
5 min read

Based on 3Blue1Brown's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Matrix exponentiation is defined by the Taylor series exp(A)=I + A + A^2/2! + …, using matrix powers and matrix addition term-by-term.

Briefing

Matrix exponentiation—written as e^(At)—turns out to be a precise way to solve systems of differential equations where a state changes at a rate proportional to a matrix times the state. That matters because many real problems in science and engineering (and especially quantum mechanics) can be written in the form v'(t)=A v(t). Instead of guessing solutions, e^(At) packages the long-term evolution from time 0 to time t into a single matrix, so multiplying e^(At) by an initial condition vector produces the full trajectory.

The operation starts with a definition that looks like “plugging a matrix into the exponential,” but it’s not arbitrary. For real numbers, e^x is defined by its Taylor series; the matrix version uses the same infinite polynomial, term by term: exp(A)=I + A + A^2/2! + A^3/3! + … . This requires A to be square so that powers like A^2, A^3, and so on are well-defined via repeated matrix multiplication. Adding matrices works term-by-term as well, so the Taylor-series construction makes sense as a formal infinite sum—at least in cases where it converges.

A concrete example makes the definition feel less like notation abuse. Take the 2×2 matrix with −π and π on the off-diagonal: [[0, −π],[π, 0]]. When its Taylor-series exponential is computed, the sum approaches −I, the negative identity matrix. That result isn’t a coincidence: it’s the matrix analogue of Euler’s identity, and it reflects what happens when the underlying matrix represents a 90-degree rotation. Exponentiating a rotation operator for time t generates the rotation by angle t; after π units of time, a 90-degree-per-unit-time rotation becomes a 180-degree flip, which is exactly multiplication by −1.

The “why” becomes clearer through differential equations. The Romeo-and-Juliet relationship model encodes two coupled rates of change: one person’s feelings increase when the other’s decrease, and vice versa. Writing the pair (x(t), y(t)) as a vector v(t) turns the system into v'(t)=M v(t), where M is a specific 2×2 matrix. Geometrically, M acts like a 90-degree rotation, so the velocity vector is always perpendicular to the position vector. That constraint forces circular motion at a constant angular speed, giving explicit solutions that match what e^(Mt) produces.

The same framework scales up. In one dimension, x'(t)=r x(t) has solutions x(t)=x0 e^(rt), showing that the exponential acts on an initial condition. In higher dimensions, the exponential becomes a matrix-valued operator: e^(At) plays the role of “the thing that turns initial data into the evolving state.” A quick calculus check supports this: differentiating the Taylor-series expression for e^(At) yields A times the original expression, so it satisfies v'(t)=A v(t) (with convergence details handled separately).

Finally, the discussion connects the idea to quantum mechanics. Schrödinger’s equation has the same structural core—time evolution governed by a matrix acting on a state vector—but the matrix includes the imaginary unit i, which enforces oscillatory behavior. More generally, the evolution can be visualized with a vector field where each point v has velocity Mv; flowing along that field for time t produces the transformation encoded by e^(Mt).

Cornell Notes

Matrix exponentiation defines e^(A) using the same Taylor series that defines the real exponential, replacing powers of a number with powers of a square matrix: exp(A)=I + A + A^2/2! + A^3/3! + … . This construction is not just formal: it solves linear systems of differential equations of the form v'(t)=A v(t). In the Romeo-and-Juliet example, the matrix acts like a 90-degree rotation, so the solution rotates in time; computing e^(Mt) via the Taylor series reproduces the rotation matrix. A key payoff is that e^(At) turns initial conditions into full time evolution: v(t)=e^(At)v(0). The same mechanism underlies time evolution in quantum mechanics, where i introduces oscillations and the “rotation” happens in a space of states or functions.

How is e^(A) defined for a matrix A, and what makes the definition legitimate?

For a square matrix A, e^(A) is defined by the Taylor series for the real exponential, term-by-term: e^(A)=I + A + A^2/2! + A^3/3! + … . Matrix powers mean repeated multiplication (A^2=A·A, etc.), and factorial denominators scale each matrix term. The construction relies on the fact that matrix addition and multiplication are well-defined, and—when convergence holds—the infinite sum approaches a stable matrix.

Why does the example with off-diagonal ±π produce −I?

The matrix [[0, −π],[π, 0]] acts like a 90-degree rotation generator. Exponentiating a rotation operator for time t yields a rotation by angle t. After π units of time, the rotation angle is 180 degrees, which flips every vector to its negative—equivalent to multiplying by −1, i.e., the result is −I. The Taylor-series computation matches this geometric interpretation.

How does e^(At) solve v'(t)=A v(t)?

Start with the Taylor-series expression for e^(At) and multiply it by an initial vector v(0). Differentiating with respect to t brings down factors that cancel the factorials, leaving an extra A on each term. That extra A factors out, giving d/dt(e^(At)v(0)) = A(e^(At)v(0)). So v(t)=e^(At)v(0) satisfies the differential equation, aside from technical convergence details.

What does the Romeo-and-Juliet system have to do with rotation?

The coupled equations for x(t) and y(t) can be written as v'(t)=M v(t) for a specific 2×2 matrix M. Reading M’s action shows it rotates vectors by 90 degrees: the velocity vector is always perpendicular to the position vector. Perpendicular velocity forces circular motion around the origin, with angular speed 1 radian per unit time, matching the explicit solution produced by e^(Mt).

How does this connect to quantum mechanics and the role of i?

Schrödinger’s equation has the same structural form: a state vector changes over time according to a matrix acting on that state. In many formulations, the matrix includes the imaginary unit i, which makes the evolution oscillatory rather than purely growing/decaying. The “rotation” intuition extends, but the rotation may occur in a space of states or functions rather than a visible 2D plane.

Review Questions

  1. What conditions must a matrix satisfy for the Taylor-series definition of e^(A) to be meaningful in terms of matrix powers?
  2. In the rotation example, how does the angle of rotation relate to the exponent parameter t?
  3. Why does differentiating the Taylor-series form of e^(At) naturally produce the factor A needed for v'(t)=A v(t)?

Key Points

  1. 1

    Matrix exponentiation is defined by the Taylor series exp(A)=I + A + A^2/2! + …, using matrix powers and matrix addition term-by-term.

  2. 2

    For square matrices, powers like A^2, A^3, and so on are defined by repeated matrix multiplication, making the series well-formed.

  3. 3

    The identity exp(Mt) for a 90-degree-rotation generator reproduces rotation by angle t, so exp(Mπ) becomes −I.

  4. 4

    Linear systems v'(t)=A v(t) have solutions v(t)=e^(At)v(0), turning initial conditions into full time evolution.

  5. 5

    Differentiating the Taylor-series expression for e^(At) yields A times the original expression, matching the differential equation’s structure (subject to convergence).

  6. 6

    Coupled differential equations can be visualized with a vector field where each point v has velocity Mv; flowing along the field for time t corresponds to applying e^(Mt).

  7. 7

    Schrödinger’s equation fits the same evolution pattern, with the imaginary unit i driving oscillatory behavior and “rotations” occurring in state/function space.

Highlights

Exponentiating a matrix that generates 90-degree rotation produces rotation by angle t; after π time units, the result is −I.
The Taylor-series definition of e^(A) isn’t just notation—it directly leads to solutions of v'(t)=A v(t) via differentiation.
In the Romeo-and-Juliet model, the coupled rates of change become a single vector equation v'(t)=M v(t), making geometry do the heavy lifting.
Quantum time evolution follows the same linear-algebraic template, with i enforcing oscillations rather than monotone growth.

Topics