Get AI summaries of any video or article — Sign up free
Ordinary Differential Equations 6 | Separation of Variables [dark version] thumbnail

Ordinary Differential Equations 6 | Separation of Variables [dark version]

4 min read

Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Separation of variables applies to non-autonomous ODEs that can be written as ẋ = G(t)·h(x), separating time and state into different factors.

Briefing

Separation of variables provides a practical route to solving certain non-autonomous ordinary differential equations by rewriting them so the time-dependent part and the state-dependent part sit in separate factors. For an ODE of the form ẋ = G(t)·h(x), the method works when h(x) is not zero at the initial state, letting the equation be rearranged into 1/h(x)·dx = G(t)·dt. With an initial value condition x(t0) = x0, integrating both sides from t0 to an arbitrary time T produces an implicit solution that can be expressed using antiderivatives and then inverted to obtain x(T). The key payoff is that the same calculus steps—rearrange, integrate, apply initial data, invert—repeat in a predictable pattern.

The procedure begins by checking the “nontrivial” case: if h(x0) = 0, the constant solution x(t) ≡ x0 already satisfies the equation, so there’s nothing to separate. Otherwise, near x0 the function h(x) stays nonzero, so dividing by h(x) is legitimate. After rearranging, the method leverages the fundamental theorem of calculus by integrating both sides with respect to their natural variables. Antiderivatives appear on both sides: one for 1/h(x) and one for G(t). Because antiderivatives differ only by an additive constant, the constants can be merged into a single real constant C. The final step is to invert the resulting relationship to solve explicitly for the solution function α(t) (the trajectory satisfying the initial condition).

A worked example makes the mechanics concrete. For ẋ = T^3·x^2 with x(0) = x0, separation is possible because the right-hand side is a product of a pure time factor T^3 and a pure state factor x^2. Rearranging yields dx/x^2 = T^3 dt, and integrating produces an expression involving the natural logarithm of |x| on the left and a power of t on the right, plus a constant. Exponentiating (using the inverse of the natural logarithm) and then enforcing x(0) = x0 determines the constant, giving the explicit solution in terms of an exponential factor scaled by x0.

A second example shows the method still applies when the state factor is exponential. For ẋ = sin(t)·e^x, separation leads to dx/e^x = sin(t) dt. Integrating gives an implicit relation involving e^{-x} on the left and cos(t) on the right, again with an added constant. Solving for x requires applying the inverse of the exponential relationship, producing a logarithmic expression. The constant is then fixed by the initial condition x(0) = x0, using cos(0) = 1 to simplify the algebra. The result is a fully determined solution α(t), obtained through the same repeated steps: separate, integrate, combine constants, invert, and apply initial data.

Overall, the method’s importance lies in turning a differential equation into an algebraic problem after integration. Rather than memorizing a general formula, the emphasis is on executing the separation workflow reliably on each example—especially for non-autonomous equations where time dependence would otherwise complicate direct solution.

Cornell Notes

Separation of variables solves non-autonomous ODEs that can be written as ẋ = G(t)·h(x). With an initial value problem x(t0)=x0, the method requires h(x0)≠0; otherwise the constant solution x(t)≡x0 already works. When separation is possible, the equation is rearranged into (1/h(x))dx = G(t)dt, then both sides are integrated from t0 to T. Antiderivatives introduce an additive constant, which is merged into a single C. The final implicit relation is inverted to obtain x(T), and the initial condition determines C.

What structural form of a non-autonomous ODE makes separation of variables possible?

Separation works when the ODE can be written as ẋ = G(t)·h(x), meaning the right-hand side is a product of a function of time only (G(t)) and a function of the state only (h(x)). In that case, rearranging produces (1/h(x))dx = G(t)dt, so time and state variables sit on opposite sides for integration.

Why does the method treat the case h(x0)=0 differently?

If h(x0)=0, then the constant function x(t)≡x0 satisfies the ODE immediately, because ẋ = G(t)·h(x0)=0. The “interesting” separation step requires dividing by h(x), so the method assumes h(x)≠0 in a neighborhood of x0 to avoid division by zero.

How does the fundamental theorem of calculus enter the separation workflow?

After rearranging into (1/h(x))dx = G(t)dt, integrating both sides from t0 to T is the key move. On the state side, the integral becomes an antiderivative evaluated between x0 and α(t). On the time side, the integral becomes an antiderivative evaluated between t0 and T. This evaluation step is what the fundamental theorem of calculus formalizes.

Why can the constants from antiderivatives be combined into a single constant C?

Any two antiderivatives of the same function differ by an additive constant. Since both sides of the separated equation introduce their own “+ constant,” those constants can be absorbed into one combined constant C in the final integrated equation. That simplification makes the inversion and initial-condition step cleaner.

In the example ẋ = T^3·x^2, how does the solution get determined from the initial condition?

After separation and integration, the result involves ln|x| on the left and a t^4 term on the right, plus a constant. Exponentiating removes the logarithm, producing an expression for x(t) containing an exponential factor and an undetermined constant. Plugging in x(0)=x0 fixes that constant, yielding the explicit solution with x0 as the scaling factor.

In the example ẋ = sin(t)·e^x, what role does cos(0)=1 play?

Integration produces a relation involving cos(t) and a constant. When applying x(0)=x0, evaluating cos(0)=1 simplifies the constant determination. The constant ends up expressed using e^{-x0} and the value of cos(0), which then feeds back into the final logarithmic expression for x(t).

Review Questions

  1. What conditions on h(x0) determine whether a constant solution exists without needing separation?
  2. After separating ẋ = G(t)·h(x) into (1/h(x))dx = G(t)dt, what are the limits of integration you would use for an initial value problem at t0?
  3. Why is it often better to follow the separation-and-integration steps rather than memorize a closed-form template?

Key Points

  1. 1

    Separation of variables applies to non-autonomous ODEs that can be written as ẋ = G(t)·h(x), separating time and state into different factors.

  2. 2

    If h(x0)=0, the constant solution x(t)≡x0 satisfies the ODE, so the nontrivial separation step is unnecessary.

  3. 3

    When h(x0)≠0, rearrange to (1/h(x))dx = G(t)dt and integrate both sides from t0 to T.

  4. 4

    Antiderivatives introduce additive constants that can be merged into a single constant C after integration.

  5. 5

    After integration, invert the resulting relationship to solve for x(T) explicitly (when possible).

  6. 6

    Use the initial condition x(t0)=x0 to determine the constant C and finalize the solution.

Highlights

Separation of variables turns ẋ = G(t)·h(x) into an integrable form by rearranging to (1/h(x))dx = G(t)dt.
The constants from antiderivatives collapse into one combined constant C, simplifying the final implicit equation.
For ẋ = T^3·x^2, the integration leads to ln|x|, and exponentiating plus x(0)=x0 yields an explicit exponential-form solution.
For ẋ = sin(t)·e^x, integration produces terms with cos(t), and applying x(0)=x0 uses cos(0)=1 to fix the constant.
The method emphasizes doing the separation-and-integration steps in examples rather than memorizing a general formula.

Topics