Ordinary Differential Equations

Existence and uniqueness, linear systems, stability, and bifurcation theory.


Ordinary differential equations are the mathematical language in which nature writes its laws of change. Wherever a quantity evolves in time — the position of a planet, the voltage across a circuit, the concentration of a chemical species, the size of a population — an ODE lurks behind the phenomenon, relating the rate of change of the quantity to its current value. The subject stretches from the seventeenth century, when Newton and Leibniz invented the calculus partly in order to solve such equations, through Poincaré’s late-nineteenth-century geometric revolution, all the way to modern dynamical systems theory and numerical computation.

First-Order Differential Equations

A first-order ordinary differential equation (ODE) is a relation of the form F(x,y,y)=0F(x, y, y') = 0, where y=y(x)y = y(x) is an unknown function and y=dy/dxy' = dy/dx is its derivative. In most situations of practical interest we can write it in the explicit form y=f(x,y)y' = f(x, y) for some given function ff. A solution on an interval II is a differentiable function ϕ:IR\phi : I \to \mathbb{R} that satisfies the equation identically: ϕ(x)=f(x,ϕ(x))\phi'(x) = f(x, \phi(x)) for all xIx \in I.

The simplest and most important class is the separable equation, where ff factors as a product g(x)h(y)g(x) h(y). The equation y=g(x)h(y)y' = g(x) h(y) can be formally separated and integrated:

dyh(y)=g(x)dx+C.\int \frac{dy}{h(y)} = \int g(x)\, dx + C.

This method goes back to Leibniz and Johann Bernoulli, who used it to solve the catenary problem in 1691. The result is typically an implicit relation between yy and xx, which may or may not yield an explicit formula for yy.

The next class is the linear first-order equation y+P(x)y=Q(x)y' + P(x) y = Q(x). The key tool is the integrating factor μ(x)=eP(x)dx\mu(x) = e^{\int P(x)\, dx}, which converts the left side into an exact derivative:

ddx[μ(x)y]=μ(x)Q(x).\frac{d}{dx}\bigl[\mu(x)\, y\bigr] = \mu(x)\, Q(x).

A single integration then yields the general solution. The structure is transparent: every solution is a sum of the homogeneous solution (the solution when Q=0Q = 0) and a particular solution driven by QQ. This superposition principle is the defining feature of linearity and will pervade everything that follows.

Beyond separable and linear equations lie the classical special types. A Bernoulli equation y+P(x)y=Q(x)yny' + P(x) y = Q(x) y^n is nonlinear but is tamed by the substitution v=y1nv = y^{1-n}, which reduces it to a linear equation. A homogeneous equation y=f(y/x)y' = f(y/x) is handled by v=y/xv = y/x. An exact equation M(x,y)dx+N(x,y)dy=0M(x,y)\,dx + N(x,y)\,dy = 0 is one for which M/y=N/x\partial M/\partial y = \partial N/\partial x, so that Mdx+NdyM\,dx + N\,dy is the exact differential of some potential function Ψ(x,y)\Psi(x,y); the general solution is then Ψ(x,y)=C\Psi(x,y) = C. When exactness fails, one seeks an integrating factor μ(x,y)\mu(x,y) that restores it.

The geometric picture is indispensable. The equation y=f(x,y)y' = f(x,y) assigns to each point (x,y)(x,y) in the plane a slope f(x,y)f(x,y), forming a direction field (or slope field). Solution curves are precisely those that are tangent to the direction field at every point. Drawing a direction field — even roughly, by hand — reveals qualitative behaviour (equilibria, funnels, separatrices) without solving the equation. This geometric thinking, systematized by Poincaré in the 1880s, was a conceptual revolution: it shifted attention from explicit formulas to the global structure of the solution set.

Existence and Uniqueness Theorems

A fundamental question precedes any computation: does the equation even have a solution, and if so, is it unique? The answer depends delicately on the regularity of the right-hand side ff.

An initial value problem (IVP) pairs the ODE y=f(x,y)y' = f(x, y) with the condition y(x0)=y0y(x_0) = y_0. The solution must pass through the specified point (x0,y0)(x_0, y_0). The foundational result is the Picard-Lindelöf theorem (also called the Cauchy-Lipschitz theorem):

Theorem (Picard-Lindelöf). If ff is continuous on a rectangle R={xx0a,yy0b}R = \{|x - x_0| \le a,\, |y - y_0| \le b\} and satisfies a Lipschitz condition in yy on RR, meaning there exists L>0L > 0 such that

f(x,y1)f(x,y2)Ly1y2|f(x, y_1) - f(x, y_2)| \le L\,|y_1 - y_2|

for all (x,y1),(x,y2)R(x, y_1), (x, y_2) \in R, then there exists h>0h > 0 such that the IVP has a unique solution on [x0h,x0+h][x_0 - h, x_0 + h].

The proof is constructive. Define the Picard iteration ϕ0(x)=y0\phi_0(x) = y_0 and

ϕn+1(x)=y0+x0xf(t,ϕn(t))dt.\phi_{n+1}(x) = y_0 + \int_{x_0}^{x} f\bigl(t, \phi_n(t)\bigr)\, dt.

One shows that {ϕn}\{\phi_n\} is a Cauchy sequence in the Banach space C[x0h,x0+h]C[x_0-h, x_0+h] under the sup norm, and its limit is the unique solution. The argument is an instance of the Banach contraction mapping theorem, making this one of the places where real analysis — completeness, uniform convergence — enters ODE theory decisively. This is why real analysis is a prerequisite: the existence and uniqueness theory is functional-analytic at its core.

Without the Lipschitz condition, uniqueness can fail. The classical example is y=y1/2y' = y^{1/2}, y(0)=0y(0) = 0: both y0y \equiv 0 and y=x2/4y = x^2/4 are solutions, because f(x,y)=y1/2f(x,y) = y^{1/2} fails to be Lipschitz near y=0y = 0. Peano’s theorem (1886) guarantees existence under mere continuity of ff, but not uniqueness.

A solution need not exist for all time. The ODE y=y2y' = y^2, y(0)=1y(0) = 1 has the explicit solution y=1/(1x)y = 1/(1-x), which blows up at x=1x = 1. The maximal interval of existence is the largest interval on which the solution remains defined. A general fact is that if the solution does not exist globally, it must “escape to infinity” in finite time: either y(x)|y(x)| \to \infty or xx approaches a boundary of the domain of ff.

Continuous dependence on initial data is equally important. If (x0,y0)(x_0, y_0) is perturbed slightly to (x0,y~0)(x_0, \tilde y_0), how much does the solution change? Under the Lipschitz condition, the answer is controlled by Gronwall’s inequality: if two solutions ϕ\phi and ϕ~\tilde\phi satisfy ϕ(x0)ϕ~(x0)δ|\phi(x_0) - \tilde\phi(x_0)| \le \delta, then

ϕ(x)ϕ~(x)δeLxx0|\phi(x) - \tilde\phi(x)| \le \delta\, e^{L|x - x_0|}

on any common interval of existence. Solutions thus depend continuously — indeed, smoothly — on initial conditions and on parameters appearing in ff. This is the mathematical foundation for the physical intuition that small measurement errors lead only to small prediction errors, at least over bounded time horizons.

Linear Systems and Matrix Exponentials

Many real-world models involve several interacting quantities described by a system of first-order ODEs. When written in vector form, the system x=Ax\mathbf{x}' = A\mathbf{x} (where xRn\mathbf{x} \in \mathbb{R}^n and AA is an n×nn \times n constant matrix) is the centerpiece of linear ODE theory and the direct analogue of the scalar equation y=ayy' = ay.

The scalar equation y=ayy' = ay has the solution y(t)=eaty(0)y(t) = e^{at} y(0). By analogy, the vector system x=Ax\mathbf{x}' = A\mathbf{x}, x(0)=x0\mathbf{x}(0) = \mathbf{x}_0, has the solution x(t)=eAtx0\mathbf{x}(t) = e^{At} \mathbf{x}_0, where the matrix exponential is defined by the convergent power series:

eAt=k=0(At)kk!=I+At+A2t22!+A3t33!+e^{At} = \sum_{k=0}^{\infty} \frac{(At)^k}{k!} = I + At + \frac{A^2 t^2}{2!} + \frac{A^3 t^3}{3!} + \cdots

Computing eAte^{At} in practice is easiest when AA is diagonalizable: if A=PDP1A = PDP^{-1} with D=diag(λ1,,λn)D = \mathrm{diag}(\lambda_1, \ldots, \lambda_n), then eAt=PeDtP1=Pdiag(eλ1t,,eλnt)P1e^{At} = P\, e^{Dt}\, P^{-1} = P\, \mathrm{diag}(e^{\lambda_1 t}, \ldots, e^{\lambda_n t})\, P^{-1}. Each column of PeDtPe^{Dt} is an independent solution of the form eλitvie^{\lambda_i t} \mathbf{v}_i, where vi\mathbf{v}_i is an eigenvector corresponding to λi\lambda_i. The general solution is thus a superposition of these normal modes:

x(t)=c1eλ1tv1+c2eλ2tv2++cneλntvn.\mathbf{x}(t) = c_1 e^{\lambda_1 t} \mathbf{v}_1 + c_2 e^{\lambda_2 t} \mathbf{v}_2 + \cdots + c_n e^{\lambda_n t} \mathbf{v}_n.

When AA has complex eigenvalues λ=α±iβ\lambda = \alpha \pm i\beta, the corresponding complex normal modes combine via Euler’s formula into real oscillatory solutions of the form eαtcos(βt)e^{\alpha t} \cos(\beta t) and eαtsin(βt)e^{\alpha t} \sin(\beta t). When AA is not diagonalizable, it possesses a Jordan normal form J=P1APJ = P^{-1}AP, and the matrix exponential involves polynomials in tt multiplied by exponentials.

For non-autonomous systems x=A(t)x\mathbf{x}' = A(t)\mathbf{x}, the solution is expressed through the fundamental matrix Φ(t)\Phi(t), which satisfies Φ=A(t)Φ\Phi' = A(t)\Phi with Φ(0)=I\Phi(0) = I. The general solution is x(t)=Φ(t)x0\mathbf{x}(t) = \Phi(t)\mathbf{x}_0. When a forcing term is added, giving x=A(t)x+b(t)\mathbf{x}' = A(t)\mathbf{x} + \mathbf{b}(t), the variation of parameters formula (or Duhamel’s principle) gives the particular solution:

x(t)=Φ(t)x0+Φ(t)0tΦ(s)1b(s)ds.\mathbf{x}(t) = \Phi(t)\mathbf{x}_0 + \Phi(t) \int_0^t \Phi(s)^{-1} \mathbf{b}(s)\, ds.

Higher-order scalar equations also fit into this framework. The equation y(n)+an1y(n1)++a1y+a0y=g(t)y^{(n)} + a_{n-1} y^{(n-1)} + \cdots + a_1 y' + a_0 y = g(t) is converted to a first-order system by setting x1=yx_1 = y, x2=yx_2 = y', and so on, yielding a system with the companion matrix of the characteristic polynomial λn+an1λn1++a0\lambda^n + a_{n-1}\lambda^{n-1} + \cdots + a_0. The characteristic roots λi\lambda_i determine the nature of solutions entirely: real roots give exponential behaviour, complex roots give oscillatory behaviour, and the multiplicity of a root determines whether polynomial factors tkt^k appear.

Phase Portraits and Qualitative Analysis

For an autonomous two-dimensional system x=F(x)\mathbf{x}' = \mathbf{F}(\mathbf{x}) (where F\mathbf{F} does not depend explicitly on tt), the phase portrait is the collection of all trajectories in the xyxy-plane. Each trajectory is a solution curve parameterized by time, and the phase portrait reveals the global geometric structure of the dynamical system at a glance.

Equilibrium points (also called fixed points or critical points) are solutions x\mathbf{x}^* satisfying F(x)=0\mathbf{F}(\mathbf{x}^*) = \mathbf{0}. Near an isolated equilibrium, the behaviour of trajectories is determined by the linearization: replace F\mathbf{F} by its Jacobian matrix J=DF(x)J = D\mathbf{F}(\mathbf{x}^*) evaluated at the equilibrium. The eigenvalues of JJ classify the local type of the equilibrium:

  • If both eigenvalues are real and negative: stable node (all trajectories approach x\mathbf{x}^*).
  • If both are real and positive: unstable node (all trajectories recede from x\mathbf{x}^*).
  • If one eigenvalue is positive and one negative: saddle point (stable and unstable manifolds cross at x\mathbf{x}^*).
  • If eigenvalues are complex α±iβ\alpha \pm i\beta with α<0\alpha < 0: stable spiral (trajectories spiral in).
  • If α>0\alpha > 0: unstable spiral.
  • If α=0\alpha = 0: center (closed orbits in the linearization; nonlinear behaviour requires higher-order analysis).

The Hartman-Grobman theorem makes the linearization rigorous: near a hyperbolic equilibrium (one where no eigenvalue has zero real part), the nonlinear flow is topologically conjugate to the linear flow of its Jacobian. In other words, the phase portrait of the nonlinear system is homeomorphically equivalent to that of the linearization near hyperbolic fixed points.

Beyond equilibria, limit cycles are isolated closed trajectories — periodic orbits that are not part of a continuous family of closed orbits. The van der Pol oscillator x¨μ(1x2)x˙+x=0\ddot{x} - \mu(1-x^2)\dot{x} + x = 0 (introduced by Balthasar van der Pol in 1920 while studying vacuum tube circuits) is the archetype: for any μ>0\mu > 0, it possesses a unique, globally attracting limit cycle. The Bendixson-Dulac criterion provides a sufficient condition for the absence of limit cycles in a region: if (hF)\nabla \cdot (h\mathbf{F}) does not change sign in a simply connected region (for some smooth hh), then no closed orbit lies entirely within it. The Poincaré-Bendixson theorem is its constructive companion: if a trajectory in R2\mathbb{R}^2 is bounded and its ω\omega-limit set contains no equilibria, then that ω\omega-limit set must be a limit cycle.

Global objects in the phase portrait include stable and unstable manifolds of saddle points, which partition the phase plane into regions of qualitatively different behaviour, homoclinic orbits (trajectories that connect a saddle to itself), and heteroclinic orbits (connecting two distinct saddle points). These global invariant manifolds are the “skeletons” around which the rest of the dynamics organizes itself.

Stability Theory

The informal question “do solutions eventually settle down near the equilibrium?” is made precise by Lyapunov stability theory, developed by Aleksandr Lyapunov in his 1892 doctoral thesis — one of the most influential documents in applied mathematics.

An equilibrium x\mathbf{x}^* is stable in the sense of Lyapunov if for every ε>0\varepsilon > 0 there exists δ>0\delta > 0 such that x(0)x<δ|\mathbf{x}(0) - \mathbf{x}^*| < \delta implies x(t)x<ε|\mathbf{x}(t) - \mathbf{x}^*| < \varepsilon for all t0t \ge 0. It is asymptotically stable if it is stable and there exists δ>0\delta > 0 such that x(0)x<δ|\mathbf{x}(0) - \mathbf{x}^*| < \delta implies x(t)x\mathbf{x}(t) \to \mathbf{x}^* as tt \to \infty. It is globally asymptotically stable if the convergence holds for all initial data.

The power of Lyapunov’s approach is that it can certify stability without solving the ODE explicitly. A Lyapunov function is a smooth function V:U[0,)V : U \to [0, \infty) (for some neighbourhood UU of x\mathbf{x}^*) satisfying:

  1. V(x)=0V(\mathbf{x}^*) = 0 and V(x)>0V(\mathbf{x}) > 0 for xx\mathbf{x} \ne \mathbf{x}^* (positive definiteness),
  2. V˙(x)=V(x)F(x)0\dot{V}(\mathbf{x}) = \nabla V(\mathbf{x}) \cdot \mathbf{F}(\mathbf{x}) \le 0 along every trajectory (non-increase of VV).

If V˙0\dot{V} \le 0, the equilibrium is stable; if V˙<0\dot{V} < 0 (negative definiteness), it is asymptotically stable. The function VV plays the role of an abstract energy: the system can only stay the same or “lose energy” over time. For the pendulum, the physical energy E=12θ˙2+(1cosθ)E = \frac{1}{2}\dot\theta^2 + (1 - \cos\theta) is a Lyapunov function; its non-increase along trajectories confirms stability of the rest position.

Constructing Lyapunov functions is an art rather than a science: there is no general algorithm. For linear systems x=Ax\mathbf{x}' = A\mathbf{x}, if all eigenvalues of AA have negative real parts, then the Lyapunov equation ATP+PA=QA^T P + PA = -Q has a unique positive-definite solution PP for any positive-definite QQ, and V(x)=xTPxV(\mathbf{x}) = \mathbf{x}^T P \mathbf{x} is a Lyapunov function. For nonlinear systems, quadratic functions, logarithmic functions, and combinations tailored to the specific vector field must be found heuristically. LaSalle’s invariance principle is a powerful refinement: even if V˙0\dot{V} \le 0 rather than V˙<0\dot{V} < 0, trajectories converge to the largest invariant subset of {V˙=0}\{\dot{V} = 0\}, which can be shown to be {x}\{\mathbf{x}^*\} in many cases, still giving asymptotic stability.

The linearization stability theorem unifies the two approaches: for a hyperbolic equilibrium, the eigenvalue sign conditions of the Jacobian give the same conclusion as any Lyapunov analysis. Stability at non-hyperbolic equilibria (where some eigenvalue has zero real part) is genuinely harder and requires either normal forms, center manifold reduction, or a direct Lyapunov construction.

Sturm-Liouville Theory and Boundary Value Problems

A qualitatively different type of problem arises when conditions are imposed at two separate points rather than at a single initial time. A two-point boundary value problem (BVP) consists of a second-order ODE together with conditions at both endpoints of an interval [a,b][a, b].

The central framework is the Sturm-Liouville problem: find functions yy and numbers λ\lambda satisfying

ddx ⁣[p(x)y]q(x)y+λw(x)y=0,x[a,b],\frac{d}{dx}\!\left[p(x) y'\right] - q(x) y + \lambda w(x) y = 0, \quad x \in [a,b],

subject to separated boundary conditions α1y(a)+α2y(a)=0\alpha_1 y(a) + \alpha_2 y'(a) = 0 and β1y(b)+β2y(b)=0\beta_1 y(b) + \beta_2 y'(b) = 0, where p>0p > 0, w>0w > 0, and pp, qq, ww are sufficiently smooth. This problem was introduced by Jacques Charles François Sturm and Joseph Liouville in a celebrated series of memoirs published between 1836 and 1838 — a landmark moment in the development of spectral theory.

The numbers λ\lambda for which nontrivial solutions exist are called eigenvalues and the corresponding solutions yny_n are eigenfunctions. The fundamental results are:

  • The eigenvalues form an infinite sequence λ1<λ2<λ3<+\lambda_1 < \lambda_2 < \lambda_3 < \cdots \to +\infty.
  • The eigenfunctions are orthogonal with respect to the weight ww:

abym(x)yn(x)w(x)dx=0for mn.\int_a^b y_m(x)\, y_n(x)\, w(x)\, dx = 0 \quad \text{for } m \ne n.

  • The eigenfunctions are complete: any sufficiently regular function ff on [a,b][a,b] can be expanded in a generalized Fourier series f(x)=n=1cnyn(x)f(x) = \sum_{n=1}^\infty c_n y_n(x), with coefficients cn=f,ynw/ynw2c_n = \langle f, y_n \rangle_w / \|y_n\|_w^2, and the series converges in the weighted L2L^2 sense.
  • The nn-th eigenfunction has exactly n1n-1 zeros in the open interval (a,b)(a,b) — the Sturm oscillation theorem.

The Sturm-Liouville framework unifies a host of classical special function theories. The Bessel equation x2y+xy+(λ2x2n2)y=0x^2 y'' + x y' + (\lambda^2 x^2 - n^2) y = 0 is a singular Sturm-Liouville problem (singular because p(0)=0p(0) = 0) whose eigenfunctions are Bessel functions JnJ_n, essential in problems with cylindrical symmetry. The Legendre equation (1x2)y2xy+(+1)y=0(1-x^2)y'' - 2x y' + \ell(\ell+1) y = 0 on [1,1][-1,1] is another singular problem, with Legendre polynomials PP_\ell as eigenfunctions, fundamental to problems with spherical symmetry. These arise naturally when separation of variables is applied to PDEs such as Laplace’s, heat, and wave equations.

When a nonhomogeneous source h(x)h(x) is present, the solution of Ly=hLy = h (where LL is the Sturm-Liouville operator) is written using a Green’s function G(x,ξ)G(x, \xi):

y(x)=abG(x,ξ)h(ξ)dξ.y(x) = \int_a^b G(x, \xi)\, h(\xi)\, d\xi.

The Green’s function is constructed by piecing together solutions of the homogeneous equation on either side of ξ\xi, with a jump discontinuity in the derivative at x=ξx = \xi determined by the source singularity. Green’s functions are the ODE-level precursors of the integral operators that become central in functional analysis and the study of PDEs.

Bifurcation Theory and Special Functions

In many physical models, the equations governing a system depend on a parameter μ\mu (temperature, flow speed, coupling strength). As μ\mu varies, equilibria can appear, disappear, merge, or change stability. A bifurcation is a qualitative change in the structure of the solution set as μ\mu passes through a critical value μc\mu_c, called the bifurcation point.

The catalogue of local bifurcations is organized by normal forms — simplified model equations that capture the essential dynamics near each bifurcation type:

The saddle-node bifurcation is the most generic: two equilibria (one stable, one unstable) collide and annihilate as μ\mu decreases past μc\mu_c. The normal form is y=μy2y' = \mu - y^2. For μ>0\mu > 0 there are two equilibria y=±μy = \pm\sqrt{\mu}; at μ=0\mu = 0 they merge; for μ<0\mu < 0 there are none. This bifurcation underlies the sudden “tipping point” transitions observed in climate science, ecology, and economics.

The transcritical bifurcation occurs when two equilibria cross and exchange stability; its normal form is y=μyy2y' = \mu y - y^2. Both equilibria persist for all μ\mu, but at μ=0\mu = 0 they swap their stability character. The logistic model of population growth features this structure.

The pitchfork bifurcation appears in systems with a symmetry that prevents the saddle-node: the normal form y=μyy3y' = \mu y - y^3 has a single equilibrium y=0y = 0 for μ<0\mu < 0, and for μ>0\mu > 0 the origin becomes unstable while two new symmetric equilibria y=±μy = \pm\sqrt{\mu} are born (the supercritical or forward pitchfork). Buckling of an elastic beam is the prototypical example.

The most dynamically rich local bifurcation is the Hopf bifurcation, formalized by Eberhard Hopf in 1942. Here an equilibrium loses stability as a pair of complex conjugate eigenvalues of the Jacobian crosses the imaginary axis. Simultaneously, a limit cycle is born (in the supercritical case) or annihilated (in the subcritical case). The frequency of the newborn oscillation near the bifurcation is approximately β=Im(λ)\beta = \mathrm{Im}(\lambda) at μ=μc\mu = \mu_c, and its amplitude grows like μμc\sqrt{|\mu - \mu_c|}. The Hopf bifurcation is the mathematical explanation for spontaneous oscillations in biological clocks, chemical reactions (the Belousov-Zhabotinsky reaction), and fluid dynamics (the von Kármán vortex street).

Global bifurcations involve large-scale changes in the phase portrait that cannot be detected by local analysis near a single equilibrium. In a homoclinic bifurcation, a limit cycle grows until it collides with a saddle point, forming a homoclinic orbit at the critical parameter value; for μ\mu beyond this value, the limit cycle disappears. Heteroclinic bifurcations involve connections between distinct saddle points. These global events can trigger transitions between qualitatively different modes of behaviour — from periodic oscillation to chaos — making their analysis essential in understanding complex physical systems.

The study of how families of ODEs depend on parameters is now the province of dynamical systems theory, a field shaped decisively by Poincaré’s geometric vision, by Birkhoff’s work in the 1920s and 1930s, and by the flowering of chaos theory after Lorenz’s 1963 discovery of sensitive dependence on initial conditions in a three-dimensional ODE. Modern bifurcation theory, backed by center manifold theorems, normal form algorithms, and numerical continuation software, remains one of the most active and applicable areas of mathematics.