Partial Differential Equations
Elliptic, parabolic, and hyperbolic equations — Sobolev spaces and weak solutions.
Partial differential equations are the mathematical language through which the physical world speaks: heat spreads, waves propagate, electric fields arrange themselves, and fluids flow according to laws that are fundamentally equations relating a function to its partial derivatives. Unlike ordinary differential equations, which describe how a quantity evolves along a single dimension, PDEs govern phenomena that vary simultaneously across space and time, making them both incomparably richer and far more difficult to solve. Their study has driven the development of much of modern analysis, from Fourier’s decomposition of functions into sine waves to the Sobolev spaces and weak solution theory that underpin the entire contemporary framework.
Classification of PDEs
Before one can hope to solve a PDE, one must understand what kind of equation one is dealing with. The classification of PDEs is not a purely aesthetic exercise — different types demand fundamentally different methods, encode different physical intuitions, and possess different qualitative properties.
A partial differential equation involves an unknown function of several variables and its partial derivatives up to some order. A second-order linear PDE in two variables and has the general form
where the coefficients may depend on and . The behavior of such an equation is governed by the discriminant , borrowed from the classification of conic sections. When , the equation is elliptic; when , it is parabolic; when , it is hyperbolic. The canonical examples are, respectively, the Laplace equation , the heat equation , and the wave equation .
This classification extends to higher dimensions and variable coefficients. For a general second-order operator , the type at a point is determined by the eigenvalues of the coefficient matrix : the equation is elliptic if all eigenvalues have the same sign, hyperbolic if one eigenvalue has the opposite sign from the rest, and parabolic if one eigenvalue vanishes. The deeper reason this classification matters is that each type corresponds to a distinct information-propagation structure. Elliptic equations encode equilibrium states where influence is felt everywhere simultaneously. Parabolic equations describe irreversible diffusion with infinite propagation speed. Hyperbolic equations model wave-like phenomena with a finite speed of propagation — information travels only within a cone, called the light cone or domain of dependence.
Well-posedness, formalized by Jacques Hadamard around 1902, requires that a problem possess a solution, that the solution be unique, and that it depend continuously on the data. Each PDE type pairs naturally with a specific class of auxiliary conditions: elliptic equations are paired with boundary conditions on a closed domain, parabolic equations require an initial condition plus boundary conditions, and hyperbolic equations typically require initial conditions on position and velocity. Matching the wrong conditions to the wrong type — such as imposing an initial value problem on Laplace’s equation — leads to the spectacularly ill-posed backward heat equation, where arbitrarily small perturbations in the data produce unboundedly large changes in the solution.
The method of characteristics provides the deepest geometric insight into this classification for first-order and hyperbolic equations. Characteristics are curves (or surfaces) along which information propagates. For the first-order equation , characteristics are the integral curves of the vector field , and the solution is constant along each characteristic. For hyperbolic second-order equations, two families of real characteristics exist and govern the finite-speed propagation of singularities. For elliptic equations, characteristics are complex-valued — there are no real curves along which data propagates, which is why boundary data on the entire boundary must be specified.
Elliptic Equations: Laplace, Poisson, and General Theory
Elliptic equations describe systems in equilibrium. When a thin conducting plate reaches thermal equilibrium, its temperature satisfies Laplace’s equation , where is the Laplace operator. Solutions to Laplace’s equation are called harmonic functions, and they possess extraordinary regularity: every harmonic function is real-analytic. The Poisson equation is the non-homogeneous version, describing, for instance, the electrostatic potential generated by a charge distribution .
The first remarkable property of harmonic functions is the mean value property: the value of a harmonic function at any point equals the average of its values on any sphere centered at that point,
where is a ball of radius centered at . This single identity implies an extraordinary wealth of consequences. The maximum principle follows immediately: a harmonic function attains its maximum and minimum on the boundary of any bounded domain, never in the interior. This principle provides uniqueness for the Dirichlet problem — the problem of finding with in a domain and on — because if and are two solutions, their difference is harmonic and zero on the boundary, hence zero everywhere. The French mathematician Siméon Denis Poisson gave the explicit formula for the solution in the disk in the early nineteenth century, and Bernhard Riemann elevated the Dirichlet problem to a central place in function theory through his use of Dirichlet’s principle.
The fundamental tool for representing solutions is the Green’s function for the Laplacian on a domain . The Green’s function encodes how the solution at depends on the source at and satisfies with for . Using Green’s representation formula, any harmonic function in can be written as
where is the outward normal at . In free space with , the fundamental solution is , where is a dimensional constant; in it is .
For general elliptic operators beyond the Laplacian, the modern approach relies on the Lax-Milgram theorem from functional analysis. Given a bilinear form that is continuous and coercive on a Hilbert space , Lax-Milgram guarantees a unique solution to for all test functions . This abstract framework simultaneously provides existence and uniqueness for a broad class of elliptic boundary value problems. The Fredholm alternative handles the more delicate case where coercivity fails, establishing that either the homogeneous problem has only the trivial solution (and the inhomogeneous problem has a unique solution for every right-hand side) or the homogeneous problem has nontrivial solutions (and the inhomogeneous problem has solutions only for right-hand sides orthogonal to those nontrivial solutions).
Regularity theory for elliptic equations is one of the crowning achievements of twentieth-century analysis. The central result of elliptic regularity states that if is a weak solution of and if (the Sobolev space of functions with square-integrable derivatives), then , gaining two derivatives relative to the data. By iterating this bootstrapping argument, a solution with smooth data becomes smooth itself. Near the boundary, additional regularity requires the domain boundary to be sufficiently smooth. The Schauder estimates provide an analogous theory in Holder spaces, asserting that if then .
Parabolic Equations and the Heat Equation
The heat equation was introduced by Joseph Fourier in his 1822 masterpiece Théorie analytique de la chaleur. Here is temperature at position and time , and is thermal diffusivity. To solve it, Fourier decomposed the initial temperature distribution into sinusoidal modes and showed that each mode decays exponentially in time — an idea that gave birth to Fourier analysis. The heat equation is the prototypical parabolic PDE, and its study illuminates the entire parabolic theory.
The solution to the Cauchy problem (initial data on all of ) is given by convolution with the heat kernel:
The heat kernel is a Gaussian that broadens and flattens as increases, expressing the physical spreading of heat. Several qualitative features stand out immediately. First, the solution is instantly smooth for no matter how rough the initial data — a smoothing effect absent from hyperbolic equations. Second, and perhaps surprisingly, the solution at any point with depends on the initial data at every point : perturbations propagate at infinite speed, in stark contrast to wave equations. This is a mathematical artifact of the idealized model; real heat conduction is governed by quantum mechanics at short scales.
For bounded domains, the method of separation of variables and eigenfunction expansion provides explicit solutions. Seeking in with on reduces the heat equation to two ODEs: and . The second is the eigenvalue problem for the Laplacian, which has a sequence of eigenvalues with corresponding eigenfunctions that form a complete orthonormal basis of . The solution is then
Each mode decays exponentially, with the lowest eigenvalue governing the long-time behavior: as .
The parabolic maximum principle asserts that the maximum of a solution to over the parabolic boundary (the bottom and sides of a space-time cylinder) is also the maximum over the entire closed cylinder. This powerful principle governs comparison, uniqueness, and stability: if two solutions agree at and on for all , they agree everywhere. Duhamel’s principle extends the framework to non-homogeneous equations : the solution is built by treating each instantaneous source as an initial condition for a separate heat equation started at time , then integrating over .
The backward heat equation is the time-reversal of the heat equation and is catastrophically ill-posed in the sense of Hadamard. The reason is that the forwards heat equation destroys information by smoothing, so reversing time requires reconstructing information from a smoothed state — an exponentially unstable operation. This ill-posedness is of practical importance in inverse problems, where one wishes to recover a past temperature distribution from present measurements.
Hyperbolic Equations and Wave Phenomena
The wave equation governs vibrating strings (), sound (), and electromagnetic radiation (). Here is the propagation speed. Unlike the heat equation, the wave equation is time-reversible and preserves information: what is true now was also true in the past, at least in the classical setting. The wave equation was studied intensely in the eighteenth century by Jean le Rond d’Alembert, Leonhard Euler, and Daniel Bernoulli, who disagreed vigorously about the nature of its solutions — a dispute that helped crystallize the modern concept of a function.
In one spatial dimension, d’Alembert’s formula gives the complete solution to the initial value problem , :
This formula makes transparent the defining feature of hyperbolic equations: the domain of dependence. The value depends only on initial data in the interval , the segment swept out by characteristics running backward from at speed . Any perturbation of data outside this interval has no effect on . Conversely, the domain of influence of a point is the cone — the set of space-time points that can be affected by data at .
In three spatial dimensions, the solution is given by Kirchhoff’s formula:
A crucial observation: in three dimensions, the value depends only on data on the sphere , not on data inside the sphere. This is Huygens’ principle — sharp signals in three dimensions remain sharp, whereas in two dimensions (and other even dimensions), a signal spreads into a trailing wake. The reason Huygens’ principle holds in odd dimensions but fails in even ones is one of the more beautiful results in PDE theory, connected to the theory of the wave operator in different dimensions.
The wave equation conserves a natural energy. For the homogeneous wave equation, the quantity
satisfies , so for all time. Energy conservation provides uniqueness (if two solutions have the same initial data, their difference has zero energy, hence is zero) and stability. Nonlinear hyperbolic equations, such as conservation laws of the form , exhibit more complex behavior: smooth initial data can develop jump discontinuities (shocks) in finite time. The Rankine-Hugoniot condition governs the speed of a shock, and entropy conditions select the physically meaningful weak solution among multiple candidates.
Variational Methods and Weak Solutions
The history of variational methods in PDE theory begins with Dirichlet’s principle, which asserts that the harmonic function on a domain with prescribed boundary values is the function minimizing the Dirichlet energy
among all functions agreeing with on . Riemann used this principle freely, but Weierstrass showed in 1870 that a minimizing sequence need not converge — the infimum might not be attained. This crisis was resolved by David Hilbert around 1900 through the development of functional analysis, culminating in the direct method in the calculus of variations: one shows that a minimizing sequence is bounded in a suitable function space, extracts a weakly convergent subsequence by compactness, and then uses lower semicontinuity of the energy to confirm that the limit is indeed a minimizer.
The notion of a weak solution is the central conceptual innovation that enables modern PDE theory. Classical solutions require enough regularity to substitute directly into the PDE; weak solutions broaden the solution concept by moving derivatives onto smooth test functions through integration by parts. For the Poisson equation , the weak formulation asks for (functions with one weak derivative that vanish on , in the sense) such that
This formulation makes sense even when is merely in and lacks the two classical derivatives needed to write pointwise. The Lax-Milgram theorem guarantees existence and uniqueness of weak solutions for a broad class of elliptic problems, and elliptic regularity then shows that the weak solution is as smooth as the data permit.
The Euler-Lagrange equation connects variational problems to PDEs. If minimizes a functional of the form , then satisfies . For the Dirichlet energy, , giving . For the minimal surface functional , the Euler-Lagrange equation becomes the minimal surface equation
a quasi-linear elliptic PDE. Weak formulations and variational methods unite large swaths of PDE theory into a single coherent framework, providing existence results for elliptic, parabolic, and even hyperbolic problems in energy spaces naturally suited to each type.
Function Spaces and Regularity Theory
The modern treatment of PDEs is inseparable from a hierarchy of function spaces designed to measure the regularity of solutions. At the top of the hierarchy sit the classical smooth spaces and Holder spaces ; at the foundational level sit the Sobolev spaces , which are the workhorses of contemporary PDE analysis.
The Sobolev space consists of all functions in whose weak derivatives up to order also lie in , equipped with the norm
The space is a Hilbert space, making it particularly amenable to the techniques of functional analysis. The weak derivative is defined by the integration-by-parts identity for all test functions , bypassing any requirement that be classically differentiable. The weak derivative was developed independently by Sergei Sobolev and Jean Leray in the 1930s, largely motivated by the needs of fluid mechanics and the Navier-Stokes equations.
Sobolev embedding theorems are among the deepest and most useful results in analysis. They assert that, under appropriate dimensional conditions, membership in implies pointwise regularity. The critical case is governed by the Sobolev exponent for : the Sobolev embedding holds continuously. When , functions in are Holder continuous: for appropriate . These embeddings explain why solutions to elliptic equations gain regularity: if and satisfies an elliptic equation, then elliptic regularity lifts to , which by Sobolev embedding may already be continuous.
Distribution theory, developed systematically by Laurent Schwartz in the 1940s (for which he received the Fields Medal in 1950), provides the broadest possible framework for generalized functions. A distribution is a continuous linear functional on the space of test functions. Every locally integrable function defines a distribution via , but there are distributions — such as the Dirac delta and its derivatives — that correspond to no function at all. The power of distributions is that every distribution can be differentiated arbitrarily many times: is the distribution defined by . This allows one to speak rigorously of the Laplacian of a Green’s function or the derivative of a shock wave, without any classical pointwise meaning.
The trace theorem handles boundary values in the Sobolev framework. Functions in need not be continuous up to , so their boundary values cannot be defined pointwise. Nevertheless, the trace operator is a well-defined, bounded linear map that extends the restriction map from smooth functions. This allows boundary conditions to be imposed in the weak formulation in a mathematically rigorous way.
Compact embeddings — specifically the Rellich-Kondrachov theorem, which asserts that is compact when is bounded — are essential for extracting convergent subsequences in existence proofs. This compactness result is the functional-analytic backbone of the direct method: a bounded sequence in has a subsequence converging strongly in , which is precisely the convergence needed to pass limits through nonlinear terms.
Green’s Functions and Numerical Methods
Green’s functions are the quintessential tool for converting PDE boundary value problems into explicit integral representations. For a linear differential operator on a domain with specified boundary conditions, the Green’s function is the response of the system at to a unit point source at . Once is known, the solution to is
Green’s functions were introduced by the self-taught British mathematician George Green in his 1828 Essay on the Application of Mathematical Analysis to the Theories of Electricity and Magnetism, a work that was largely ignored for two decades before being recognized as foundational. The construction of the Green’s function for specific domains exploits symmetry and the method of images: for the Laplacian on the half-space , the Green’s function is , where is the reflection of across the boundary, so that vanishes on .
For the heat equation, the fundamental solution plays the role of the Green’s function on all of . More generally, the heat kernel on a bounded domain carries geometric information about : the spectrum of the Laplacian can be recovered from the short-time asymptotics of , a connection captured by the celebrated formula of Mark Kac: “Can one hear the shape of a drum?” — can one determine from the eigenvalues of its Laplacian? The answer, provided by Carolyn Gordon, David Webb, and Scott Wolpert in 1992, is no: non-isometric domains can share the same eigenvalue spectrum.
When explicit Green’s functions are unavailable — which is the generic case for irregular domains or variable-coefficient operators — numerical methods take over. The three main classes are finite difference methods, finite element methods, and spectral methods. Finite difference methods replace derivatives by difference quotients on a grid: . They are simple to implement and analyze, and the Courant-Friedrichs-Lewy (CFL) condition dictates the stability constraint for explicit schemes applied to hyperbolic equations: the time step must satisfy , ensuring that the numerical domain of dependence contains the true domain of dependence.
Finite element methods (FEM) work directly with the weak formulation. One approximates in a finite-dimensional subspace (typically piecewise polynomial functions on a triangulation of ) and finds satisfying the weak equation for all . Galerkin’s method — named after Boris Galerkin, who developed it for structural mechanics in 1915 — yields a linear system where is the stiffness matrix with entries . The method is supported by a complete error theory: Cea’s lemma shows that the FEM approximation is quasi-optimal, , and polynomial approximation theory then gives convergence rates in terms of mesh size and polynomial degree. Adaptive refinement driven by a posteriori error estimators allows one to concentrate computational effort where the solution is least regular.
Spectral methods expand the solution in a globally smooth basis — Fourier modes or Chebyshev polynomials — and achieve exponential convergence rates for smooth solutions, far surpassing the algebraic rates of finite differences and finite elements. They are most effective for problems on simple geometries with smooth data, such as in weather prediction, fluid simulations, and quantum mechanics.
The interplay of analysis and computation that PDE theory demands has made it one of the richest and most active areas of mathematics. From Fourier’s original heat equation to the Navier-Stokes equations governing turbulent flow (whose global regularity remains one of the Clay Millennium Prize Problems) and to the emerging field of physics-informed neural networks, the subject continues to grow, always driven by the same essential tension: between the physical world’s inexhaustible complexity and mathematics’ drive to find structure within it.