Partial Differential Equations

Elliptic, parabolic, and hyperbolic equations — Sobolev spaces and weak solutions.


Partial differential equations are the mathematical language through which the physical world speaks: heat spreads, waves propagate, electric fields arrange themselves, and fluids flow according to laws that are fundamentally equations relating a function to its partial derivatives. Unlike ordinary differential equations, which describe how a quantity evolves along a single dimension, PDEs govern phenomena that vary simultaneously across space and time, making them both incomparably richer and far more difficult to solve. Their study has driven the development of much of modern analysis, from Fourier’s decomposition of functions into sine waves to the Sobolev spaces and weak solution theory that underpin the entire contemporary framework.

Classification of PDEs

Before one can hope to solve a PDE, one must understand what kind of equation one is dealing with. The classification of PDEs is not a purely aesthetic exercise — different types demand fundamentally different methods, encode different physical intuitions, and possess different qualitative properties.

A partial differential equation involves an unknown function uu of several variables x1,x2,,xnx_1, x_2, \ldots, x_n and its partial derivatives up to some order. A second-order linear PDE in two variables xx and yy has the general form

Auxx+2Buxy+Cuyy+Dux+Euy+Fu=G,A u_{xx} + 2B u_{xy} + C u_{yy} + D u_x + E u_y + F u = G,

where the coefficients A,B,C,D,E,F,GA, B, C, D, E, F, G may depend on xx and yy. The behavior of such an equation is governed by the discriminant Δ=B2AC\Delta = B^2 - AC, borrowed from the classification of conic sections. When Δ<0\Delta < 0, the equation is elliptic; when Δ=0\Delta = 0, it is parabolic; when Δ>0\Delta > 0, it is hyperbolic. The canonical examples are, respectively, the Laplace equation uxx+uyy=0u_{xx} + u_{yy} = 0, the heat equation ut=uxxu_t = u_{xx}, and the wave equation utt=uxxu_{tt} = u_{xx}.

This classification extends to higher dimensions and variable coefficients. For a general second-order operator Lu=i,jaij(x)iju+lower-order terms\mathcal{L}u = \sum_{i,j} a_{ij}(x) \partial_i \partial_j u + \text{lower-order terms}, the type at a point xx is determined by the eigenvalues of the coefficient matrix (aij)(a_{ij}): the equation is elliptic if all eigenvalues have the same sign, hyperbolic if one eigenvalue has the opposite sign from the rest, and parabolic if one eigenvalue vanishes. The deeper reason this classification matters is that each type corresponds to a distinct information-propagation structure. Elliptic equations encode equilibrium states where influence is felt everywhere simultaneously. Parabolic equations describe irreversible diffusion with infinite propagation speed. Hyperbolic equations model wave-like phenomena with a finite speed of propagation — information travels only within a cone, called the light cone or domain of dependence.

Well-posedness, formalized by Jacques Hadamard around 1902, requires that a problem possess a solution, that the solution be unique, and that it depend continuously on the data. Each PDE type pairs naturally with a specific class of auxiliary conditions: elliptic equations are paired with boundary conditions on a closed domain, parabolic equations require an initial condition plus boundary conditions, and hyperbolic equations typically require initial conditions on position and velocity. Matching the wrong conditions to the wrong type — such as imposing an initial value problem on Laplace’s equation — leads to the spectacularly ill-posed backward heat equation, where arbitrarily small perturbations in the data produce unboundedly large changes in the solution.

The method of characteristics provides the deepest geometric insight into this classification for first-order and hyperbolic equations. Characteristics are curves (or surfaces) along which information propagates. For the first-order equation a(x,y)ux+b(x,y)uy=ca(x,y) u_x + b(x,y) u_y = c, characteristics are the integral curves of the vector field (a,b)(a, b), and the solution is constant along each characteristic. For hyperbolic second-order equations, two families of real characteristics exist and govern the finite-speed propagation of singularities. For elliptic equations, characteristics are complex-valued — there are no real curves along which data propagates, which is why boundary data on the entire boundary must be specified.

Elliptic Equations: Laplace, Poisson, and General Theory

Elliptic equations describe systems in equilibrium. When a thin conducting plate reaches thermal equilibrium, its temperature satisfies Laplace’s equation Δu=0\Delta u = 0, where Δ=x12++xn2\Delta = \partial_{x_1}^2 + \cdots + \partial_{x_n}^2 is the Laplace operator. Solutions to Laplace’s equation are called harmonic functions, and they possess extraordinary regularity: every harmonic function is real-analytic. The Poisson equation Δu=f\Delta u = f is the non-homogeneous version, describing, for instance, the electrostatic potential generated by a charge distribution ff.

The first remarkable property of harmonic functions is the mean value property: the value of a harmonic function at any point equals the average of its values on any sphere centered at that point,

u(x)=1Br(x)Br(x)udS,u(x) = \frac{1}{|\partial B_r(x)|} \int_{\partial B_r(x)} u \, dS,

where Br(x)B_r(x) is a ball of radius rr centered at xx. This single identity implies an extraordinary wealth of consequences. The maximum principle follows immediately: a harmonic function attains its maximum and minimum on the boundary of any bounded domain, never in the interior. This principle provides uniqueness for the Dirichlet problem — the problem of finding uu with Δu=0\Delta u = 0 in a domain Ω\Omega and u=gu = g on Ω\partial\Omega — because if uu and vv are two solutions, their difference w=uvw = u - v is harmonic and zero on the boundary, hence zero everywhere. The French mathematician Siméon Denis Poisson gave the explicit formula for the solution in the disk in the early nineteenth century, and Bernhard Riemann elevated the Dirichlet problem to a central place in function theory through his use of Dirichlet’s principle.

The fundamental tool for representing solutions is the Green’s function G(x,y)G(x, y) for the Laplacian on a domain Ω\Omega. The Green’s function encodes how the solution at xx depends on the source at yy and satisfies ΔxG(x,y)=δ(xy)-\Delta_x G(x, y) = \delta(x - y) with G(x,y)=0G(x, y) = 0 for xΩx \in \partial\Omega. Using Green’s representation formula, any harmonic function in Ω\Omega can be written as

u(x)=Ωu(y)Gνy(x,y)dS(y)+ΩG(x,y)f(y)dy,u(x) = -\int_{\partial\Omega} u(y) \frac{\partial G}{\partial \nu_y}(x, y) \, dS(y) + \int_\Omega G(x, y) f(y) \, dy,

where νy\nu_y is the outward normal at yΩy \in \partial\Omega. In free space Rn\mathbb{R}^n with n3n \geq 3, the fundamental solution is Φ(x)=cnx2n\Phi(x) = c_n |x|^{2-n}, where cnc_n is a dimensional constant; in R2\mathbb{R}^2 it is Φ(x)=12πlogx\Phi(x) = -\frac{1}{2\pi} \log|x|.

For general elliptic operators beyond the Laplacian, the modern approach relies on the Lax-Milgram theorem from functional analysis. Given a bilinear form a(u,v)a(u, v) that is continuous and coercive on a Hilbert space HH, Lax-Milgram guarantees a unique solution uHu \in H to a(u,v)=f,va(u, v) = \langle f, v \rangle for all test functions vv. This abstract framework simultaneously provides existence and uniqueness for a broad class of elliptic boundary value problems. The Fredholm alternative handles the more delicate case where coercivity fails, establishing that either the homogeneous problem has only the trivial solution (and the inhomogeneous problem has a unique solution for every right-hand side) or the homogeneous problem has nontrivial solutions (and the inhomogeneous problem has solutions only for right-hand sides orthogonal to those nontrivial solutions).

Regularity theory for elliptic equations is one of the crowning achievements of twentieth-century analysis. The central result of elliptic regularity states that if uu is a weak solution of Lu=f\mathcal{L}u = f and if fHk(Ω)f \in H^k(\Omega) (the Sobolev space of functions with kk square-integrable derivatives), then uHk+2(Ω)u \in H^{k+2}(\Omega), gaining two derivatives relative to the data. By iterating this bootstrapping argument, a solution with smooth data becomes smooth itself. Near the boundary, additional regularity requires the domain boundary Ω\partial\Omega to be sufficiently smooth. The Schauder estimates provide an analogous theory in Holder spaces, asserting that if fCk,αf \in C^{k,\alpha} then uCk+2,αu \in C^{k+2,\alpha}.

Parabolic Equations and the Heat Equation

The heat equation ut=kΔuu_t = k \Delta u was introduced by Joseph Fourier in his 1822 masterpiece Théorie analytique de la chaleur. Here u(x,t)u(x, t) is temperature at position xx and time tt, and k>0k > 0 is thermal diffusivity. To solve it, Fourier decomposed the initial temperature distribution into sinusoidal modes and showed that each mode decays exponentially in time — an idea that gave birth to Fourier analysis. The heat equation is the prototypical parabolic PDE, and its study illuminates the entire parabolic theory.

The solution to the Cauchy problem (initial data on all of Rn\mathbb{R}^n) is given by convolution with the heat kernel:

u(x,t)=RnΦ(xy,t)u0(y)dy,Φ(x,t)=1(4πkt)n/2ex2/(4kt).u(x, t) = \int_{\mathbb{R}^n} \Phi(x - y, t) \, u_0(y) \, dy, \qquad \Phi(x, t) = \frac{1}{(4\pi k t)^{n/2}} e^{-|x|^2/(4kt)}.

The heat kernel is a Gaussian that broadens and flattens as tt increases, expressing the physical spreading of heat. Several qualitative features stand out immediately. First, the solution is instantly smooth for t>0t > 0 no matter how rough the initial data — a smoothing effect absent from hyperbolic equations. Second, and perhaps surprisingly, the solution at any point (x,t)(x, t) with t>0t > 0 depends on the initial data at every point yRny \in \mathbb{R}^n: perturbations propagate at infinite speed, in stark contrast to wave equations. This is a mathematical artifact of the idealized model; real heat conduction is governed by quantum mechanics at short scales.

For bounded domains, the method of separation of variables and eigenfunction expansion provides explicit solutions. Seeking u(x,t)=X(x)T(t)u(x, t) = X(x) T(t) in Ω×(0,)\Omega \times (0, \infty) with u=0u = 0 on Ω\partial\Omega reduces the heat equation to two ODEs: T+λkT=0T' + \lambda k T = 0 and ΔX=λX-\Delta X = \lambda X. The second is the eigenvalue problem for the Laplacian, which has a sequence of eigenvalues 0<λ1λ20 < \lambda_1 \leq \lambda_2 \leq \cdots \to \infty with corresponding eigenfunctions ϕn\phi_n that form a complete orthonormal basis of L2(Ω)L^2(\Omega). The solution is then

u(x,t)=n=1cneλnktϕn(x),cn=Ωu0(y)ϕn(y)dy.u(x, t) = \sum_{n=1}^\infty c_n e^{-\lambda_n k t} \phi_n(x), \qquad c_n = \int_\Omega u_0(y) \phi_n(y) \, dy.

Each mode decays exponentially, with the lowest eigenvalue λ1\lambda_1 governing the long-time behavior: u(,t)L2eλ1kt\|u(\cdot, t)\|_{L^2} \sim e^{-\lambda_1 k t} as tt \to \infty.

The parabolic maximum principle asserts that the maximum of a solution to utΔu=0u_t - \Delta u = 0 over the parabolic boundary (the bottom and sides of a space-time cylinder) is also the maximum over the entire closed cylinder. This powerful principle governs comparison, uniqueness, and stability: if two solutions agree at t=0t = 0 and on Ω\partial\Omega for all t>0t > 0, they agree everywhere. Duhamel’s principle extends the framework to non-homogeneous equations utΔu=fu_t - \Delta u = f: the solution is built by treating each instantaneous source f(,s)f(\cdot, s) as an initial condition for a separate heat equation started at time ss, then integrating over ss.

The backward heat equation ut=Δuu_t = -\Delta u is the time-reversal of the heat equation and is catastrophically ill-posed in the sense of Hadamard. The reason is that the forwards heat equation destroys information by smoothing, so reversing time requires reconstructing information from a smoothed state — an exponentially unstable operation. This ill-posedness is of practical importance in inverse problems, where one wishes to recover a past temperature distribution from present measurements.

Hyperbolic Equations and Wave Phenomena

The wave equation utt=c2Δuu_{tt} = c^2 \Delta u governs vibrating strings (n=1n = 1), sound (n=3n = 3), and electromagnetic radiation (n=3n = 3). Here c>0c > 0 is the propagation speed. Unlike the heat equation, the wave equation is time-reversible and preserves information: what is true now was also true in the past, at least in the classical setting. The wave equation was studied intensely in the eighteenth century by Jean le Rond d’Alembert, Leonhard Euler, and Daniel Bernoulli, who disagreed vigorously about the nature of its solutions — a dispute that helped crystallize the modern concept of a function.

In one spatial dimension, d’Alembert’s formula gives the complete solution to the initial value problem u(x,0)=ϕ(x)u(x, 0) = \phi(x), ut(x,0)=ψ(x)u_t(x, 0) = \psi(x):

u(x,t)=ϕ(x+ct)+ϕ(xct)2+12cxctx+ctψ(s)ds.u(x, t) = \frac{\phi(x + ct) + \phi(x - ct)}{2} + \frac{1}{2c} \int_{x - ct}^{x + ct} \psi(s) \, ds.

This formula makes transparent the defining feature of hyperbolic equations: the domain of dependence. The value u(x,t)u(x, t) depends only on initial data in the interval [xct,x+ct][x - ct, x + ct], the segment swept out by characteristics running backward from (x,t)(x, t) at speed ±c\pm c. Any perturbation of data outside this interval has no effect on u(x,t)u(x, t). Conversely, the domain of influence of a point (x0,0)(x_0, 0) is the cone {(x,t):xx0ct}\{(x, t) : |x - x_0| \leq ct\} — the set of space-time points that can be affected by data at x0x_0.

In three spatial dimensions, the solution is given by Kirchhoff’s formula:

u(x,t)=t(14πc2tyx=ctϕ(y)dS(y))+14πc2tyx=ctψ(y)dS(y).u(x, t) = \frac{\partial}{\partial t}\left(\frac{1}{4\pi c^2 t} \int_{|y - x| = ct} \phi(y) \, dS(y)\right) + \frac{1}{4\pi c^2 t} \int_{|y - x| = ct} \psi(y) \, dS(y).

A crucial observation: in three dimensions, the value u(x,t)u(x, t) depends only on data on the sphere yx=ct|y - x| = ct, not on data inside the sphere. This is Huygens’ principle — sharp signals in three dimensions remain sharp, whereas in two dimensions (and other even dimensions), a signal spreads into a trailing wake. The reason Huygens’ principle holds in odd dimensions but fails in even ones is one of the more beautiful results in PDE theory, connected to the theory of the wave operator in different dimensions.

The wave equation conserves a natural energy. For the homogeneous wave equation, the quantity

E(t)=12Ω(ut2+c2u2)dxE(t) = \frac{1}{2} \int_\Omega \left(u_t^2 + c^2 |\nabla u|^2\right) dx

satisfies ddtE(t)=0\frac{d}{dt} E(t) = 0, so E(t)=E(0)E(t) = E(0) for all time. Energy conservation provides uniqueness (if two solutions have the same initial data, their difference has zero energy, hence is zero) and stability. Nonlinear hyperbolic equations, such as conservation laws of the form ut+f(u)x=0u_t + f(u)_x = 0, exhibit more complex behavior: smooth initial data can develop jump discontinuities (shocks) in finite time. The Rankine-Hugoniot condition governs the speed of a shock, and entropy conditions select the physically meaningful weak solution among multiple candidates.

Variational Methods and Weak Solutions

The history of variational methods in PDE theory begins with Dirichlet’s principle, which asserts that the harmonic function on a domain Ω\Omega with prescribed boundary values gg is the function minimizing the Dirichlet energy

E[u]=Ωu2dxE[u] = \int_\Omega |\nabla u|^2 \, dx

among all functions agreeing with gg on Ω\partial\Omega. Riemann used this principle freely, but Weierstrass showed in 1870 that a minimizing sequence need not converge — the infimum might not be attained. This crisis was resolved by David Hilbert around 1900 through the development of functional analysis, culminating in the direct method in the calculus of variations: one shows that a minimizing sequence is bounded in a suitable function space, extracts a weakly convergent subsequence by compactness, and then uses lower semicontinuity of the energy to confirm that the limit is indeed a minimizer.

The notion of a weak solution is the central conceptual innovation that enables modern PDE theory. Classical solutions require enough regularity to substitute directly into the PDE; weak solutions broaden the solution concept by moving derivatives onto smooth test functions φCc(Ω)\varphi \in C_c^\infty(\Omega) through integration by parts. For the Poisson equation Δu=f-\Delta u = f, the weak formulation asks for uH01(Ω)u \in H^1_0(\Omega) (functions with one weak derivative that vanish on Ω\partial\Omega, in the L2L^2 sense) such that

Ωuφdx=Ωfφdxfor all φH01(Ω).\int_\Omega \nabla u \cdot \nabla \varphi \, dx = \int_\Omega f \varphi \, dx \qquad \text{for all } \varphi \in H^1_0(\Omega).

This formulation makes sense even when ff is merely in L2(Ω)L^2(\Omega) and uu lacks the two classical derivatives needed to write Δu\Delta u pointwise. The Lax-Milgram theorem guarantees existence and uniqueness of weak solutions for a broad class of elliptic problems, and elliptic regularity then shows that the weak solution is as smooth as the data permit.

The Euler-Lagrange equation connects variational problems to PDEs. If uu minimizes a functional of the form F[u]=ΩL(x,u,u)dx\mathcal{F}[u] = \int_\Omega L(x, u, \nabla u) \, dx, then uu satisfies ixiLpi+Lu=0-\sum_i \partial_{x_i} L_{p_i} + L_u = 0. For the Dirichlet energy, L=u2/2L = |\nabla u|^2 / 2, giving Δu=0-\Delta u = 0. For the minimal surface functional L=1+u2L = \sqrt{1 + |\nabla u|^2}, the Euler-Lagrange equation becomes the minimal surface equation

div(u1+u2)=0,\operatorname{div}\left(\frac{\nabla u}{\sqrt{1 + |\nabla u|^2}}\right) = 0,

a quasi-linear elliptic PDE. Weak formulations and variational methods unite large swaths of PDE theory into a single coherent framework, providing existence results for elliptic, parabolic, and even hyperbolic problems in energy spaces naturally suited to each type.

Function Spaces and Regularity Theory

The modern treatment of PDEs is inseparable from a hierarchy of function spaces designed to measure the regularity of solutions. At the top of the hierarchy sit the classical smooth spaces Ck(Ω)C^k(\Omega) and Holder spaces Ck,α(Ω)C^{k,\alpha}(\Omega); at the foundational level sit the Sobolev spaces Wk,p(Ω)W^{k,p}(\Omega), which are the workhorses of contemporary PDE analysis.

The Sobolev space Wk,p(Ω)W^{k,p}(\Omega) consists of all functions in Lp(Ω)L^p(\Omega) whose weak derivatives up to order kk also lie in Lp(Ω)L^p(\Omega), equipped with the norm

uWk,p(Ω)=(αkΩDαupdx)1/p.\|u\|_{W^{k,p}(\Omega)} = \left(\sum_{|\alpha| \leq k} \int_\Omega |D^\alpha u|^p \, dx\right)^{1/p}.

The space Hk(Ω)=Wk,2(Ω)H^k(\Omega) = W^{k,2}(\Omega) is a Hilbert space, making it particularly amenable to the techniques of functional analysis. The weak derivative DαuD^\alpha u is defined by the integration-by-parts identity Ω(Dαu)φ=(1)αΩuDαφ\int_\Omega (D^\alpha u) \varphi = (-1)^{|\alpha|} \int_\Omega u D^\alpha \varphi for all test functions φ\varphi, bypassing any requirement that uu be classically differentiable. The weak derivative was developed independently by Sergei Sobolev and Jean Leray in the 1930s, largely motivated by the needs of fluid mechanics and the Navier-Stokes equations.

Sobolev embedding theorems are among the deepest and most useful results in analysis. They assert that, under appropriate dimensional conditions, membership in Wk,pW^{k,p} implies pointwise regularity. The critical case is governed by the Sobolev exponent p=np/(nkp)p^* = np/(n - kp) for kp<nkp < n: the Sobolev embedding Wk,p(Ω)Lp(Ω)W^{k,p}(\Omega) \hookrightarrow L^{p^*}(\Omega) holds continuously. When kp>nkp > n, functions in Wk,pW^{k,p} are Holder continuous: Wk,p(Ω)Ckn/p1,α(Ω)W^{k,p}(\Omega) \hookrightarrow C^{k - n/p - 1, \alpha}(\Omega) for appropriate α\alpha. These embeddings explain why solutions to elliptic equations gain regularity: if fL2f \in L^2 and uH1u \in H^1 satisfies an elliptic equation, then elliptic regularity lifts uu to H2H^2, which by Sobolev embedding may already be continuous.

Distribution theory, developed systematically by Laurent Schwartz in the 1940s (for which he received the Fields Medal in 1950), provides the broadest possible framework for generalized functions. A distribution is a continuous linear functional on the space D(Ω)=Cc(Ω)\mathcal{D}(\Omega) = C_c^\infty(\Omega) of test functions. Every locally integrable function ff defines a distribution via φΩfφdx\varphi \mapsto \int_\Omega f \varphi \, dx, but there are distributions — such as the Dirac delta δx\delta_x and its derivatives — that correspond to no function at all. The power of distributions is that every distribution can be differentiated arbitrarily many times: iT\partial_i T is the distribution defined by φT(iφ)\varphi \mapsto -T(\partial_i \varphi). This allows one to speak rigorously of the Laplacian of a Green’s function or the derivative of a shock wave, without any classical pointwise meaning.

The trace theorem handles boundary values in the Sobolev framework. Functions in H1(Ω)H^1(\Omega) need not be continuous up to Ω\partial\Omega, so their boundary values cannot be defined pointwise. Nevertheless, the trace operator γ0:H1(Ω)H1/2(Ω)\gamma_0 : H^1(\Omega) \to H^{1/2}(\partial\Omega) is a well-defined, bounded linear map that extends the restriction map from smooth functions. This allows boundary conditions to be imposed in the weak formulation in a mathematically rigorous way.

Compact embeddings — specifically the Rellich-Kondrachov theorem, which asserts that W1,p(Ω)Lp(Ω)W^{1,p}(\Omega) \hookrightarrow L^p(\Omega) is compact when Ω\Omega is bounded — are essential for extracting convergent subsequences in existence proofs. This compactness result is the functional-analytic backbone of the direct method: a bounded sequence in H1H^1 has a subsequence converging strongly in L2L^2, which is precisely the convergence needed to pass limits through nonlinear terms.

Green’s Functions and Numerical Methods

Green’s functions are the quintessential tool for converting PDE boundary value problems into explicit integral representations. For a linear differential operator L\mathcal{L} on a domain Ω\Omega with specified boundary conditions, the Green’s function G(x,y)G(x, y) is the response of the system at xx to a unit point source at yy. Once GG is known, the solution to Lu=f\mathcal{L}u = f is

u(x)=ΩG(x,y)f(y)dy.u(x) = \int_\Omega G(x, y) f(y) \, dy.

Green’s functions were introduced by the self-taught British mathematician George Green in his 1828 Essay on the Application of Mathematical Analysis to the Theories of Electricity and Magnetism, a work that was largely ignored for two decades before being recognized as foundational. The construction of the Green’s function for specific domains exploits symmetry and the method of images: for the Laplacian on the half-space R+n\mathbb{R}^n_+, the Green’s function is G(x,y)=Φ(xy)Φ(xy)G(x, y) = \Phi(x - y) - \Phi(x - y^*), where yy^* is the reflection of yy across the boundary, so that GG vanishes on R+n\partial\mathbb{R}^n_+.

For the heat equation, the fundamental solution Φ(x,t)=(4πt)n/2ex2/(4t)\Phi(x, t) = (4\pi t)^{-n/2} e^{-|x|^2/(4t)} plays the role of the Green’s function on all of Rn\mathbb{R}^n. More generally, the heat kernel on a bounded domain carries geometric information about Ω\Omega: the spectrum of the Laplacian can be recovered from the short-time asymptotics of ΩG(x,x,t)dx\int_\Omega G(x, x, t) \, dx, a connection captured by the celebrated formula of Mark Kac: “Can one hear the shape of a drum?” — can one determine Ω\Omega from the eigenvalues of its Laplacian? The answer, provided by Carolyn Gordon, David Webb, and Scott Wolpert in 1992, is no: non-isometric domains can share the same eigenvalue spectrum.

When explicit Green’s functions are unavailable — which is the generic case for irregular domains or variable-coefficient operators — numerical methods take over. The three main classes are finite difference methods, finite element methods, and spectral methods. Finite difference methods replace derivatives by difference quotients on a grid: uxx(x)(u(x+h)2u(x)+u(xh))/h2u_{xx}(x) \approx (u(x+h) - 2u(x) + u(x-h))/h^2. They are simple to implement and analyze, and the Courant-Friedrichs-Lewy (CFL) condition dictates the stability constraint for explicit schemes applied to hyperbolic equations: the time step Δt\Delta t must satisfy cΔt/h1c \Delta t / h \leq 1, ensuring that the numerical domain of dependence contains the true domain of dependence.

Finite element methods (FEM) work directly with the weak formulation. One approximates uu in a finite-dimensional subspace VhH01(Ω)V_h \subset H^1_0(\Omega) (typically piecewise polynomial functions on a triangulation of Ω\Omega) and finds uhVhu_h \in V_h satisfying the weak equation for all vhVhv_h \in V_h. Galerkin’s method — named after Boris Galerkin, who developed it for structural mechanics in 1915 — yields a linear system Ku=fKu = f where KK is the stiffness matrix with entries Kij=a(ϕj,ϕi)K_{ij} = a(\phi_j, \phi_i). The method is supported by a complete error theory: Cea’s lemma shows that the FEM approximation is quasi-optimal, uuhH1CinfvhVhuvhH1\|u - u_h\|_{H^1} \leq C \inf_{v_h \in V_h} \|u - v_h\|_{H^1}, and polynomial approximation theory then gives convergence rates in terms of mesh size hh and polynomial degree. Adaptive refinement driven by a posteriori error estimators allows one to concentrate computational effort where the solution is least regular.

Spectral methods expand the solution in a globally smooth basis — Fourier modes or Chebyshev polynomials — and achieve exponential convergence rates for smooth solutions, far surpassing the algebraic rates of finite differences and finite elements. They are most effective for problems on simple geometries with smooth data, such as in weather prediction, fluid simulations, and quantum mechanics.

The interplay of analysis and computation that PDE theory demands has made it one of the richest and most active areas of mathematics. From Fourier’s original heat equation to the Navier-Stokes equations governing turbulent flow (whose global regularity remains one of the Clay Millennium Prize Problems) and to the emerging field of physics-informed neural networks, the subject continues to grow, always driven by the same essential tension: between the physical world’s inexhaustible complexity and mathematics’ drive to find structure within it.