Analytic Number Theory

The Riemann zeta function, the Prime Number Theorem, and L-functions.


Analytic number theory is the branch of mathematics that brings the tools of real and complex analysis to bear on questions about the integers — above all, on the distribution of prime numbers. The field was born in the mid-nineteenth century when Bernhard Riemann showed, in a single landmark 1859 paper, that the prime-counting function is controlled by the zeros of a complex-analytic function now called the Riemann zeta function, and it has grown into one of the richest and most active areas of modern mathematics, connecting elementary arithmetic to complex analysis, harmonic analysis, and representation theory. At its heart lies a stunning fact: the seemingly erratic sequence of prime numbers 2,3,5,7,11,2, 3, 5, 7, 11, \ldots is governed by a precise, beautiful law, and the tools needed to prove it come not from arithmetic at all, but from analysis.

The Riemann Zeta Function

The Riemann zeta function is the central object of analytic number theory. For a complex number ss with real part greater than one, it is defined by the absolutely convergent Dirichlet series

ζ(s)=n=11ns=1+12s+13s+14s+\zeta(s) = \sum_{n=1}^{\infty} \frac{1}{n^s} = 1 + \frac{1}{2^s} + \frac{1}{3^s} + \frac{1}{4^s} + \cdots

The connection to prime numbers appears immediately. Leonhard Euler, working in the eighteenth century, observed that the fundamental theorem of arithmetic — unique factorization of integers into primes — translates directly into a product formula. Since every positive integer has a unique prime factorization, the sum over all integers splits into an independent product over all primes pp:

ζ(s)=p prime11ps\zeta(s) = \prod_{p \text{ prime}} \frac{1}{1 - p^{-s}}

This Euler product encodes all information about primes into the analytic function ζ(s)\zeta(s). When s=1s = 1, the sum becomes the harmonic series 1+12+13+1 + \frac{1}{2} + \frac{1}{3} + \cdots, which diverges — and from this divergence Euler extracted a new proof that there are infinitely many primes, since if there were only finitely many, the product would converge to a finite value.

Riemann’s 1859 paper Ueber die Anzahl der Primzahlen unter einer gegebenen Grösse (On the number of primes less than a given magnitude) transformed the subject by treating ζ(s)\zeta(s) as a function of a complex variable. The series converges absolutely for Re(s)>1\text{Re}(s) > 1, but Riemann showed that ζ(s)\zeta(s) extends to a meromorphic function on the entire complex plane, with a single simple pole at s=1s = 1 with residue 11. The key tool is analytic continuation: starting from the Dirichlet series, one uses integral representations and functional equations to extend the function far beyond its original domain.

The central symmetry of ζ(s)\zeta(s) is expressed by its functional equation, which Riemann established using the completed zeta function ξ(s)=12s(s1)πs/2Γ(s/2)ζ(s)\xi(s) = \frac{1}{2} s(s-1) \pi^{-s/2} \Gamma(s/2) \zeta(s). This completed function satisfies the elegant relation

ξ(s)=ξ(1s)\xi(s) = \xi(1 - s)

reflecting a deep symmetry about the line Re(s)=12\text{Re}(s) = \frac{1}{2}. The gamma factor Γ(s/2)\Gamma(s/2) introduces trivial zeros at the negative even integers s=2,4,6,s = -2, -4, -6, \ldots, where the gamma function has poles. All other zeros — the non-trivial zeros — lie inside the critical strip 0<Re(s)<10 < \text{Re}(s) < 1. Riemann computed the first few non-trivial zeros and noticed they all lie on the critical line Re(s)=12\text{Re}(s) = \frac{1}{2}. He conjectured, cautiously, that this is always the case.

The Riemann Hypothesis — that all non-trivial zeros of ζ(s)\zeta(s) satisfy Re(s)=12\text{Re}(s) = \frac{1}{2} — remains unproven, and is widely regarded as the most important unsolved problem in mathematics. It is one of the Clay Millennium Prize Problems. Computational verification has confirmed the hypothesis for the first 101310^{13} non-trivial zeros, all lying precisely on the critical line, but no proof is in sight. The stakes are high: the Riemann Hypothesis is equivalent to the sharpest possible error bound for the prime-counting function, and its truth (or falsity) would have sweeping consequences throughout number theory and beyond.

Prime Number Theorem and its Proof

The prime-counting function π(x)\pi(x) counts the number of primes less than or equal to xx. The fundamental question — how fast does π(x)\pi(x) grow? — occupied mathematicians for centuries. Empirical evidence accumulated by Gauss and Legendre in the late eighteenth century suggested an asymptotic law, and Gauss conjectured that π(x)\pi(x) grows roughly like x/logxx / \log x or like the logarithmic integral

Li(x)=2xdtlogt\mathrm{Li}(x) = \int_2^x \frac{dt}{\log t}

The Prime Number Theorem makes this precise: as xx \to \infty,

π(x)xlogx\pi(x) \sim \frac{x}{\log x}

which means π(x)/(x/logx)1\pi(x) / (x / \log x) \to 1. Equivalently, using the Chebyshev function ψ(x)=pkxlogp\psi(x) = \sum_{p^k \leq x} \log p (which weights prime powers by their logarithms), the theorem is equivalent to ψ(x)x\psi(x) \sim x.

The Prime Number Theorem was proved independently in 1896 by Jacques Hadamard and Charles-Jean de la Vallée Poussin, using Riemann’s approach. The key step is to show that ζ(s)\zeta(s) has no zeros on the line Re(s)=1\text{Re}(s) = 1 — the boundary of the critical strip. Specifically, one proves the estimate ζ(1+it)0\zeta(1 + it) \neq 0 for all real t0t \neq 0. A clever trigonometric argument using the inequality 3+4cosθ+cos2θ03 + 4\cos\theta + \cos 2\theta \geq 0 establishes this zero-free region. Once this is known, contour integration via the explicit formula — which expresses ψ(x)\psi(x) as a sum over zeros of ζ(s)\zeta(s) — yields the asymptotic ψ(x)x\psi(x) \sim x.

Riemann’s explicit formula is the jewel at the center of this proof. It states, in a precise sense:

ψ(x)=xρxρρlog(2π)12log(1x2)\psi(x) = x - \sum_{\rho} \frac{x^{\rho}}{\rho} - \log(2\pi) - \frac{1}{2} \log(1 - x^{-2})

where the sum runs over all non-trivial zeros ρ\rho of ζ(s)\zeta(s). This formula shows that the zeros of ζ(s)\zeta(s) act as “harmonics” in the distribution of primes: each zero contributes an oscillating term xρ/ρx^{\rho}/\rho, and the prime distribution is the superposition of all these oscillations. If all zeros satisfy Re(ρ)=12\text{Re}(\rho) = \frac{1}{2}, as the Riemann Hypothesis asserts, then xρ=x|x^{\rho}| = \sqrt{x} for each term, and one obtains the conditional error bound

ψ(x)=x+O ⁣(xlog2x)\psi(x) = x + O\!\left(\sqrt{x} \log^2 x\right)

or equivalently π(x)=Li(x)+O(xlogx)\pi(x) = \mathrm{Li}(x) + O(\sqrt{x} \log x). Unconditionally, de la Vallée Poussin established a zero-free region of the form Re(s)>1c/logt\text{Re}(s) > 1 - c/\log |t| for an absolute constant c>0c > 0, which yields the error term

π(x)=Li(x)+O ⁣(xeclogx)\pi(x) = \mathrm{Li}(x) + O\!\left(x \, e^{-c\sqrt{\log x}}\right)

Improving this error term — and in particular reducing the exponent from logx\sqrt{\log x} toward x1/2x^{1/2} — is one of the central open problems in the field, and any progress is intimately tied to knowledge of the zero-free region for ζ(s)\zeta(s).

Dirichlet L-Functions and Primes in Progressions

The natural generalization asks not just how many primes there are up to xx, but how many primes fall in a given arithmetic progression a,a+q,a+2q,a, a + q, a + 2q, \ldots for integers aa and qq with gcd(a,q)=1\gcd(a, q) = 1. The intuitive expectation is that primes are roughly equidistributed among the ϕ(q)\phi(q) residue classes modulo qq that are coprime to qq.

This was made precise by Peter Gustav Lejeune Dirichlet in 1837 in his celebrated theorem: if gcd(a,q)=1\gcd(a, q) = 1, then there are infinitely many primes pa(modq)p \equiv a \pmod{q}. Dirichlet’s proof introduced two fundamental innovations that shaped the next century of number theory.

The first innovation is Dirichlet characters: completely multiplicative functions χ:ZC\chi : \mathbb{Z} \to \mathbb{C} of period qq satisfying χ(n)=0\chi(n) = 0 when gcd(n,q)>1\gcd(n, q) > 1 and χ(n)=1|\chi(n)| = 1 when gcd(n,q)=1\gcd(n, q) = 1. The characters modulo qq form a group of order ϕ(q)\phi(q) under pointwise multiplication, and they satisfy the crucial orthogonality relations:

n=1qχ(n)χ(n)={ϕ(q)if χ=χ0otherwise\sum_{n=1}^{q} \chi(n) \overline{\chi'(n)} = \begin{cases} \phi(q) & \text{if } \chi = \chi' \\ 0 & \text{otherwise} \end{cases}

These orthogonality relations allow one to “detect” a residue class: the indicator of the condition na(modq)n \equiv a \pmod{q} is expressed as

1[na]=1ϕ(q)χmodqχ(a)χ(n)\mathbf{1}[n \equiv a] = \frac{1}{\phi(q)} \sum_{\chi \bmod q} \overline{\chi(a)} \chi(n)

The second innovation is the Dirichlet L-function associated to a character χ\chi:

L(s,χ)=n=1χ(n)ns=p11χ(p)psL(s, \chi) = \sum_{n=1}^{\infty} \frac{\chi(n)}{n^s} = \prod_{p} \frac{1}{1 - \chi(p) p^{-s}}

Like the zeta function, each L(s,χ)L(s, \chi) has an Euler product reflecting unique factorization, and extends to an entire function (if χ\chi is non-principal) or a function with a single pole at s=1s = 1 (for the principal character χ0\chi_0, where L(s,χ0)L(s, \chi_0) is essentially ζ(s)\zeta(s)). By combining the logarithmic derivative of L(s,χ)L(s, \chi) with the orthogonality of characters, Dirichlet reduced the count of primes in the progression a(modq)a \pmod{q} to a sum over all characters:

pxpa(modq)logp1ϕ(q)x\sum_{\substack{p \leq x \\ p \equiv a \pmod{q}}} \log p \approx \frac{1}{\phi(q)} x

The key analytical step is proving that L(1,χ)0L(1, \chi) \neq 0 for all non-principal characters χ\chi. For complex characters this follows quickly from the Euler product, but for real characters (those with χ2=χ0\chi^2 = \chi_0) Dirichlet needed a subtle argument involving class numbers of quadratic forms — a hint of the deep connections between L-functions and algebraic number theory.

The quantitative form, the Siegel-Walfisz theorem, states that for any A>0A > 0 and for q(logx)Aq \leq (\log x)^A, one has

π(x;q,a)=Li(x)ϕ(q)+O ⁣(xeclogx)\pi(x; q, a) = \frac{\mathrm{Li}(x)}{\phi(q)} + O\!\left(x \, e^{-c\sqrt{\log x}}\right)

where the constant cc depends only on AA. The generalized Riemann hypothesis — that all zeros of all Dirichlet L-functions lie on the critical line — would sharpen this dramatically to an error of size O(xlogx)O(\sqrt{x} \log x).

The zeros of Dirichlet L-functions are controlled by the zero-free regions analogous to those for ζ(s)\zeta(s). A notorious difficulty is the possible existence of a Siegel zero: a real zero of L(s,χ)L(s, \chi) very close to s=1s = 1 for a real character χ\chi. The non-existence of Siegel zeros is essentially equivalent to improved bounds for primes in progressions, and despite enormous effort, this remains out of reach.

Additive and Multiplicative Problems

Analytic number theory also attacks problems of an additive nature: when can an integer be expressed as a sum of primes, powers, or other special integers? These questions require a different toolkit — in particular, exponential sums and the circle method.

Goldbach’s conjecture, stated in a 1742 letter from Christian Goldbach to Euler, asserts that every even integer greater than 22 is the sum of two primes. It remains open, despite being verified computationally for all even numbers up to 4×10184 \times 10^{18}. The best proven result in this direction is Chen’s theorem (1973), which establishes that every sufficiently large even integer is the sum of a prime and a number with at most two prime factors.

The ternary Goldbach conjecture — that every odd number greater than 55 is the sum of three primes — was resolved in 2013 by Harald Helfgott, who proved it for all odd numbers greater than 55 without exception. Earlier, Ivan Vinogradov had shown in 1937 that every sufficiently large odd number is the sum of three primes, using his method for bounding Vinogradov exponential sums of the form pxe2πiαp\sum_{p \leq x} e^{2\pi i \alpha p}.

The fundamental tool for these additive problems is the Hardy-Littlewood circle method, developed by G. H. Hardy and John Edensor Littlewood in the 1920s. The idea is to express the number of representations of NN as a sum of kk primes using an integral around the unit circle:

rk(N)=01S(α)ke2πiNαdαr_k(N) = \int_0^1 S(\alpha)^k e^{-2\pi i N \alpha} \, d\alpha

where S(α)=pNe2πiαpS(\alpha) = \sum_{p \leq N} e^{2\pi i \alpha p} is an exponential sum over primes. The circle is divided into major arcs (where α\alpha is close to a rational number with small denominator) and minor arcs (the rest). On major arcs, S(α)S(\alpha) is well-approximated using knowledge of primes in arithmetic progressions, giving a main term called the singular series S(N)\mathfrak{S}(N). On minor arcs, the exponential sum cancels due to equidistribution, giving a negligible error. The singular series is an explicit product over primes that equals 00 for even NN (three-prime Goldbach) and is bounded away from zero for odd NN, explaining why the conjecture should hold only for odd numbers.

Waring’s problem asks for the minimum number g(k)g(k) such that every positive integer is the sum of at most g(k)g(k) perfect kk-th powers. Lagrange’s four-square theorem (k=2k = 2, g(2)=4g(2) = 4) is the classical case. For general kk, it was shown by David Hilbert in 1909 that g(k)g(k) is always finite. The exact values are known for all kk: g(3)=9g(3) = 9, g(4)=19g(4) = 19, and for large kk one has g(k)=2k+(3/2)k2g(k) = 2^k + \lfloor (3/2)^k \rfloor - 2. The Hardy-Littlewood method again provides the main tool, though the minor arc estimates require substantial work.

On the multiplicative side, analytic number theory studies the average behavior of arithmetic functions. The average order of the divisor function d(n)d(n) (which counts divisors) satisfies

nxd(n)=xlogx+(2γ1)x+O(x)\sum_{n \leq x} d(n) = x \log x + (2\gamma - 1)x + O(\sqrt{x})

where γ\gamma is the Euler-Mascheroni constant. The error term O(x)O(\sqrt{x}) is classical (due to Dirichlet), but improving the exponent 12\frac{1}{2} to the conjectured 14+ε\frac{1}{4} + \varepsilon is the Dirichlet divisor problem, still open. The Erdős-Kac theorem (1940) shows that the number of prime factors ω(n)\omega(n) of a “typical” integer near xx is normally distributed with mean and variance both equal to loglogx\log \log x — a striking probabilistic statement about a completely deterministic function.

Modular Forms and L-Functions

The deepest direction of modern analytic number theory leads into the theory of modular forms and their associated L-functions. A modular form is a complex-analytic function on the upper half-plane H={zC:Im(z)>0}\mathbb{H} = \{z \in \mathbb{C} : \text{Im}(z) > 0\} that transforms in a controlled way under the action of the modular group SL2(Z)\mathrm{SL}_2(\mathbb{Z}). Specifically, a modular form of weight kk for SL2(Z)\mathrm{SL}_2(\mathbb{Z}) is a holomorphic function f:HCf : \mathbb{H} \to \mathbb{C} satisfying

f ⁣(az+bcz+d)=(cz+d)kf(z)for all (abcd)SL2(Z)f\!\left(\frac{az + b}{cz + d}\right) = (cz + d)^k f(z) \quad \text{for all } \begin{pmatrix} a & b \\ c & d \end{pmatrix} \in \mathrm{SL}_2(\mathbb{Z})

together with a growth condition at the cusp ii\infty. If ff vanishes at the cusp, it is called a cusp form. Every modular form has a Fourier expansion f(z)=n=0anqnf(z) = \sum_{n=0}^{\infty} a_n q^n where q=e2πizq = e^{2\pi i z}, and the coefficients ana_n carry profound arithmetic information.

The L-function of a cusp form ff is defined by

L(s,f)=n=1annsL(s, f) = \sum_{n=1}^{\infty} \frac{a_n}{n^s}

These L-functions share all the key properties of the Riemann zeta function and Dirichlet L-functions: they have Euler products, analytic continuations, and functional equations. For an eigenform — a cusp form that is simultaneously an eigenfunction of all the Hecke operators TnT_n — the Euler product takes the factored form

L(s,f)=p11apps+pk12sL(s, f) = \prod_{p} \frac{1}{1 - a_p p^{-s} + p^{k-1-2s}}

where apa_p is the Hecke eigenvalue at pp. The multiplicativity of Hecke eigenvalues, established by Erich Hecke in the 1930s, implies the multiplicativity amn=amana_{mn} = a_m a_n for gcd(m,n)=1\gcd(m,n) = 1, which is precisely what gives the Euler product.

The arithmetic significance of modular forms is best illustrated by the Ramanujan tau function τ(n)\tau(n), defined by the expansion of the cusp form Δ(z)=qn=1(1qn)24=n=1τ(n)qn\Delta(z) = q \prod_{n=1}^{\infty} (1-q^n)^{24} = \sum_{n=1}^{\infty} \tau(n) q^n. In 1916, Srinivasa Ramanujan conjectured that τ(p)2p11/2|\tau(p)| \leq 2p^{11/2} for all primes pp. This was proved in 1974 by Pierre Deligne as a consequence of his proof of the Weil conjectures — a profound connection between the theory of modular forms and algebraic geometry over finite fields, for which Deligne received the Fields Medal.

The most spectacular application of modular forms to number theory is the proof of Fermat’s Last Theorem. In 1955, Yutaka Taniyama and Goro Shimura conjectured that every elliptic curve over Q\mathbb{Q} is modular, meaning its L-function equals the L-function of some weight-2 cusp form. In 1986, Ken Ribet proved that the Taniyama-Shimura conjecture (for semistable curves) implies Fermat’s Last Theorem. In 1995, Andrew Wiles, with crucial help from Richard Taylor, proved the semistable case of the conjecture, thereby establishing Fermat’s Last Theorem after 358 years. The full modularity theorem — that every elliptic curve over Q\mathbb{Q} is modular — was completed in 2001 by Breuil, Conrad, Diamond, and Taylor.

The general philosophy connecting arithmetic objects to automorphic L-functions is the Langlands program, initiated by Robert Langlands in a celebrated 1967 letter to André Weil. The program predicts a vast web of correspondences between Galois representations, automorphic forms, and L-functions. It subsumes class field theory, the modularity theorem, and Dirichlet’s theorem as special cases, and its full realization would constitute a unification of number theory on a scale comparable to what the Langlands program itself calls “a kind of arithmetic Fourier analysis.” Current research — on the geometric Langlands program, on the local and global correspondences for reductive groups, and on the applications to Diophantine equations — represents the frontier of analytic number theory and algebraic number theory combined, pointing toward a deep and still-unfolding synthesis of analysis, algebra, and arithmetic.