Real Analysis
Sequences, continuity, differentiation, and Riemann/Lebesgue integration.
Real analysis is the rigorous foundation of calculus — the discipline that supplies precise definitions and airtight proofs for the intuitions that Newton and Leibniz set in motion in the seventeenth century. It is where the notion of “approaching a limit” is given an exact meaning, where the familiar rules of differentiation and integration are derived from first principles, and where the real number line is examined with enough care to reveal its subtle completeness properties. Studying real analysis is, above all, an exercise in mathematical maturity: it trains the mind to distrust intuition just enough, and to trust proof absolutely.
Foundations of the Real Number System
The story begins not with calculus but with numbers. The rational numbers — fractions with , — seem, at first glance, to fill up the number line. The ancient Greeks believed precisely this until they proved, to their dismay, that cannot be rational. The rationals have gaps, and those gaps are precisely the obstacle that prevents a naive theory of limits from working.
The real numbers are constructed to fill those gaps. There are two classical constructions. Richard Dedekind, in his 1872 essay Stetigkeit und irrationale Zahlen, defined each real number as a Dedekind cut: a partition of into two non-empty sets such that every element of is less than every element of and has no largest element. The real number , for instance, corresponds to the cut where . Simultaneously, Georg Cantor proposed defining real numbers as equivalence classes of Cauchy sequences of rationals — sequences that become arbitrarily close to each other without necessarily converging to a rational limit. Both constructions yield the same object: an ordered field with the crucial additional property called completeness.
Formally, is characterized as the unique complete ordered field. It satisfies the usual field axioms (addition, multiplication, their inverses), an ordering compatible with the field structure, and the Completeness Axiom (also called the Least Upper Bound Property): every non-empty subset of that is bounded above has a supremum (least upper bound) in . This single axiom is what distinguishes from . The supremum of a set is written ; the greatest lower bound is the infimum .
Two important consequences flow immediately from completeness. The Archimedean property states that for every real number , there exists a natural number with — the natural numbers are not bounded above in . This rules out infinitely large or infinitely small elements. The density of the rationals states that between any two distinct real numbers there exists a rational number with . Remarkably, there also exists an irrational between and , so both and are dense in , even though they have very different cardinalities (countable versus uncountable).
The absolute value measures distance from the origin: if and if . Its most important property is the triangle inequality:
The triangle inequality is the engine that drives most convergence arguments. A variant, the reverse triangle inequality , is equally useful in bounding differences.
Sequences and Series
A sequence of real numbers is a function , written or simply . The central question is: does the sequence settle down to a definite value? A sequence converges to a limit , written or , if for every there exists such that for all . In plain English: the terms of the sequence eventually stay within any prescribed distance of . This is the - definition, and mastering it — learning to produce an given an arbitrary — is the first and most important technical skill of real analysis.
Limits, when they exist, are unique. Convergent sequences are necessarily bounded: there exists with for all . The algebra of limits is familiar from calculus but requires proof: if and , then , , and provided .
Monotone sequences are particularly well-behaved. The Monotone Convergence Theorem states that every monotone increasing sequence that is bounded above converges, and every monotone decreasing sequence that is bounded below converges. This is a direct consequence of the completeness axiom: the supremum of the sequence is its limit.
The Bolzano-Weierstrass Theorem (named for Bernard Bolzano and Karl Weierstrass, the latter of whom gave the modern rigorous treatment in the 1860s) asserts that every bounded sequence of real numbers has a convergent subsequence. A subsequence is obtained by selecting an infinite increasing chain of indices . Bolzano-Weierstrass is a compactness result in disguise, and it underpins the proofs of the Extreme Value Theorem and the Heine-Cantor Theorem.
A sequence is a Cauchy sequence if for every there exists such that for all . The terms become close to each other without reference to any proposed limit. The fundamental result is that, in , Cauchy sequences and convergent sequences are the same thing: a sequence converges if and only if it is Cauchy. This property, called completeness of , is what the Cantor construction is designed to guarantee.
An infinite series is defined as the limit of the partial sums . The series converges if and only if converges as a sequence. A necessary condition for convergence is that , but this condition is far from sufficient — the harmonic series diverges even though , a fact first proved by Nicole Oresme in the fourteenth century. The standard convergence tests — comparison, ratio, root, and the alternating series test — each carve out sufficient conditions. A series is absolutely convergent if converges; absolute convergence implies convergence, and absolutely convergent series can be rearranged without changing their sum. Conditionally convergent series — those that converge but not absolutely — are far more delicate: by the Riemann Rearrangement Theorem, any conditionally convergent series can be rearranged to converge to any prescribed real number, or even to diverge to .
Limits, Continuity, and Differentiation
For functions , the limit means: for every , there exists such that implies . The condition ensures we do not require anything about itself. A function is continuous at if — the limit exists, equals the function value, and the function is defined at . Continuity has an equivalent sequential characterization: is continuous at if and only if implies for every sequence .
Continuous functions on closed bounded intervals enjoy two celebrated properties. The Extreme Value Theorem (proved rigorously by Weierstrass) states that if is continuous, then attains its maximum and minimum values — there exist with for all . The Intermediate Value Theorem (traced to Bolzano’s 1817 paper Rein analytischer Beweis) states that if is continuous and , then there exists with . The Intermediate Value Theorem is the rigorous underpinning of root-finding algorithms: every continuous function that changes sign must have a zero.
Uniform continuity is a stronger condition: is uniformly continuous on a set if for every there exists a single (independent of the point) such that implies for all . The Heine-Cantor Theorem asserts that every continuous function on a closed bounded interval is uniformly continuous — a powerful result with no analogue on open intervals ( on is continuous but not uniformly so).
The derivative of at a point is defined as the limit of the difference quotient:
provided this limit exists. Differentiability at implies continuity at , but not conversely — continuity is strictly weaker. The standard differentiation rules (sum, product, quotient, chain) are theorems, not axioms. The most important theorems in differential calculus are the mean value theorems. Rolle’s Theorem states that if is continuous on , differentiable on , and , then there exists with . The Lagrange Mean Value Theorem (the form most often called “the mean value theorem”) generalizes this:
This single result implies that a function with a positive derivative on an interval is increasing, that a function with a zero derivative is constant, and that differentiable functions cannot oscillate faster than their derivative allows. Taylor’s Theorem extends the idea, approximating a sufficiently smooth function by a polynomial and providing an explicit formula for the remainder.
Riemann Integration
The Riemann integral, formalized by Bernhard Riemann in his 1854 Habilitation thesis Über die Darstellbarkeit einer Function durch eine trigonometrische Reihe, gives a precise meaning to the area under a curve. The construction begins with partitions: a partition of is a finite collection of points . For each subinterval , define the upper sum where , and the lower sum where .
A bounded function is Riemann integrable on if the infimum of all upper sums equals the supremum of all lower sums:
The Riemann-Darboux criterion for integrability states that is integrable if and only if for every there exists a partition with . From this criterion, two large classes of integrable functions emerge: all continuous functions on are integrable (Weierstrass), and all monotone functions on are integrable. More generally, functions with only finitely many discontinuities are integrable, as are functions whose set of discontinuities has measure zero.
The relationship between differentiation and integration is codified in the two Fundamental Theorems of Calculus. The first theorem states that if is continuous on and , then is differentiable and — integration produces an antiderivative. The second theorem states that if is integrable on and is any antiderivative of , then . Together, these theorems establish that differentiation and integration are inverse operations, a fact that Newton and Leibniz used instinctively but that required two centuries of effort to prove with full rigor.
Metric Spaces and Topology of Euclidean Space
Real analysis on the real line generalizes naturally to higher dimensions and to abstract spaces. A metric space is a pair where is a set and is a metric satisfying, for all : (i) if and only if ; (ii) (symmetry); (iii) (triangle inequality). The Euclidean metric on is . Other examples include the discrete metric, the supremum metric on spaces of functions, and the -adic metric on number fields.
An open ball of radius centered at is the set . A set is open if every point of has an open ball entirely contained in . A set is closed if its complement is open — equivalently, if it contains all its limit points. The open sets define the topology of the metric space, encoding the notion of nearness.
The critical topological notion for analysis is compactness. A subset of a metric space is compact if every open cover of has a finite subcover. In , the Heine-Borel Theorem gives a much simpler characterization: a subset of is compact if and only if it is closed and bounded. Compact sets are the natural domain for the strongest theorems: continuous functions on compact sets are uniformly continuous (Heine-Cantor), attain their extreme values (Extreme Value Theorem), and their images are compact.
A subset of a metric space is connected if it cannot be written as a union of two disjoint non-empty open sets. In , the connected sets are precisely the intervals (including rays and the whole line). A stronger notion is path connectedness: is path connected if any two points in can be joined by a continuous path lying entirely in . In , path connectedness implies connectedness.
A metric space is complete if every Cauchy sequence in it converges to a point in the space. Euclidean space is complete; the rationals are not. Completeness is the precise property that prevents sequences from converging to “missing” points. The Banach Fixed Point Theorem (also called the Contraction Mapping Theorem) states that every contraction on a complete metric space has a unique fixed point, and that iterating the contraction from any starting point converges to it. This theorem is the backbone of many existence proofs in differential equations and numerical analysis.
Sequences and Series of Functions
When functions, rather than numbers, form the terms of a sequence, a new subtlety emerges: there are two distinct notions of convergence. A sequence of functions on a set converges pointwise to if for each fixed — a separate limit condition for each point. It converges uniformly to if for every there exists (independent of ) such that for all and all . Uniform convergence is a joint condition on the entire function simultaneously, while pointwise convergence permits the rate of convergence to vary arbitrarily from point to point.
The distinction matters enormously because pointwise limits can destroy properties that each individual function possesses. A pointwise limit of continuous functions need not be continuous: the sequence on converges pointwise to the function that is on and at , which is discontinuous. Under uniform convergence, however, the limit of a sequence of continuous functions is continuous, the limit can be integrated term-by-term, and (with an additional condition on the derivatives) the limit can be differentiated term-by-term. These exchange-of-limits results are among the most useful theorems in analysis, and their failure under mere pointwise convergence is one of the lessons that the nineteenth century had to learn painfully.
A series of functions converges uniformly if the partial sums converge uniformly. The Weierstrass M-test gives a clean sufficient condition: if for all and , then converges uniformly and absolutely. Power series are the most important examples: each power series has a radius of convergence (given by the Cauchy-Hadamard formula ) such that the series converges absolutely and uniformly on compact subsets of and diverges for . Within its interval of convergence, a power series can be differentiated and integrated term-by-term, and it defines a function whose Taylor coefficients are exactly . This is the theory of analytic functions in the real setting, a precursor to complex analysis.
The Weierstrass Approximation Theorem (1885) states that every continuous function on a closed bounded interval can be approximated uniformly by polynomials. The theorem was a surprise: polynomials are special, yet they are dense in the space of all continuous functions. Karl Weierstrass proved the theorem constructively; later, Sergei Bernstein gave an elegant probabilistic proof in 1912 using the polynomials now bearing his name. The Stone-Weierstrass Theorem vastly generalizes the result, replacing polynomials by any subalgebra of continuous functions that separates points and contains constants.
Introduction to Lebesgue Integration
The Riemann integral has a fundamental limitation: it struggles with functions that oscillate wildly or have too many discontinuities. The Dirichlet function — defined to be on the rationals and on the irrationals — is not Riemann integrable on any interval, even though it “should” have integral (the rationals are negligible). The Riemann integral is also poorly behaved with respect to limits: pointwise limits of Riemann integrable functions need not be Riemann integrable, and even when they are, one cannot always exchange the limit and the integral.
Henri Lebesgue, in his 1902 doctoral thesis Intégrale, longueur, aire, introduced a radically different approach. Instead of partitioning the domain (the -axis) as Riemann did, Lebesgue partitioned the range (the -axis) and measured the size of the set of -values where the function takes values in each strip. This requires a theory of measure: a way to assign a “size” to subsets of that generalizes the length of intervals.
A -algebra on a set is a collection of subsets of that is closed under complements and countable unions (and hence countable intersections). A measure is a function satisfying and countable additivity: if are pairwise disjoint sets in , then . The Lebesgue measure on is the unique measure on the Borel -algebra that assigns to every interval. Sets of Lebesgue measure zero — null sets — can be ignored for the purposes of integration. The Cantor set is a striking example: it is uncountable yet has Lebesgue measure zero.
A function is Lebesgue measurable if the preimage is a measurable set for every . The Lebesgue integral of a non-negative measurable function is built up in stages: first for simple functions (finite linear combinations of indicator functions of measurable sets), then for general non-negative functions as a supremum over simple functions bounded below them, and finally for general functions by splitting into positive and negative parts.
The power of Lebesgue integration lies in its convergence theorems. The Monotone Convergence Theorem states that if is a sequence of non-negative measurable functions increasing pointwise to , then . Fatou’s Lemma gives a lower bound for lim inf. The crown jewel is the Dominated Convergence Theorem: if pointwise almost everywhere and for an integrable dominating function , then . This theorem is the rigorous tool that makes the exchange of limits and integrals legitimate, and it is indispensable in functional analysis, probability theory, and the theory of partial differential equations.
Every Riemann integrable function is Lebesgue integrable and the two integrals agree. The Lebesgue integral is strictly more general: the Dirichlet function is Lebesgue integrable with integral . The precise characterization of Riemann integrability in Lebesgue’s language is elegant: a bounded function is Riemann integrable if and only if its set of discontinuities has Lebesgue measure zero. Real analysis, which began with the effort to make calculus rigorous, ends by opening the door to measure theory, functional analysis, and modern probability — the landscape of twentieth-century mathematics.