Generating Functions are Maclaurin Series
The probability generating function GX(t) encodes an entire probability distribution — Binomial, Geometric, Poisson — as a single function of t. What is less often made explicit is that GX(t) is always a Maclaurin series, and the coefficients of that series are the probabilities P(X = k). The Maclaurin formula is the mechanism by which GX(t) generates the distribution.
Probability generating functions sit within the Statistics strand of A-level Further Maths. The underlying machinery — Maclaurin series — is assessed in compulsory Further Pure at all boards. Both strands meet here.
The unifying idea
For any discrete random variable X taking non-negative integer values, the probability generating function is defined by:
This is a Maclaurin series in t. The coefficient of tk is P(X = k). Consequently, the Maclaurin formula — differentiate k times, evaluate at t = 0, divide by k! — is precisely the rule for reading a probability out of GX(t):
The standard textbook presentation
A typical Further Maths statistics course encounters generating functions through specific distributions in roughly this order:
- Binomial distribution — GX(t) = (q + pt)n. This is a polynomial in t, obtained by applying the binomial theorem.
- Geometric distribution — GX(t) = pt / (1 − qt). This is an infinite series, obtained by summing a geometric progression.
- Poisson distribution — GX(t) = eλ(t−1). This involves the exponential function, obtained using the series for ex.
These look like three unrelated techniques. They are not: in each case GX(t) is a Maclaurin series, and the Maclaurin coefficient of tk is P(X = k). The different algebraic forms arise only because the underlying distributions have different probability mass functions.
GX(t) as a Maclaurin series
The Maclaurin series formula is in the compulsory Further Pure content at all boards. The probability generating function GX(t) and its application to specific distributions is assessed in the optional Statistics modules listed above.
Suppose GX(t) can be written as a power series with unknown coefficients:
Set t = 0: immediately a₀ = GX(0) = P(X = 0). Differentiate once:
Set t = 0: a₁ = GX′(0) = P(X = 1). Differentiate again:
Set t = 0: a₂ = GX″(0) / 2 = P(X = 2). After k differentiations:
But aₖ is the coefficient of tk in GX(t), which by definition equals P(X = k). So:
This is the Maclaurin formula. There is no separate "PGF technique" for reading off probabilities — it is differentiation at zero, exactly as in every other Maclaurin series.
Binomial distribution — a polynomial PGF
For X ~ B(n, p) with q = 1 − p, the probability mass function is P(X = k) = C(n, k) pᵏ qn−k. Substituting into the definition:
This is the binomial theorem applied to (q + pt)n:
This is a polynomial of degree n in t — a Maclaurin series that terminates at n = k because P(X = k) = 0 for k > n. The binomial theorem is not a separate method; it produces the same coefficients that the Maclaurin formula would give.
Verification by the Maclaurin formula
Let X ~ B(3, p) so GX(t) = (q + pt)³. Extract P(X = 2) by differentiating twice:
Geometric distribution — a geometric series PGF
For X ~ Geo(p) — the number of trials until the first success — P(X = k) = qk−1p for k = 1, 2, 3, … Substituting:
The sum Σ (qt)j is a geometric series with ratio qt. Its Maclaurin expansion is:
Therefore:
Reading P(X = k) back from GX(t)
Expanding GX(t) = pt · Σ (qt)j = Σ pqk−1 tk, the coefficient of tk is pqk−1 = P(X = k) ✓. The closed form pt / (1 − qt) contains the entire geometric distribution in its Maclaurin expansion.
Poisson distribution — an exponential series PGF
For X ~ Po(λ), P(X = k) = e−λλk/k!. Substituting:
The sum Σ (λt)k/k! is the Maclaurin series for eλt:
Therefore:
Reading P(X = k) back from GX(t)
Expanding GX(t) = e−λ · eλt = e−λ Σ (λt)k/k!, the coefficient of tk is e−λλk/k! = P(X = k) ✓. The entire Poisson distribution is encoded in the single expression eλ(t−1).
Extracting the mean and variance from GX(t)
Once GX(t) is known, the mean and variance follow by differentiation. Differentiating GX(t) = Σ P(X = k) tk with respect to t:
Setting t = 1 (where the series is guaranteed to converge) gives:
Differentiating a second time:
The variance is then:
Convergence — why t = 1 always works
Every PGF converges for |t| ≤ 1. This is not a coincidence: since all P(X = k) ≥ 0 and Σ P(X = k) = 1, for |t| ≤ 1 we have Σ P(X = k)|t|k ≤ Σ P(X = k) = 1. The series is dominated by a convergent series of constants. This guarantees that GX′(1) and GX″(1) exist (whenever the relevant moments exist), which is why evaluating at t = 1 is the standard route to the mean and variance.
| Distribution | GX(t) | Underlying series | Valid for |
|---|---|---|---|
| Bernoulli(p) | q + pt | polynomial (degree 1) | all t |
| Binomial B(n, p) | (q + pt)n | polynomial (degree n) | all t |
| Geometric Geo(p) | pt / (1 − qt) | geometric series 1/(1−qt) | |t| < 1/q |
| Poisson Po(λ) | eλ(t−1) | exponential series eλt | all t |
The Binomial PGF is a polynomial — a Maclaurin series that terminates. The Geometric and Poisson PGFs are infinite series. The Maclaurin formula applies to all of them uniformly: the coefficient of tk is always P(X = k).
The deeper picture — Taylor series
Note on exam boards: Taylor series appears only in the optional Further Pure units. Maclaurin series (expanding around t = 0) is covered in compulsory Further Pure at all four boards; Taylor series (expanding around an arbitrary point a) is the optional extension.
The Maclaurin series is a special case of the Taylor series, which expands GX around an arbitrary point a rather than zero:
Setting a = 0 recovers the Maclaurin series and the PGF formula P(X = k) = GX(k)(0) / k!. Setting a = 1 recovers the moment formulae E[X] = GX′(1) and Var(X) = GX″(1) + GX′(1) − [GX′(1)]² — these are Taylor coefficients at t = 1, not Maclaurin coefficients. Both views are the same theorem applied at different points.
The PGF also has a clean algebraic property: if X and Y are independent, then GX+Y(t) = GX(t) · GY(t). Multiplying two power series corresponds to convolving their coefficients — which is exactly the formula for the distribution of a sum of independent random variables. This is the algebra of generating functions: operations on the series mirror operations on the distributions.
