Evaluating the solution at recovers the original problem and gives . In the figure below we visualize the solution of as a function of .
For this very simple example, we essentially chose a placement of such that the unperturbed problem (for ) becomes trivial. We also have many choices for the power of . In this example, the perturbation series terminates and yields the exact solution.
Since then, we have presented some methods to sum series (Padé approximants, Euler summation, Borel summation, Borel-Écalle summation, generic summation and Zeta summation). We have also seen techniques to accelerate convergence (Shanks transformation and Richardson transformation).
When a problem is not solvable exactly (extracting the roots of a polynomial, solving a specific differential equation, …) the idea of perturbation theory is to obtain an approximate solution (analytically) by inserting a small parameter and assuming the answer has the form .
Procedure:
Insert
Set (the problem must be exactly solvable for )
The solution has the form
Set (the solution of the initial problem is now )
Study the convergence of
If necessary use a summation method
Solve a very simple ‘problem’ using perturbation theory:
Because coefficients of Taylor series are unique (see proof) we are allowed to compare powers of .
This implies that (setting ).
In the figure below, we visualize the solution of as a function of .
Solution of the problem using perturbation theory. For the solution is zero and for the solution is 2. By making approach 1 we obtain an increasingly precise answer to the initial problem.
In physics, most problems—from the anharmonic oscillator to Quantum Electrodynamics (QED)— cannot be solved exactly. We rely on Perturbation Theory, expanding the solution in powers of a small coupling constant (often denoted λ or α).
A classic example is the quantum anharmonic oscillator, where the Hamiltonian is
(with the quartic term treated as a small perturbation λ ≪ 1). Another famous case is QED, where the fine-structure constant α ≈ 1/137 serves as the small parameter for expanding processes involving electrons and photons.
The challenge? These perturbative expansions are almost always asymptotic series with a null radius of convergence. The coefficients typically grow factorially as (or even faster), making the series eventually diverge regardless of how small the coupling constant is.
For instance, in the anharmonic oscillator, the perturbative corrections to the energy levels involve terms whose coefficients grow like , leading to a divergent series for any nonzero λ. Similarly, in QED, the perturbative series for quantities like the electron’s anomalous magnetic moment (g−2) is asymptotic: it gives spectacular agreement with experiment up to high orders, but diverges if pushed too far.
This is where our journey pays off. By applying Borel resummation and Padé approximants to these divergent series, we can extract non-perturbative information and obtain remarkably accurate physical predictions. In quantum field theory, Borel summation helps recover meaningful results from perturbative expansions (for example, in certain Euclidean field theories or by handling contributions from instantons and renormalons). Padé approximants are routinely used to accelerate convergence and estimate the behavior of series in QCD and other models.
In our upcoming posts, we will see how this mathematical framework becomes the essential language of modern theoretical physics.
After several months of exploring various summation and acceleration methods, it is time to synthesize these tools before we apply them to one of their most important domains: Perturbation Theory.
When a power series has a small or even zero radius of convergence, or converges too slowly to be numerically useful, we require more than standard arithmetic. Here is a recap of the methods we have covered and how they bridge the gap between formal series and numerical values.
Summation Methods: Taming Divergence
Borel and Borel–Écalle: By using the Borel transform, we transform factorially growing coefficients into a convergent series in the Borel plane. The Borel–Écalle framework further allows us to handle singularities via resurgence theory, providing a unique “resummed” value even for non-Borel-summable series.
Euler and Zeta summation: These methods assign finite values to divergent sums by analytic continuation. While Euler summation is ideal for alternating series, Zeta summation provides a powerful way to handle the infinite sums that frequently appear in quantum vacuum energy calculations.
Generic summation: Beyond these classical approaches, we have also explored the idea of generic summation, which provides a unifying framework for assigning values to divergent or slowly convergent series.
Acceleration Techniques: Enhancing Convergence
Shanks Transformation and Padé Approximants: These nonlinear transformations (often related to continued fractions) excel at capturing the behavior of functions beyond their radius of convergence, particularly when poles are present.
Richardson Extrapolation: A fundamental tool for numerical analysis that cancels out the leading error terms, allowing us to estimate the limit of a sequence Sn with much higher precision from only a few terms.
An important consistency check underlies all these techniques: whenever different summation or acceleration methods apply to the same series, they agree on a common value. This convergence of independent approaches is not accidental, it reflects the fact that these methods capture an underlying analytic object beyond the formal series itself. When they work, they do not merely assign a value: they reveal a coherent extension of the function that the original divergent expansion was hinting at.
Richardson extrapolation acts as a powerful convergence accelerator in numerical methods. Instead of relying solely on extremely fine discretizations—which can be computationally expensive—it combines results from coarser simulations to estimate the zero-step (continuum) limit.
This idea appears in several numerical techniques (e.g., Romberg integration, mesh convergence studies in engineering), where it can significantly improve accuracy at a modest additional cost.
When the error behaves regularly, Richardson extrapolation can dramatically reduce the error—sometimes by an order of magnitude or more—without requiring prohibitively fine meshes.
Using solely a partial series with 4 terms, we were able to speed up the convergence of the above-mentioned series considerably using Richardson extrapolation.
Another method for accelerating convergence is Richardson extrapolation. This method is similar to the Shanks transformation. Let’s model the partial sum as:
where .
equivalently:
and
Summing up the system above:
Simplifying and taking the limit :
On this basis, we define the Richardson extrapolation as:
To illustrate the Shanks transformation we will use the following series:
This series is converging very slowly. To illustrate this we will compute ln(2):
The corresponding partial sum of ln(2) is:
Let’s calculate some terms of this partial sum:
Perform the Shanks transform (using SN = S4):
Since and It is necessary to calculate up to 358 terms of the partial sum SN to obtain a relative error equal to that of the Shanks transformation based on the first 4 terms.
Therefore the Shanks transformation of the geometric series is correctly . A similar result can be obtained for the series:
The corresponding Shanks transformation is .
Remarkably, the Shanks transformation yields the exact sum of the geometric series using only a few terms, since the error model perfectly describes such a series. The Gregory series for is notoriously slow, requiring five hundred terms just to calculate two correct decimal places. However, by applying the Shanks transformation, we can accelerate this convergence dramatically, extracting high-precision results from only a few initial partial sums. This demonstration highlights how a simple mathematical shift can transform an inefficient infinite series into a powerful computational tool.
Imagine you’ve solved a problem using a Taylor series expansion of the solution, and the resulting series converges very slowly. The Shanks transformation is one way of improving the speed of convergence of a convergent series. Given the partial sum:
Suppose that this partial sum converges to:
We can model the difference between S and SN as follows:
Where :
and
Therefore:
According to this result, the Shanks transformation is defined as:
The various summation methods presented in this table are not arbitrary assignments of values to divergent series; they are governed by the fundamental principle of regularity and consistency. A method is “regular” if it preserves the sum of any convergent series, and “consistent” if different methods yield the same value when they both apply to a given series.
In theoretical physics and asymptotic analysis, these techniques are indispensable. Borel and Euler summations are frequently used to handle “weakly” divergent perturbative expansions in quantum mechanics. Zeta-function regularization is a cornerstone of modern physics, particularly in the calculation of the Casimir effect and in string theory, where it provides a rigorous way to subtract infinities from physical observables. Meanwhile, the Borel–Écalle theory of resurgent functions offers the most sophisticated framework for “decoding” the non-perturbative information hidden within divergent power series. By bridging the gap between divergent expansions and their underlying analytic structures, these methods allow mathematicians and physicists to extract precise physical predictions from seemingly ill-defined mathematical objects.