Approximaths

  • Accelerate convergence (Shanks Transformation)

    Imagine you’ve solved a problem using a Taylor series expansion of the solution, and the resulting series converges very slowly. The Shanks transformation is one way of improving the speed of convergence of a convergent series.
    Given the partial sum:

    \displaystyle S_N = a_0 + a_1 + a_2 + \dots + a_N

    Suppose that this partial sum converges to:

    \displaystyle S = a_0 + a_1 + a_2 + \dots
    \displaystyle S = \lim_{N \to \infty} \sum_{k=0}^{N} a_k

    We can model the difference between S and SN as follows:

    \displaystyle S_N = S + \alpha \beta^{N}

    Where |\beta| < 1:

    \displaystyle S_{N+1} = S + \alpha \beta^{N+1}
    \displaystyle S_N = S + \alpha \beta^{N}
    \displaystyle S_{N-1} = S + \alpha \beta^{N-1}

    and

    \displaystyle S_{N+1} - S = \alpha \beta^{N+1}
    \displaystyle S_N - S = \alpha \beta^{N}
    \displaystyle S_{N-1} - S = \alpha \beta^{N-1}

    Therefore:

    \displaystyle \frac{S_N - S}{S_{N-1}-S} = \frac{\alpha \beta^{N}}{\alpha \beta^{N-1}} = \beta
    \displaystyle \frac{S_{N+1} - S}{S_{N}-S} = \frac{\alpha \beta^{N+1}}{\alpha \beta^{N}} = \beta
    \displaystyle \frac{S_N - S}{S_{N-1}-S} = \frac{S_{N+1} - S}{S_{N}-S}
    \displaystyle (S_N - S)(S_{N}-S) = (S_{N+1} - S)(S_{N-1}-S)
    \displaystyle S_N^2 - 2S_NS + S^2 = S_{N+1}S_{N-1}- SS_{N+1} - SS_{N-1} + S^2
    \displaystyle S_N^2 - 2S_NS + S^2= S_{N+1}S_{N-1} + S^{2}- S(S_{N+1}+ S_{N-1})
    \displaystyle S_N^2 - 2S_NS = S_{N+1}S_{N-1}- S(S_{N+1}+ S_{N-1})
    \displaystyle S_N^2 - S_{N+1}S_{N-1} = - S(S_{N+1}+ S_{N-1}) - 2S_NS
    \displaystyle S_N^2 - S_{N+1}S_{N-1} = S(-S_{N+1}- S_{N-1})+ 2S_N
    \displaystyle S = \frac{S_N^2 - S_{N+1}S_{N-1}}{2S_N -S_{N+1}- S_{N-1}}

    According to this result, the Shanks transformation is defined as:

    \displaystyle \mathcal{S} := \frac{S_N^2 - S_{N+1}S_{N-1}}{2S_N -S_{N+1}- S_{N-1}}
  • Summation summary

    Series Standard Euler Borel Borel–Écalle Zeta
    \displaystyle 1+\frac{1}{4}+\frac{1}{9}+\dots
    \displaystyle \frac{\pi^2}{6}
    \displaystyle \frac{\pi^2}{6}
    \displaystyle \frac{\pi^2}{6}
    \displaystyle \frac{\pi^2}{6}
    \displaystyle \frac{\pi^2}{6}
    \displaystyle 1-1+1-1+\dots
    \displaystyle \frac12
    \displaystyle \frac12
    \displaystyle \frac12
    \displaystyle \frac12
    \displaystyle 1!-2!+3!-4!+\dots
    \displaystyle \approx 0.5963
    \displaystyle \approx 0.5963
    \displaystyle 1+2+3+4+\dots
    \displaystyle -\frac{1}{12}
    \displaystyle 1!+2!+3!+4!+\dots

    The various summation methods presented in this table are not arbitrary assignments of values to divergent series; they are governed by the fundamental principle of regularity and consistency. A method is “regular” if it preserves the sum of any convergent series, and “consistent” if different methods yield the same value when they both apply to a given series. \newline

    In theoretical physics and asymptotic analysis, these techniques are indispensable. Borel and Euler summations are frequently used to handle “weakly” divergent perturbative expansions in quantum mechanics. Zeta-function regularization is a cornerstone of modern physics, particularly in the calculation of the Casimir effect and in string theory, where it provides a rigorous way to subtract infinities from physical observables. Meanwhile, the Borel–Écalle theory of resurgent functions offers the most sophisticated framework for “decoding” the non-perturbative information hidden within divergent power series. By bridging the gap between divergent expansions and their underlying analytic structures, these methods allow mathematicians and physicists to extract precise physical predictions from seemingly ill-defined mathematical objects.

  • Classic Summation Axioms

    So far, we’ve seen several ways of summing series (the usual method, Euler, Borel, generic, Borel-Écalle and Zeta summation). All of these methods fulfill all three properties — except Zeta summation, which fulfills none of them.

    \displaystyle \mathcal{S}(a_0+a_1+a_2+ ...) = \lim_{n \to \infty} \sum_{k =0}^{n}a_k \quad\text{(Regularity)}
    \displaystyle \mathcal{S}(\sum(\alpha a_n) + \sum(\beta b_n)) = \alpha \mathcal{S}(\sum a_n) + \beta \mathcal{S}(\sum b_n) \quad\text{(Linearity)}
    \displaystyle \mathcal{S}(a_0+a_1+a_2+ ...) = a_0 + \mathcal{S}(a_1+a_2+ ...) \quad\text{(Stability)}

    Concerning the Zeta summation we should restrain our enthusiasm by saying that a certain meromorphic complex function called \zeta which have the value \zeta(s)=\sum_{n=1}^{\infty} \frac{1}{n^s} for all Re(s)>1 can be defined on the whole complex plane except at 1, in such a way that, \zeta(0) = -\frac{1}{2}. This ‘summation’ does not rely on any of the properties cited above. So it’s important to be clear about the method we are using and what properties it fulfills.

  • Zeta summation VI

    If we want to calculate \zeta(-2) for example we need to have a proper value for \Gamma(s/2) in for Re(s) \leq 0. The problem is that the Bernoulli representation of the Gamma function presented above:

    \displaystyle \Gamma(s) = \int_{0}^{\infty} e^{-t} t^{s-1}\,dt

    is only valid for Re(s) > 0. We have to provide another representation of the Gamma function for negative values of s:

     \Gamma(s) = \int_{0}^{\infty} e^{-t}\, t^{s-1}\, dt
     = \int_{1}^{\infty} e^{-t}\, t^{s-1}\, dt + \int_{0}^{1} e^{-t}\, t^{s-1}\, dt
     = \int_{1}^{\infty} e^{-t}\, t^{s-1}\, dt + \int_{0}^{1} \sum_{n=0}^{\infty} \frac{(-1)^n}{n!} t^{n+s-1}\, dt
     = \int_{1}^{\infty} e^{-t}\, t^{s-1}\, dt + \sum_{n=0}^{\infty} \frac{(-1)^n}{n!} \int_{0}^{1} t^{n+s-1}\, dt
     = \int_{1}^{\infty} e^{-t}\, t^{s-1}\, dt + \sum_{n=0}^{\infty} \frac{(-1)^n}{n!} \left[ \frac{t^{n+s}}{n+s} \right]_{0}^{1}
     = \int_{1}^{\infty} e^{-t}\, t^{s-1}\, dt + \sum_{n=0}^{\infty} \frac{(-1)^n}{n!\,(n+s)}

    The sum \sum_{n=0}^{\infty} \frac{(-1)^n}{n!(n+s)} exists for negative s except for s= -1,-2,-3,... This implies that \frac{1}{\Gamma(s/2)} is entire with simple zeros at s= -2,-4,-6,.... The simple pole of \xi(s) at zero is cancelled by the corresponding zero of \frac{1}{\Gamma(s/2)}. As a consequence the only singularity of \zeta(s) is a single pole at s=1. Since

    \displaystyle \zeta(s) = \pi^{s/2}\frac{\xi(s)}{\Gamma(s/2)}

    and \frac{1}{\Gamma(s/2)} is entire with simple zeros at s= -2,-4,-6,..., this implies that for example:

    \displaystyle \zeta(-2) = 0

    We have in general

    \displaystyle \zeta(-2n)=0

    The so-called “trivial zeros” of the Zeta function. They occur at the negative even integers.

  • Zeta summation V

    In a previous post we have computed the meromorphic continuation into the entire complex plane of the Zeta function:

    \displaystyle \zeta(s) = \pi^{s/2}\frac{\xi(s)}{\Gamma(s/2)}

    This analytic continuation allows us to sum the Zeta function in the whole complex plane. For special (‘known’) values of the Xi function we can calculate the Zeta function. For example for

    s = 2 we have \xi(2) = \frac{\pi}{6} and \Gamma(2/2) = \Gamma(1) = 1 and therefore:

    \displaystyle \zeta(2) = \frac{\pi^2}{6}
    \displaystyle \frac{1}{1^2} + \frac{1}{2^2} + \frac{1}{3^2} + \frac{1}{4^2} + ... = \frac{\pi^2}{6}

    For s = 4 we have \xi(4) = \frac{\pi^2}{90} and \Gamma(4/2) = \Gamma(2) = 1:

    \displaystyle \zeta(4) = \frac{\pi^4}{90}
    \displaystyle \frac{1}{1^4} + \frac{1}{2^4} + \frac{1}{3^4} + \frac{1}{4^4} + ... = \frac{\pi^4}{90}

    For s = 6 we have \xi(6) = \frac{2\pi^3}{945} and \Gamma(6/2) = \Gamma(3) = 2:

    \displaystyle \zeta(6) = \frac{\pi^6}{945}
    \displaystyle \frac{1}{1^6} + \frac{1}{2^6} + \frac{1}{3^6} + \frac{1}{4^6} + ... = \frac{\pi^6}{945}

    So, Zeta of all the positive even integers has the form

    \displaystyle\zeta(2n) = (-1)^{n+1} \frac{B_{2n} (2\pi)^{2n}}{2(2n)!}

    Where B_{2n} are Bernoulli numbers.

  • 1 + 2 + 3 + 4 + …

    To calculate the value of the Riemann zeta function at s = -1, we use the functional equation (presented here):

    \displaystyle \zeta(s) = \pi^{s/2} \frac{\xi(s)}{\Gamma(s/2)}

    The Dirichlet series for the Riemann zeta function is defined for \Re(s) > 1 as:

    \displaystyle \zeta(s) = \sum_{n=1}^{\infty} \frac{1}{n^s}

    For s = -1, the formal series (which is divergent in the classical sense) is:

    \displaystyle \zeta(-1) = \sum_{n=1}^{\infty} \frac{1}{n^{-1}}= \sum_{n=1}^{\infty} n = 1 + 2 + 3 + 4 + \dots

    Substituting s = -1 into the formula, we obtain:

    \displaystyle \zeta(-1) = \pi^{-1/2} \frac{\xi(-1)}{\Gamma(-1/2)}

    Using the recurrence property of the Gamma function, \Gamma(z) = \frac{\Gamma(z+1)}{z}, and the fact that \Gamma(1/2) = \sqrt{\pi}:

    \displaystyle \Gamma(-1/2) = \frac{\Gamma(-1/2+1)}{-1/2} = \frac{\Gamma(1/2)}{-1/2} = -2\sqrt{\pi}

    We exploit the functional equation \xi(s) = \xi(1-s), which implies:

    \displaystyle \xi(-1) = \xi(1 - (-1)) = \xi(2)

    From the definition of \xi(s), we have:

    \displaystyle \xi(2) = \pi^{-2/2} \Gamma(2/2) \zeta(2)

    Using \Gamma(1) = 1 and the solution of the Basel problem presented in the previous post \zeta(2) = \frac{\pi^2}{6}:

    \displaystyle \xi(-1) = \pi^{-1} \cdot 1 \cdot \frac{\pi^2}{6} = \frac{\pi}{6}

    Combining these results into our original expression:

    \displaystyle \zeta(-1) =\frac{1}{\sqrt{\pi}} \cdot \frac{\frac{\pi}{6}}{-2\sqrt{\pi}}
    \displaystyle = \frac{\pi}{-12 \cdot (\sqrt{\pi} \cdot \sqrt{\pi})}
    \displaystyle = \frac{\pi}{-12\pi}
    \displaystyle \boxed{\zeta(-1) = -\frac{1}{12}}

    Though surprising at first, turns out to be extremely useful in physics — most notably in string theory and the calculation of the Casimir effect, where the infinite sum 1 + 2 + 3 + 4 + … = −1/12 naturally appears when regularizing the vacuum energy between two conducting plates.

  • Zeta summation: Basel problem

    For s= 2, the zeta series is:

    \displaystyle \zeta(2) = \sum_{n=1}^{\infty} \frac{1}{n^{2}}=  1 + \frac{1}{4} + \frac{1}{9} + \frac{1}{16} + \dots

    We would like to evaluate this series using the representation presented the previous post in we obtain

    \displaystyle \zeta(s) = \pi^{s/2} \frac{\xi(s)}{\Gamma(s/2)}

    For s = 2,

    \displaystyle \zeta(2) = \pi^{2/2} \frac{\xi(2)}{\Gamma(2/2)} = \pi \, \xi(2)

    since \Gamma(1) = 1. We need to evaluate

    \displaystyle \xi(2) = \int_0^\infty \frac{\vartheta(u) - 1}{2} \, du.

    We split the integral as follows:

    \displaystyle \int_0^1 \frac{\vartheta(u) - 1}{2} \, du + \int_1^\infty \frac{\vartheta(u) - 1}{2} \, du

    For the second part, we make the change of variable u = 1/v, so du = -dv/v^2. Using the functional equation of the theta function presented in a previous post.

    \displaystyle \vartheta(u) = u^{-1/2} \vartheta(1/u), \qquad u>0.

    This can be rewritten as

    \displaystyle \vartheta(1/v) = \sqrt{v} \, \vartheta(v)

    one can show (after standard calculations) that the two parts combine to yield a convergent integral. The final evaluation, which relies on the Poisson summation formula applied to the Gaussian, gives

    \displaystyle \xi(2) = \frac{\pi}{6}

    We obtain

    \displaystyle \zeta(2) = \pi \cdot \xi(2) = \pi \cdot \frac{\pi}{6} = \frac{\pi^2}{6}

    Finally

    \displaystyle 1 + \frac{1}{4} + \frac{1}{9} + \frac{1}{16} + \dots = \frac{\pi^2}{6}

    The Basel problem was sloved by Euler in 1735, who showed that the sum equals \displaystyle \frac{\pi^2}{6}.

  • Zeta summation IV

    In the previous post we derived the functional equation of the theta function

    \displaystyle \vartheta(t) = t^{-1/2} \vartheta(1/t), \qquad t>0.

    With t > 0. We observe that:

    \displaystyle \vartheta(t) \leq C t^{-1/2} \text{ as } t \to 0

    note that

    \displaystyle \sum_{n \geq 1}e^{-\pi n^2 t} \leq \sum_{n \geq 1}e^{-\pi n t} \leq Ce^{-\pi t}, t \geq 1

    Remember the definition of the Gamma function. For s >0 the Gamma function is defined by:

    \displaystyle \Gamma(s) = \int_{0}^{\infty} e^{-t} t^{s-1}\,dt
    \displaystyle \Gamma(s/2) = \int_{0}^{\infty} e^{-t} t^{\frac{s}{2}-1}\,dt

    Define u:= \frac{t}{\pi n^2} \implies du = \frac{1}{\pi n^2}\,dt and dt = \pi n^2 \,du

    \displaystyle \Gamma(s/2) = \int_{0}^{\infty} e^{-\pi n^2 u} (\pi n^2 u)^{\frac{s}{2}-1} \pi n^2\,du
    \displaystyle \Gamma(s/2) = \pi^{s/2} n^{s}\int_{0}^{\infty} e^{-\pi n^2 u} u^{\frac{s}{2}-1}\,du
    \displaystyle \pi^{-s/2} \Gamma(s/2) n^{-s} = \int_{0}^{\infty} e^{-\pi n^2 u} u^{\frac{s}{2}-1}\,du \qquad (n \geq 1)

    Observe that for the \vartheta(t) defined above:

    \displaystyle \vartheta(u) = \sum_{n \in \mathbb{Z}} e^{-\pi n^2 u}
    \displaystyle = 2 \sum_{n = 1}^{\infty} e^{-\pi n^2 u} + 1
    \displaystyle \frac{\vartheta(u) -1}{2} = \sum_{n = 1}^{\infty} e^{-\pi n^2 u}

    It follows that:

    \displaystyle \int_{0}^{\infty} u^{\frac{s}{2}-1}  \biggl(\frac{\vartheta(u) -1}{2}\biggr) \,du = \int_{0}^{\infty} u^{\frac{s}{2}-1} \sum_{n = 1}^{ \infty} e^{-\pi n^2 u}  \,du
    \displaystyle = \sum_{n = 1}^{\infty} \int_{0}^{\infty} u^{\frac{s}{2}-1} e^{-\pi n^2 u}\,du
    \displaystyle = \sum_{n = 1}^{\infty} \int_{0}^{\infty} e^{-\pi n^2 u} u^{\frac{s}{2}-1} \,du
    \displaystyle = \sum_{n = 1}^{\infty} \pi^{-s/2} \Gamma(s/2) n^{-s}
    \displaystyle =  \pi^{-s/2} \Gamma(s/2) \sum_{n = 1}^{\infty} n^{-s}
    \displaystyle =  \pi^{-s/2} \Gamma(s/2) \zeta(s)

    Therefore:

    \displaystyle \zeta(s) = \frac{\int_{0}^{\infty} u^{\frac{s}{2}-1}  \biggl(\frac{\vartheta(u) -1}{2}\biggr) \,du}{\pi^{-s/2} \Gamma(s/2) }

    The define the Xi function as :

    \displaystyle \xi(s) := \int_{0}^{\infty} u^{\frac{s}{2}-1}  \biggl(\frac{\vartheta(u) -1}{2}\biggr) \,du

    We finally have

    \boxed{\displaystyle \zeta(s) = \pi^{s/2}\frac{\xi(s)}{\Gamma(s/2)}}

    which is the meromorphic continuation into the complex plane of the Zeta function with a simple pole at s = 1.

  • Zeta summation III

    Recall that in the previous section our goal was to analytically extend the Riemann zeta function.

    To this end, we showed that the Gaussian function e^{-\pi x^2} is invariant under the Fourier transform:

    \displaystyle \mathcal{F}(e^{-\pi x^2})(\xi) = \int_{-\infty}^{\infty} e^{-\pi x^2} e^{-2\pi i x \xi}\,dx = e^{-\pi \xi^2}.

    We recall that the Fourier transform is defined by

    \displaystyle \mathcal{F}(f)(\xi) = \int_{-\infty}^{\infty} f(x)\,e^{-2\pi i x \xi}\,dx.

    For a>0, a change of variables shows that

    \displaystyle \mathcal{F}(f(ax))(\xi) = \frac{1}{a}\,\mathcal{F}\!\left(f\right)\!\left(\frac{\xi}{a}\right).

    Applying this to the Gaussian f(x)=e^{-\pi x^2} with a=\sqrt{t}, we obtain

    \displaystyle \mathcal{F}\!\left(e^{-\pi t x^2}\right)(\xi) = t^{-1/2} e^{-\pi \xi^2/t}.

    Both functions belong to the class \mathcal{S} of functions (The Schwartz class consists of infinitely differentiable functions on \mathbb{R}^n that, together with all their derivatives, decay faster than any polynomial at infinity (rapidly decreasing functions)) and Poisson summation applies:

    \displaystyle \sum_{n\in\mathbb{Z}} f(n) = \sum_{n\in\mathbb{Z}} \mathcal{F}(f)(n).

    In the present case, this yields

    \displaystyle \sum_{n\in\mathbb{Z}} e^{-\pi t n^2} = \sum_{n\in\mathbb{Z}} t^{-1/2} e^{-\pi n^2/t}.

    The left-hand side defines the theta function

    \displaystyle \vartheta(t) := \sum_{n\in\mathbb{Z}} e^{-\pi t n^2}.

    We immediately obtain the functional equation

    \displaystyle \boxed{\vartheta(t) = t^{-1/2} \vartheta(1/t),\qquad t>0.}
  • Zeta summation II

    Now look at:

    \displaystyle f(z) = e^{-\pi z^2}

    where z \in \mathbb{C}. This function is an entire function (differentiable in the entire complex plane). Since this function is entire the Cauchy’s integral theorem applies:

    \displaystyle \int_{\gamma_1} f(z)\,dz = 0

    Where \gamma_1 is any closed contour in the complex plane. Let us consider the contour presented in the figure below:

    \displaystyle \int_{-R}^{R} f(z) \,dz + \int_{R}^{R+ i \xi} f(z) \,dz + \int_{R+ i \xi}^{-R+ i \xi} f(z) \,dz + \int_{-R+ i \xi}^{-R} f(z) \,dz = 0

    For the first term of the integral we set: z(t) = t.
    For the second term: z(t) = R + it.
    For the third term: z(t) = t + i \xi.
    For the fourth term: z(t) = -R + it.
    The integral becomes:

    \displaystyle \int_{-R}^{R} f(t) \,dt + \int_{R}^{R+ i \xi} f(R + it) i\,dt + \int_{R+ i \xi}^{-R+ i \xi} f(t + i \xi) \,dt + \int_{-R+ i \xi}^{-R} f(-R + it) i\,dt =
    \displaystyle \int_{-R}^{R} f(t) \,dt + \int_{0}^{\xi} f(R + it) i\,dt + \int_{R}^{-R} f(t + i \xi) \,dt + \int_{\xi}^{0} f(-R + it) i\,dt =
    \displaystyle \int_{-R}^{R} e^{-\pi t^2} \,dt + \int_{0}^{\xi} e^{-\pi (R + it)^2} i\,dt + \int_{R}^{-R} e^{-\pi (t + i \xi)^2} \,dt + \int_{\xi}^{0} e^{-\pi (-R + it)^2} i\,dt = 0

    Observe that the last term is equal to (using u = -t):

    \displaystyle  \int_{\xi}^{0} e^{-\pi (-R + it)^2} i\,dt = -\int_{-\xi}^{0} e^{-\pi (-R - iu)^2} i\,du

    The term (-R - iu)^2 can be written (R + iu)^2:

    \displaystyle  \int_{\xi}^{0} e^{-\pi (-R + it)^2} i\,dt = -\int_{-\xi}^{0} e^{-\pi (R + iu)^2} i\,du

    now:

    \displaystyle -\int_{-\xi}^{0} e^{-\pi (R + iu)^2} i\,du = - \left( -\int_{0}^{-\xi} e^{-\pi (R + iu)^2} i\,du \right) = \int_{0}^{-\xi} e^{-\pi (R + iu)^2} i\,du

    The second and fourth term cancel and the equation above becomes:

    \displaystyle  \int_{-R}^{R} e^{-\pi t^2} dt + \int_{R}^{-R} e^{-\pi (t + i \xi)^2}\,dt = 0 \\ \int_{-R}^{R} e^{-\pi t^2} dt = - \int_{R}^{-R} e^{-\pi (t + i \xi)^2}\,dt \\ \int_{-R}^{R} e^{-\pi t^2} dt = \int_{-R}^{R} e^{-\pi (t^2 + 2 t i \xi - \xi^2)}\,dt \\ \int_{-R}^{R} e^{-\pi t^2} dt = e^{\pi \xi^2} \int_{-R}^{R} e^{-\pi t^2} e^{-2\pi i t \xi}\,dt \\

    Using

    \displaystyle \lim_{R \to \infty} \int_{-R}^{R} e^{-\pi t^2}\,dt = 1

    Therefore, if \xi \in \mathbb{R}:

    \displaystyle     e^{-\pi \xi^2} = \int_{-\infty}^{\infty} e^{-\pi x^2} e^{-2\pi i x \xi} \,dx

    We conclude from this post and the previous post that the Gaussian function e^{-\pi \xi^2} is its own Fourier transform.