## Introduction

Practically all complex adaptive systems exhibit fat-tailed distribution functions in the statistics of their dynamical variables. Often these distribution functions are exact or almost exact asymptotic power-laws, $$p(x)\sim {x}^{-\lambda }$$. While Gaussian statistics can often be traced back to a single origin, the central limit theorem, for the origin of power laws there exist several routes. These include (i) Yule-Simon processes (preferential attachment)1,2,3, (ii) multiplicative processes with constraints, see refs 4 and 5, (iii) criticality6, (iv) self-organized criticality and cascading processes7,8,9,10, (v) constraint optimization11,12,13, and (vi) sample space reducing (SSR) processes14,15,16. Most of these mechanisms are able to explain specific values of the exponent λ or a range of exponents. None of them however explain the full range of exponents from zero to infinity, λ [0, ∞) in a straightforward way. Depending on the generative process exponents belong to different ranges. For example, exponents near critical points are often fractional, and within the range $$\lambda \in [\mathrm{1/2},\,\mathrm{5/2}]$$ 6. Many exponents for avalanche processes are found within a range of λ (0, 3)17. Some processes, like the preferential attachment, can formally explain a wide range of exponents, although deviations from the standard values λ (2, 3.5) are hard to map to realistic underlying stochastic dynamics18. Here we show that the combination of cascading processes with SSR processes is able to do exactly that, to provide a single one-parameter model that produces the full spectrum of all possible scaling exponents. The parameter is nothing but the multiplication ratio of the cascading processes. Finally, we show that generic disintegration processes can be mapped one-to-one to SSR cascades with an over imposed condition of conservation of whatever magnitude is represented by the states. This mapping allows us to derive a remarkable result: The histogram of visits to each state follows an exponent −2, regardless the multiplication parameter. This result may have important consequences in order to understand generic properties of disintegration processes and the ubiquity of the exponent −2 in nature.This, for example, allows us to recover Fermi’s classic result on the energy spectrum of cosmic rays, only appealing to combinatorial properties of the cascade. In addition, we provide a rigorous proof of that result in the appendix A, as a new contribution to the study of the random partition of the interval.

Cascading processes have played an important role in the understanding of power-law statistics in granular media7,8,9,10, 19, earth quakes20,21,22, precipitation23, 24, dynamics of combinatorial evolution25, or failure in networks26,27,28. The scaling exponents of the probability distribution functions of quantities such as avalanche sizes, energy distributions, visiting times, event durations, etc., are found within a relatively narrow band. Cascading processes are often history-dependent processes in the sense that for a particular event taking place the temporal order of microscopic events is important. Recent progress in the understanding of the generic statistics of history-dependent processes and their relation to power laws was made in refs 14,15,16 and 29. Maybe the simplest history-dependent processes are the sample space reducing (SSR) processes14, which explain the origin of scaling in a very simple and intuitive way. They have been used in various applications in computational linguistics30, fragmentation processes14, and diffusion on directed networks and search processes15.

A SSR process is a stochastic processes whose sample space reduces as it evolves in time. They can be depicted in a simple way, see Fig. 1(a). Imagine a set of N states in a system, labelled by i = 1, 2, …, N. The states are ordered by the label. The only rule that defines the SSR process is that transitions between states may only occur from higher to lower labels. This means that transition from state ji is possible only if label j > i. When the lowest state i = 1 is reached, the process stops or is re-started. It was shown in ref. 14 that this dynamics leads to a Zipf’s law in the frequency of state visits, i.e. the probability to visit state i is given by p(i) = i −1. This scaling law is extremely robust and occurs for a wide class of prior probabilities15. We now show that the combination of SSR processes with the simplest cascading process allows us to obtain a mechanism that can produce power laws with any scaling exponent. We will comment on how the model can be used to recover the Fermi’s classic result on the cosmic ray spectrum31, 32. This is possible by imposing energy conservation on SSR cascading processes. We discuss other potential cases where the theory of SSR cascades might apply.

To define SSR cascades, imagine a system with a set of N ordered states, each of which has a prior probability of appearing, q 1, …, q N−1, being state N the starting point of the cascade. The process starts at t = 0 with μ balls at state N jumping to any state i 1, …, i μ  < N with a probability proportional to q i , …, q μ , respectively. Suppose that the μ balls landed on states i 1,..i μ , respectively. At the next timestep, t = 1, each of these μ balls divide into μ new balls which all jump to any state below their original state. The multiplicative process continues downwards. Whenever a ball hits the lowest state, it is eliminated from the system. Effectively we superimpose a multiplicative process that is characterized by the multiplicative parameter μ, and the SSR process described above, see Fig. 1(c). The case μ = 1 is exactly the standard SSR process, where no new elements are created, and the case μ < 1 corresponds to the noisy SSR, where there is the possibility that the process gets cut at some step, see Fig. 1(b).

The derivation of the visiting distribution of this cascading SSR process follows the arguments found in ref. 15. We first define the cumulative prior distribution function g(k),

$$g(k)=\sum _{i\le k}\,{q}_{i}.$$

Without a multiplication factor μ the transition probabilities p(i|j) determine the probability to reach state i at timestep t + 1, given that the system is in state j at time t, and are given by

$$p(i|j)=\{\begin{array}{ll}\frac{{q}_{i}}{g(j-\mathrm{1)}} & {\rm{for}}\,i < j\\ 0 & {\rm{for}}\,i\ge j.\end{array}$$
(1)

In a SSR cascade, if there is an element sitting at state j at time t, there are now μ trials to reach any state i < j at t + 1. Since the number of particles is not conserved throughout the process, we talk about the expected number of jumps from j to i. Since the jumps from j to i of each ball is independent, the expected number of jumps from j to i we denoted by n(ji) can be approximated as follows:

$$n(j\to i)=\mu p(i|j).$$
(2)

We denote the expected number of elements that will hit state i in a given SSR cascade by n i . Up to a factor the sequence n 1, …, n N is identical to the histogram of visits. From equations (1) and (2) we get

$${n}_{i}=\sum _{j > i}\,n(j\to i){n}_{j}=\mu {q}_{i}\,\sum _{j > i}\,\frac{{n}_{j}}{g(j-\mathrm{1)}}.$$

By subtracting n i+1 − n i and after re-arranging terms we find

$$\frac{{n}_{i+1}}{{q}_{i+1}}\,(1+\mu \frac{{q}_{i+1}}{g(i)})=\frac{{n}_{i}}{{q}_{i}},$$

or, when applied iteratively

$${n}_{i}={n}_{1}\,\frac{{q}_{i}}{{q}_{1}}\,\prod _{1 < j\le i}\,{(1+\mu \frac{{q}_{j}}{g(j-\mathrm{1)}})}^{-1}.$$

Since μ(q j )/(g(j − 1)) is typically small the product term is well approximated by

$$\begin{array}{rcl}\prod _{1 < j\le i}\,{(\cdots )}^{-1} & = & \exp \,[-\sum _{1 < j\le i}\,\mathrm{log}\,(1+\mu \frac{{q}_{j}}{g(j-\mathrm{1)}})]\\ & \approx & \exp \,(-\mu \sum _{1 < j\le i}\,\frac{{q}_{j}}{g(j-\mathrm{1)}})\\ & \approx & \exp \,(-\mu {\int }_{1}^{i}\,\frac{dg}{dx}\frac{1}{g(x)}dx)\\ & \approx & \exp \,(-\mu \,\mathrm{log}\,\frac{g(i)}{{q}_{1}})\\ & = & {(\frac{g(i)}{{q}_{1}})}^{-\mu },\end{array}$$

where we used $${q}_{j}\sim {dg/dx|}_{j}$$ and $$\mathrm{log}\,\mathrm{(1}+x)\sim x$$. Finally, we have

$${n}_{i}\sim \frac{{n}_{1}}{{q}_{1}^{1-\mu }}\,(\frac{{q}_{i}}{g{(i)}^{\mu }}).$$
(3)

For equal prior probabilities, $${q}_{i}=\frac{1}{N-1}$$ for all states, we get the expected visiting probability to be

$${n}_{i}\sim {i}^{-\mu }.$$

The multiplication factor μ becomes the scaling exponent, for μ = 1 the standard SSR processes is recovered14. Figure (2) shows numerical results which are in perfect agreement with the theoretical predictions. Note that the argument also holds for non-integer μ, where, on average, μ balls are created at every step. In the numerical implementation, a non-integer μ is introduced as follows: Let $$\mu =\lfloor \mu \rfloor +\delta$$, with δ < 1. Then, with probability δ, $$\lfloor \mu \rfloor +1$$ balls are created and, with probability 1 − δ, $$\lfloor \mu \rfloor$$ balls are created. Also the case of multiplication factors μ < 1 are possible, reproducing the previously defined noisy SSR case, see Fig. (1b) and ref. 14. In this situation at each step the process can be restarted with the probability 1 − μ.

We numerically compute the cascade size distribution as a function of the number of states N and μ. For a given realisation of a cascade ψ with initial sample space N and multiplicative parameter μ, starting with a single element at N, we define the cascade size, s μ,N (ψ), as the number of elements of the cascade ψ that reach state 1, $${n}_{1}^{(\psi )}$$. Numerical analysis suggests that the cascade size distribution f(s m,μ ) can be well approximated by a Γ distribution33. For the sake of simplicity, we drop the subscripts μ and N for s. We thus find a purely phenomenological equation that reads.

$$f(s)\propto {s}^{\alpha -1}{e}^{-\lambda s},\quad \langle s\rangle \propto \frac{{N}^{\mu }}{{e}^{a\mu }},\quad {\sigma }^{2}\propto \frac{{N}^{b\mu }}{{e}^{(\frac{1}{2}+a)\mu }}\langle s\rangle ,$$
(4)

with a = 0.82, b = 0.9, α = 〈s2/σ 2, λ = 〈s〉/σ 2. Numerical results and fits are shown in Fig. (3). The inset shows that the approximation for 〈s〉 is highly accurate.

### Energy conservation and the prevalence of −2 exponent

In the sequel we derive the statistics of visits to the states of our system when our cascade observes an energy conservation constraint. Remarkably, we see that the histogram of visits to each state along the whole cascading process follows a power-law of exponent −2, regardless the multiplication parameter of the avalanche. The strategy followed here is based on the partition through renormalization and works basically as follows: out of an interval $$\mathrm{[0,}\,1]$$, a partition is performed by throwing μ random numbers between 0 and 1 and then renormalizing them such that their sum is 1 as in Fig. (4a,b)34. Our strategy can be seen as a particular choice of Dirichlet partitioning of the interval35. This strategy is different from the random selection of breaking points of the interval, described, e.g., in refs 33 and 36. In the appendix A we provide a complete proof of our result.

To study SSR cascades with a superimposed conservation law, let us assume, with any loss of generality, that the states 1, 2, …, N are associated with energy levels

$${E}_{1},{E}_{2},\ldots ,{E}_{N}.$$

Energy conservation imposes the following constraint: If a particle with energy E and splits into μ particles i 1, i 2, … i μ , with respective energies $${E}_{{i}_{1}},{E}_{{i}_{2}},\ldots ,{E}_{{i}_{\mu }}$$, then:

$$\sum _{k=1}^{\mu }\,{E}_{{i}_{k}}=E.$$
(5)

We are interested in the energy spectrum, i.e. number of observations of particles at a particular energy level at any point of the cascading process, n(E).

The first task is to impose the energy conservation constraint given in equation (5) in the schema of transition probabilities of the SSR cascade. To compute it we use a rescaling technique and we will assume that the energy spectrum is continuous. The rescaling technique is outlined in Fig. (4). Let us ignore energy conservation for the moment and define a continuous uniform random variable u on the interval $$\mathrm{[0,}\,1]$$. Let u 1, u 2, …, u μ be μ independent realisations of u, see Fig. (4a). Let us suppose that we are at level E. From this sequence of random variables one can derive the target sites of the newly created particles in a SSR avalanche with multiplicative parameter μ as

$${u}_{1}\cdot E,{u}_{2}\cdot E,\ldots ,{u}_{\mu }\cdot E.$$

This is the continuous version of what we described in section 2.1. Now we define a new random variable, ϕ μ , which is the sum of μ realisations of the random variable u:

$${\varphi }_{\mu }=\sum _{k\le \mu }\,{u}_{k}.$$

The sum of μ realisations of a random variable u uniformly distributed on the interval $$\mathrm{[0,}\,1]$$, ϕ μ , follows the Irwin-Hall distribution, f μ (ϕ μ )33. This means that one can construct, for each μ realisations of the random variable u a rescaled sequence, see Fig. (4b)

$${\varphi }_{\mu }^{-1}{u}_{1}\cdot E,\ldots ,{\varphi }_{\mu }^{-1}{u}_{\mu }\cdot E,$$

such that sum up to the total energy E,

$${\varphi }_{\mu }^{-1}\sum _{i\le \mu }\,{u}_{i}\cdot E=E.$$

Thus, by imposing energy conservation we actually expect the following sequence of rescaled energies $${E}_{{i}_{1}},{E}_{{i}_{2}},\ldots ,{E}_{{i}_{\mu }}$$ for the emerging particles, where

$${E}_{{i}_{k}}={\varphi }_{\mu }^{-1}({u}_{k}\cdot E\mathrm{).}$$

This rescaling approach assumes that the μ new particles behave independently. The crucial issue is to map this process into a cascade, see Fig. (4c,d). To approach this problem, we first study the expected number of particles that jump to a given state E if a given value of ϕ μ occurs. We then average over all potential values of ϕ μ . We assume that the expected number of particles from EE′, n(EE′) goes as ~μp(E′|E). Taking into account the rescaling imposed by energy conservation, $$p(E^{\prime} |E)\sim \frac{{\varphi }_{\mu }}{E}$$, see Fig. (4)–on has that:

$$n(E\to E^{\prime} ,{\varphi }_{\mu })=\{\begin{array}{ll}\mu \frac{{\varphi }_{\mu }}{E} & {\rm{for}}\,{\varphi }_{\mu }E^{\prime} < E\\ 0 & {\rm{for}}\,{\varphi }_{\mu }E^{\prime} \ge E.\end{array}$$
(6)

Assuming a continuous spectrum of energies, one has, for a given value of ϕ μ , that the expected number of particles that will visit state E at some point of the cascade, n(E,ϕ μ ) is:

$$n(E,{\varphi }_{\mu })={\int }_{{\varphi }_{\mu }E}^{\infty }\,\mu \frac{{\varphi }_{\mu }}{E^{\prime} }n(E^{\prime} )dE^{\prime} ,$$

where n(E′) is the total number of particles that are expected to visit state E′ during the cascade. n(E) will be obtained by averaging n(E,ϕ μ ) over all potential values of ϕ μ , distributed as the Irwin-Hall distribution, f μ :

$$n(E)={\int }_{0}^{\mu }\,{f}_{\mu }({\varphi }_{\mu })\{{\int }_{{\varphi }_{\mu }E}^{\infty }\,\mu \frac{{\varphi }_{\mu }}{E^{\prime} }n(E^{\prime} )dE^{\prime} \}\,d{\varphi }_{\mu }.$$
(7)

Differentiating n(E), one arrives at the following equation with displacement:

$$\frac{dn}{dE}=-\mu {\int }_{0}^{\mu }\,{\varphi }_{\mu }{f}_{\mu }({\varphi }_{\mu })n({\varphi }_{\mu }E)d{\varphi }_{\mu }.$$

Assuming that n(E) E α, we arrive at the following self-consistent equation for α:

$$\alpha =\mu {\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }.$$
(8)

whose only solution, for large μs converges to α = 2, leading to the general result.

$$n(E)\propto {E}^{-2}.$$
(9)

In the appendix A we provide a rigorous derivation of this result. In spite of the asymptotic nature of the proof given in the appendix A, numerical simulations show an excellent agreement with this theoretical prediction, even for α small. In Fig. (5) we show the frequency plots for 105 avalanches with μ = 2.5, 3.5, 4.5, and the convergence of the histograms to E −2 can be perfectly appreciated. This result is the same that is obtained from Fermi’s particle acceleration model to explain the spectrum of cosmic rays31, 32. Here we derived it on the basis of simple combinatorial reasonings of SSR processes with a superimposed constraint.

## Discussion

### Appendix A: Derivation of exponent −2 for cascades with energy conservation

We derive the main result of section IIB. The strategy followed is summarised in Fig. (4). An alternative view is given in Fig. (6) in this appendix.

We start with the definition of the Irwin-Hall distribution: Let u be a random variable whose probability density is uniform in the interval $$\mathrm{[0,}\,1]$$. Let u 1, …, u μ be a sequence of independent drawings of the random variable u and ϕ(μ) a random variable defined over the interval $$\mathrm{[0,}\,\mu ]$$ as:

$${\varphi }_{\mu }=\sum _{k\le \mu }\,{u}_{k}.$$
(A1)

The probability density that governs the random variable ϕ μ is the Irwin-Hall distribution. Now we go to equation (7),

$$n(E)={\int }_{0}^{\mu }\,{f}_{\mu }({\varphi }_{\mu })\,\{{\int }_{{\varphi }_{\mu }E}^{\infty }\,\mu \frac{{\varphi }_{\mu }}{E^{\prime} }n(E^{\prime} )dE^{\prime} \}\,d{\varphi }_{\mu }.$$

Differentiating,

$$\begin{array}{rcl}\frac{dn}{dE} & = & \frac{d}{dE}\,{\int }_{0}^{\mu }\,{f}_{\mu }({\varphi }_{\mu })\,\{{\int }_{{\varphi }_{\mu }E}^{\infty }\,\mu \frac{{\varphi }_{\mu }}{E^{\prime} }n(E^{\prime} )dE^{\prime} \}\,d{\varphi }_{\mu }\\ & = & {\int }_{0}^{\mu }\,{f}_{\mu }({\varphi }_{\mu })\frac{d}{dE}\,\{{\int }_{{\varphi }_{\mu }E}^{\infty }\,\mu \frac{{\varphi }_{\mu }}{E^{\prime} }n(E^{\prime} )dE^{\prime} \}\,d{\varphi }_{\mu }\\ & = & -\mu {\int }_{0}^{\mu }\,{\varphi }_{\mu }{f}_{\mu }({\varphi }_{\mu })n({\varphi }_{\mu }E)d{\varphi }_{\mu },\end{array}$$

one arrives at the following equation with displacement:

$$\frac{dn}{dE}=-\mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }{f}_{\mu }({\varphi }_{\mu })n({\varphi }_{\mu }E)d{\varphi }_{\mu }.$$

Assuming that n(E) E α, we arrive at the following self-consistent equation for α:

$$\alpha =\mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }.$$

This is equation (8) and that is what we have to solve. The following theorem states that the solution in the limit of large μ’s is α = 2, independent of μ. To demonstrate that we need to proof 5 lemmas. After the demonstration, we approach the solution α → 2 using a mean field approach. Finally, we report a side observation concerning the behaviour of the average value of a random variable following the Irwin-hall distribution.

Theorem: The only α satisfying the following equation:

$$\mathop{\mathrm{lim}}\limits_{\mu \to \infty }\,\mu {\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }=2.$$
(A2)

is α = 2.

To prove this theorem, we observe first observe that a random variable following the Irwin-Hall distribution converges to a random variable following a normal distribution with average $$\frac{\mu }{2}$$ and standard deviation $$\sqrt{\frac{\mu }{12}}$$.

Lemma 1: Let ϕ μ a random variable following the Irwin-Hall distribution. Let Y be a random variable following a normal distribution centred at 0 and with standard deviation 1. Then, the following limit holds:

$${\varphi }_{\mu }\to \frac{\mu }{2}+\sqrt{\frac{\mu }{12}}Y\mathrm{(0},1).$$

in probability.

Proof: The average value of a uniformly distributed random variable u is $${\mathbb{E}}(u)=\frac{1}{2}$$ and standard deviation $$\sigma =\sqrt{\frac{1}{12}}$$. By of the central limit theorem, one has that, for an i.i.d. sequence of random variables u 1, …, u n :

$$\frac{{\sum }_{i\le \mu }{u}_{i}-\frac{\mu }{2}}{\sqrt{\frac{\mu }{12}}}\to Y,$$

being Y a random variable following a normal distribution centred at 0 and with standard deviation 1. Therefore, by realising that ϕ μ is actually a sum of μ i.i.d random variables u, one has:

$${\varphi }_{\mu }\to \frac{\mu }{2}+\sqrt{\frac{\mu }{12}}Y\mathrm{(0},1),$$

as we wanted to demonstrate.☐

This implies that the Irwin-Hall distribution f μ can be fairly approached by a normal distribution with mean $$\frac{\mu }{2}$$ and standard deviation $$\sqrt{\frac{\mu }{12}}$$, Φ μ (x). However, one must be careful with this approach: It can lead the integral that we want to solve, $${\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }$$, to a singularity at 0 which is inexistent in the Irwin-Hall distribution. Therefore, for the interval [0, 1) we will maintain the original form of the distribution. In the following lemma we demonstrate that this has no impact in the limit of large μ’s.

Lemma 2: Let Φ μ be a normal distribution with mean at $$\frac{\mu }{2}$$ and standard deviation $${\sigma }_{\mu }=\sqrt{\frac{\mu }{12}}$$. Then, $$(\forall \epsilon > \mathrm{0)}\,(\exists N):(\forall \mu > N)$$

$$|{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }-{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{{\rm{\Phi }}}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }| < \epsilon .$$

Proof: From lemma 1 we know that $$(\forall \epsilon ^{\prime} > \mathrm{0)}\,(\exists N):(\forall S\mu > N)$$

$$|{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }-{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{{\rm{\Phi }}}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }| < \epsilon ^{\prime} .$$

Now we observe that the Irwin-Hall distribution can be defined per intervals using different polynomials. In the case of the interval [0, 1), the polynomial reads:

$${f}_{\mu }({\varphi }_{\mu })=\frac{1}{(\mu -\mathrm{1)!}}{\varphi }_{\mu }^{\mu -1};\quad {\varphi }_{\mu }\in [0,1).$$
(A3)

Computing directly the integral, one has:

$$\begin{array}{rcl}{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu } & = & {\int }_{0}^{1}\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }\\ & & +{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu },\end{array}$$

where the first integral, according to equation (A3) leads to:

$$\begin{array}{rcl}{\int }_{0}^{1}\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu } & = & \frac{1}{(\mu -\mathrm{1)!}}{\int }_{0}^{1}\,{\varphi }_{\mu }^{\mu -\alpha }d{\varphi }_{\mu }\\ & = & \frac{1}{(\mu -\mathrm{1)!(}\mu -\alpha -\mathrm{1)}}.\end{array}$$

Now take δ (0, 1) and define $$\epsilon (\mu ,\,\delta )$$ as:

$$\epsilon (\mu )\equiv \frac{1+\delta }{(\mu -\mathrm{1)!(}\mu -\alpha -\mathrm{1)}}$$

From lemma 1, $$(\forall \epsilon ^{\prime} > \mathrm{0)}\,(\exists N):(\forall \mu > N)$$ we can define the following bound:

$$|{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }-{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{{\rm{\Phi }}}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }| < \epsilon ^{\prime} +\epsilon (\mu ,\delta ).$$

Finally, we observe that

$$\mathop{\mathrm{lim}}\limits_{\mu \to \infty }\,\epsilon (\mu ,\delta )=0,$$

which demonstrates the lemma.☐

Lemma 3: The function of G(α) defined by the integral

$$G(\alpha )={\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu },$$

is strictly decreasing.

Proof: It is enough to compute the derivative:

$$\begin{array}{rcl}\frac{d}{d\alpha }{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu } & = & {\int }_{1}^{\mu }\,(\frac{d}{d\alpha }{\varphi }_{\mu }^{1-\alpha })\,{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu }\\ & = & -{\int }_{1}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }\,\mathrm{log}\,{\varphi }_{\mu }{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu }\\ & < & 0,\end{array}$$

since the term inside the integral, $${\varphi }_{\mu }^{1-\alpha }\,\mathrm{log}\,{\varphi }_{\mu }{\rm{\Phi }}({\varphi }_{\mu })$$, is strictly positive in the interval (1, μ).☐

Now take a monotonously increasing function that grows slower than the standard deviation $${\sigma }_{\mu }=\sqrt{\frac{\mu }{12}}$$, φ(μ). For convenience, we define define it as:

$$\phi (\mu )\equiv {(\frac{\mu }{12})}^{\frac{1}{4}}.$$
(A4)

Clearly:

$$\mathop{\mathrm{lim}}\limits_{\mu \to \infty }\,{\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu }=1.$$
(A5)

Now suppose, that α = 2. Then, thanks to Lemma 2, one has that $$(\forall \epsilon > \mathrm{0)}\,(\exists N):(\forall \mu > N)$$

$$|{\int }_{0}^{\mu }\,\frac{{f}_{\mu }({\varphi }_{\mu })}{{\varphi }_{\mu }}\,d{\varphi }_{\mu }-{\int }_{1}^{\mu }\,\frac{{{\rm{\Phi }}}_{\mu }({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }| < \epsilon .$$

From this we derive the third lemma of our demonstration:

Lemma 4: $$(\forall \epsilon > \mathrm{0)}\,(\exists N):(\forall \mu > N),$$

$$|{\int }_{1}^{\mu }\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }-{\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }| < \epsilon .$$

Proof: We need to compute the parts that fall outside the integration limits and see that their contribution vanishes. First, we see that:

$$\begin{array}{lll}{\int }_{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}^{\mu }\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu } & < & (\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )){e}^{-{\phi }^{2}(\mu )-O(\mathrm{log}\mu )}\\ & < & (\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )){e}^{-\sqrt{\mu }}.\end{array}$$

Analogously,

$${\int }_{1}^{\frac{\mu }{2}-{\sigma }_{\mu }\varphi (\mu )}\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu } < (\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )){e}^{-\sqrt{\mu }}.$$

Now we define:

$$\begin{array}{c}{\epsilon }_{1}(\mu )=(\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )){e}^{-\sqrt{\mu }},\\ {\epsilon }_{2}(\mu )=(\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )){e}^{-\sqrt{\mu }},\end{array}$$

$$|{\int }_{1}^{\mu }\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }-{\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }| < {\epsilon }_{1}(\mu )+{\epsilon }_{2}(\mu ),$$

demonstrating the lemma.☐

Corollary of Lemma 4: $$(\forall \epsilon > \mathrm{0)}\,(\exists N):(\forall \mu > N),$$

$$|{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }-{\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }| < \epsilon .$$

Proof: By direct application of lemmas 2 and 4.☐

Now we define the following functions of the limits of the integral $${\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,\ldots$$:

$$\begin{array}{c}{r}_{1}(\mu )\equiv {(\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu ))}^{-1}=\frac{2}{\mu }-\frac{2{\sigma }_{\mu }{\phi }_{\mu }}{\mu (\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu ))},\\ {r}_{2}(\mu )\equiv {(\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu ))}^{-1}=\frac{2}{\mu }+\frac{2{\sigma }_{\mu }{\phi }_{\mu }}{\mu (\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu ))}.\end{array}$$

Clearly,

$${r}_{1,2}(\mu )\sim \frac{2}{\mu }+{\mathscr{O}}\,({\mu }^{-\frac{5}{4}}),$$
(A6)

where the subscript 1,2 means that both functions satisfy the property.

Lemma 5: $$(\forall \epsilon > \mathrm{0)}\,(\exists N):(\forall \mu > N)$$

$$|{r}_{\mathrm{1,2}}(\mu )\cdot \mu {\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu }-\mu {\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }| < \epsilon .$$

Proof: We first observe that, by substituting $${\varphi }_{\mu }^{-1}$$ by the integration limits, we have the following chain of inequalities, in terms of the above defined functions r 1,2(μ):

$$\begin{array}{lll}{r}_{2}(\mu )\cdot {\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu } & < & {\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,\frac{{\rm{\Phi }}({\varphi }_{\mu })}{{\varphi }_{\mu }}\,d{\varphi }_{\mu }\\ & < & {r}_{1}(\mu )\cdot {\int }_{\frac{\mu }{2}-{\sigma }_{\mu }\phi (\mu )}^{\frac{\mu }{2}+{\sigma }_{\mu }\phi (\mu )}\,{\rm{\Phi }}({\varphi }_{\mu })d{\varphi }_{\mu }.\end{array}$$

Therefore, is enough to demonstrate that $$(\forall \epsilon ^{\prime} > \mathrm{0)}\,(\exists N):(\forall \mu > N)$$

$$|\mu {r}_{1}(\mu )-\mu {r}_{2}(\mu )| < \epsilon ^{\prime} .$$

This can be proven directly from equation (A6), leading to:

$$|\mu {r}_{1}(\mu )-\mu {r}_{2}(\mu )|\sim {\mathscr{O}}\,({\mu }^{-\frac{1}{4}}),$$

which demonstrates the lemma.☐

Collecting lemmas 1, 2, 4, and 5, we have demonstrated that, under the assumption that α = 2,

$$\mu {r}_{1}(\mu )\to \mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }.$$

From equation (8), the only remaining issue is to demonstrate the consistency of our hypothesis is that indeed $${\mathrm{lim}}_{{\mu }_{\to }\infty }\,\mu {r}_{1}(\mu )=2$$. It is not difficult to check, from equation (A6), that:

$$\mathop{\mathrm{lim}}\limits_{\mu \to \infty }\,\mu {r}_{1}(\mu )=\mathop{\mathrm{lim}}\limits_{\mu \to \infty }\,[2+{\mathscr{O}}\,({\mu }^{-\frac{1}{4}})]=2.$$

So far we have demonstrated that the solution α = 2 is consistent with the statement of the theorem. Now it remains to demonstrate that this is the only solution. To see that, we observe that thanks to lemma 3, we know that the function μG(α) is decreasing. In addition, we have proven that the statement of the theorem is consistent for α = 2. Therefore, if α = 2 + β, with β > 0, then:

$$\mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu } < 2+\beta ,$$

which contradicts the statement of the theorem. The same happens if one imposes if α = 2 − β, with β > 0, since one gets:

$$\mu {\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu } > 2-\beta ,$$

which is, again inconsistent, thereby proving the lemma.☐

By direct application of equation (8), Lemma 5 puts the last piece to demonstrate the theorem.☐

### Mean-field approach

In a less rigorous way, we observe that we can approach the solution as follows: We know that the expected value of a random variable ϕ μ following the Irwin-Hall distribution f μ is:

$${\mathbb{E}}({\varphi }_{\mu })=\frac{\mu }{2}.$$

Now assume that $${\varphi }_{\mu }\approx \frac{\mu }{2}$$. This implies that, in the integral of the statement of the theorem, equation (A2), we replace f μ (ϕ μ ) by $$\delta \,({\varphi }_{\mu }-\frac{\mu }{2})$$, where δ is the Dirac δ function:

$$\mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }\approx \mu \,{\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }\delta \,({\varphi }_{\mu }-\frac{\mu }{2})\,d{\varphi }_{\mu }.$$

Solving the integral, and thanks to equation (8), we obtain the following relation:

$$\frac{\alpha }{\mu }={(\frac{\mu }{2})}^{1-\alpha },$$

whose only solution is α = 2.

We end observing that the theorem that we demonstrated has a curious consequence: Let ϕ μ be a random variable following the Irwin-Hall distribution. We observe that, if α = 2, then:

$$\begin{array}{rcl}\frac{\alpha }{\mu } & = & {\int }_{0}^{\mu }\,{\varphi }_{\mu }^{1-\alpha }{f}_{\mu }({\varphi }_{\mu })d{\varphi }_{\mu }\\ & = & {\int }_{0}^{\mu }\,\frac{{f}_{\mu }({\varphi }_{\mu })}{{\varphi }_{\mu }}d{\varphi }_{\mu }\\ & = & {\mathbb{E}}\,(\frac{1}{{\varphi }_{\mu }})\\ & \to & \frac{2}{\mu }.\end{array}$$

We know that $${\mathbb{E}}({\varphi }_{\mu })=\frac{\mu }{2}$$. Therefore, a direct consequence of the theorem is that:

$${\mathbb{E}}\,(\frac{1}{{\varphi }_{\mu }})\to \frac{1}{{\mathbb{E}}({\varphi }_{\mu })}.$$