Abstract
We show aging of Glauber-type dynamics on the random energy model, in the sense that we obtain the annealed scaling limits of the clock process and of the age process. The latter encodes the Gibbs weight of the configuration occupied by the dynamics. Both limits are expressed in terms of stable subordinators.
Similar content being viewed by others
1 Introduction
Out of equilibrium statistical mechanics is notoriously so diverse (as, say, non-elephant biology is) that a general theory is hardly conceivable. Nevertheless, and although it might seem like a paradox at first sight, physicists in the 1990s suggested that the dynamical properties of disordered statistical mechanical systems (spin glasses) should be governed by a common general scheme now known under the name of aging.
From a mathematical perspective, at least as a first approximation, proving “aging” basically boils down to showing that a certain additive functional of a (family of) Markov chain(s) scales to a stable subordinator. In spite of the universality of the aging phenomenon claimed by physicists, until this paper, only one specific dynamics, called RHT, could be rigorously handled. Thus, as pointed out in their conclusion by the authors of [6, “Even in the case of the REM, we do not know at this point how to deal with different, and more usual, dynamics [...] This problem remains one of the great challenges in the field.”
It is the aim of the present paper to tackle the challenge, and show that the aging picture is actually valid for a large class of dynamics on the REM.
The random energy model (REM) was introduced in [22, 23] as a simple model of a spin glass. It is a mean-field model living on the hypercube \(V_N=\{0,1\}^N\), and the energies \((E_x)\) associated with each configuration \(x \in V_N\) are taken to be i.i.d. random variables. We assume that they are distributed as the positive part of a standard Gaussian random variable. The Gibbs weights of the model are thus of the form
where \(\beta > 0\) is the inverse temperature.
Unlike other spin glass models where the energies display some complicated correlation, as for instance the SK model, the statics of the REM may be described in detail using results on the extremes of families of independent Gaussian random variables [29, 38]. Roughly speaking, the energy landscape of the REM may be thought of as consisting of a bulk of configurations of energies of order \(1\), with a few scattered deep wells where \(\sqrt{N} E_x\) is of order \(cN\) for some constant \(c\). These deep wells carry most of the weight of the Gibbs measure (for a constant \(c\) which actually depends on \(\beta \)).
The simple structure of the energies in the REM also made it possible to study its relaxation to equilibrium under a Metroplis or Glauber dynamics. Asymptotics of the spectral gap are known [24]. We refer to [35] and [36] for further results on thermalization times.
It is believed that, for a large class of reversible dynamics, the off-equilibrium dynamics observed before thermalization should be described using the language of aging. A crude picture of the phenomenon of aging goes as follows: on well-chosen time and depth scales, the dynamics should spend most of its time in a few deep wells. Before it reaches equilibrium, one may think that the dynamics is transient: it finds a deep well, spends some amount of time around it and then moves through the bulk to a second well, and so on, without ever returning to a previously visited well. In such a scenario, at a given time \(t\), the dynamics is with high probability in a deep well whose depth is of order \(t\). Thus one can evaluate the “age” of the dynamics from the value of the energy of the current configuration.
One way to quantify this aging phenomenon is by computing correlation functions. One may for instance try to derive asymptotics of the probability that the dynamics stays in a small region of its state space between times \(t\) and \(t+s\). From the discussion above, it follows that a reasonable choice is to take \(s\) proportional to \(t\). As we expect the main contribution to such a correlation function to arise from deep wells, and since the statistics of extremes of exponentials of Gaussian fields are given by Poisson point processes with polynomial tails, it is natural to believe that asymptotics of correlation functions should be related to stable subordinators. More precisely, back-of-the-envelope computations show that correlation functions should follow the arc-sine law.
More recently, a different and more ambitious description of aging has been proposed that, instead of correlation functions, focuses on deriving the scaling limit of what we call the age process. This process gives, at any time, the value of the Gibbs weight of the configuration occupied by the dynamics. Once again, it is easy to guess how these should be related to stable subordinators. In the transient regime we are considering now, the scaling limit of the age process should be universal and given by the jumps of a stable subordinator computed at the inverse of a (different and correlated) subordinator, as in [28].
Although this description of the dynamics may seem simple, proving that it is correct turned out to be a mathematically difficult problem. As far as the REM is concerned, so far, rigorous results had only been obtained for one specific dynamics called the random hopping times (RHT) dynamics. It is the aim of the present paper to go beyond this limitation.
In the RHT dynamics, the process stays at a given configuration an exponentially distributed time whose parameter is the inverse of its Gibbs weight, and then jumps to any neighbouring configuration with uniform probability. In other words, in the RHT dynamics, one spin flips at a time, and the spin that flips is chosen uniformly at random among the \(N\) possibilities, independently of the past, independently of the current configuration and independently of the energy landscape. This feature implies that the trajectory (i.e. the sequence of successively visited configurations) follows the same law as the trajectory of a simple random walk on \(V_N\): the only way the energies enter in the dynamics is through the holding times. In other words, the RHT dynamics is a time-changed simple random walk on \(V_N\), and this simplifies the mathematical analysis a lot. The time change itself is usually called the clock process. To be coherent with our description of aging, one conjectures that this clock process converges to a stable subordinator.
Rigorous results on the RHT dynamics for the REM focused on proving the convergence of correlation functions to the arc-sine law, see [8, 9, 11, 30].
The aim of our paper is to prove that the aging picture holds for Glauber-type local dynamics in which the law of the jumps depends on the energies of both the current and the target configurations. We obtain the scaling limit of the corresponding clock process (a stable subordinator, as expected) and show that the scaling limit of the age process is the same as in [28].
When moving from RHT dynamics to Glauber dynamics, new difficulties appear. Some of these are due to the fact that the potential theory of a symmetric Markov chain on the hypercube is more involved in the generic case than in the simple symmetric case. Also, for the RHT dynamics, the distribution of the energy landscape around (but excluding), say, the first deep trap visited, is simply the Gibbs measure. This is no longer the case for a Glauber dynamics, and the crux is to understand this distribution.
Let us summarize our strategy of proof in a few words. We view the dynamics as a time-change of another process, which is the original dynamics accelerated by a factor \(\tau _{N,x}\) at every site \(x \in V_N\). Although this new process is no longer a simple random walk as for the RHT dynamics, it retains some of its properties, as for instance the reversibility with respect to the uniform measure over \(V_N\). We prove a law of large numbers for the number of sites discovered by this walk: on appropriate time scales, we show that this number grows asymptotically linearly with time, with a proportionality constant depending neither on the trajectory, nor on the environment.
Our second major step is to identify how the environment around a deep trap found by the dynamics is distributed. Roughly speaking, we follow an idea of [37] and show that asymptotically, this distribution has an explicit density with respect to the product measure, given in terms of a Green function. Care must be taken in order to make this statement precise, since the underlying graph changes with \(N\). As opposed to [37], we must thus resort to a finitary version of asymptotic convergence, in the sense that we only write statements for finite \(N\) with explicit bounds, instead of statements about limit distributions.
In order to complete the description of the distribution of the environment around a deep trap, we must then study the asymptotic behaviour of the Green function under the product measure. This can be performed via explicit computations, provided that the dynamics we study is not too predictable.
From the law of large numbers and the knowledge of the Green function at a deep trap, we finally derive the joint asymptotic behaviour of all the relevant quantities associated with the visits of traps: the time until the \(n\)th deep trap is met, the depth of the trap, and the total time spent on it. The aging picture then follows by noticing that the time spent outside of deep traps is negligible.
Many aspects of this approach do not depend too much on the choice of the dynamics, but the step of the proof we just described concerning the precise description of the Green function seems to be very model-dependent. Here we choose to deal with a Bouchaud dynamics with so-called asymmetry parameter \(a_N\) satisfying assumption (2.2) (see the related discussion there). Other choices are possible which would lead to new difficulties.
We close this introduction by a short discussion on related issues that are not directly addressed here.
There exists another spin-glass toy model on which a rather complete description of the dynamics can be easily obtained, namely the REM on the complete graph. In the RHT dynamics on the complete graph, when jumping, the dynamics chooses to go to any of the \(2^N\) possible configurations with equal probability. Aging results (in the sense of convergence of correlation functions to the arc-sine law) have been obtained in this case [10, 15–17]. For this model on the complete graph, it turns out to be also possible to consider a certain class of Glauber-type dynamics called the asymmetric dynamics, and prove aging for these with renewal techniques [31]. The definition of asymmetric dynamics is similar to the dynamics we consider here. Observe however that the asymmetric dynamics on the complete graph (and only on the complete graph!) has a very special structure: at a jump time, it chooses where to go according to a law that depends on the energies (which is reasonable) but that does not depend on the current configuration (which is less reasonable). As a consequence, the clock process is a subordinator already in finite volume, a fact that helps to prove it converges to a stable subordinator.
Limit theorems for the RHT dynamics on the \(p\)-spin model are discussed in [7, 18].
The aging picture we described above has also been investigated on models that are not related to spin glass theory. These are often referred to in the literature as trap models: one starts with a discrete-time Markov chain on a graph, which may be finite or infinite, and considers its time change through an additive functional as in the RHT dynamics, looking for scaling limits of either the correlation functions or the age process. We refer to [11] for results in this direction. In particular, models of non-simple random walks on \(\mathbb {Z}^d\) (\(d \geqslant 2\)) with aging properties have been considered in [4, 20, 32, 37].
So far we only discussed dynamics in the transient regime, long before thermalization occurs. For dynamical spin glasses, as for more general trap models on a growing sequence of graphs, it makes sense to look for scaling limits on the ergodic time scale as well. The first results in that direction dealt with the RHT dynamics on the complete graph [27]. The scaling limits that appear are known as \(K\)-processes. These interpolate between the transient regime, for small times, and the usual thermalization regime. It was later proved that \(K\)-processes describe the RHT dynamics in the REM [25] or the \(a\)-symmetric dynamics on the complete graph [13]. More recently, [34] showed how \(K\)-processes arise as a universal limit for trap models on growing sequences of graphs.
On the other side, it is also possible to derive aging results for dynamics of spin glass models on very short time scales. This is often referred to as extreme aging, see [12, 19].
The paper is organized as follows. In Sect. 2, we introduce basic definitions and notation. We then provide various quantitative estimates on the mixing properties of the accelerated walk in Sect. 3. Section 4 introduces the notion of the exploration process. The law of large numbers for the number of sites discovered by the walk is then achieved in Sect. 5. Next, Sect. 6 defines the relevant scales of our problem, in particular that of the deep traps. In Sect. 7, we obtain the joint scaling limit of the time it takes to find a deep trap, and of the depth of the deep trap. After recalling elementary properties of the process of the environment viewed by the particle in Sect. 8, we tackle the problem of controlling the distribution of the environment around the first deep trap met in Sect. 9. Section 10 is devoted to the asymptotic analysis of the Green function under the product measure. Combining the two previous sections, we obtain the scaling limit of the Green function at the first deep trap met in Sect. 11. We then derive the joint convergence of all the relevant quantities associated with the visits of the traps in Sect. 12. We conclude by deducing the convergence of the clock process and of the age process, in Sects. 13 and 14 respectively.
2 Definitions and notation
We view the hypercube \(V_N = \{0,1\}^N\) as an additive group, endowed with its usual graph structure (that is, nearest-neighbour for the Hamming distance, that we write \(| \cdot |\)). For \(x,y \in V_N\), we write \(x \sim y\) when \(|x-y| = 1\). Let \((E_x)_{x \in V_N}\) be i.i.d. random variables distributed as the positive part of a standard Gaussian. We write \(\mathbb {P}\) for their joint law (with associated expectation \(\mathbb {E}\)), and \({\mathcal {N}}_+(0,1)\) for the marginal distribution. Recall that \(\beta > 0\) is a fixed parameter, and that \(\tau _{N,x} = e^{\beta \sqrt{N} E_x}\). We write \(\tau _N = (\tau _{N,x})_{x \in V_N}\).
For \(a_N \in \mathbb {R}\), we consider the continuous-time Markov chain \((Z_t)_{t \geqslant 0}\) whose jump rate from \(x\) to a neighbour \(y\) is given by
and write \(\mathbf{P}^{\tau _N}_z\) for the law of this chain started from \(z \in V_N\) (and \(\mathbf{E}^{\tau _N}_z\) for the associated expectation). This dynamics, often called Bouchaud’s dynamics, is reversible for the measure with weights \((\tau _{N,x})_{x \in V_N}\). The case \(a_N = 0\) makes the jump rate in (2.1) independent of \(y\), and thus corresponds to the RHT dynamics. It is physically natural to consider dynamics that are more attracted to sites whose associated Gibbs weight is high, so we will only consider the case \(a_N \geqslant 0\). For technical reasons, it is convenient to assume from now on that actually \(a_N \geqslant 1\).
If \(a_N \gg \sqrt{\log N}\), then with probability tending to \(1\), the walk started from the origin will make its first jump to the neighbouring site of highest energy. This asymptotic complete predictability persists for subsequent jumps, in the sense that for every \(k\), one can predict with probability tending to \(1\) the sequence of the first \(k\) distinct sites that will be visited by the dynamics. In order to avoid such a behaviour, we will assume instead that
and moreover, that \(\overline{a} < 1/20\). It is instructive to figure out the large \(N\) behaviour of the extreme values of the jump rates from the origin to each of its neighbours in this case. Up to a multiplicative factor, this amounts to understanding the extreme values of the family \((e^{a_N E_y})_{y \sim 0}\). By the classical theory of extremes for independent Gaussian random variables, when the constraint (2.2) is replaced by an equality, the point process of the rescaled \((e^{a_N E_y})_{y \sim 0}\) converges to the point process with intensity measure \(x^{-(1+1/\overline{a})} \, \mathrm {d}x\). In particular, the jump rates remain inhomogeneous even in the large \(N\) limit.
Assumption (2.2) will be used in order to prove that when a walk moves at some fixed distance away from a deep trap, it will not manage to find it back rapidly. The parameter \(1/20\) is chosen for convenience. The proof below can be optimized to cover higher values of \(\overline{a}\), but a more detailed analysis would be required to cover every \(\overline{a} < 1\) for instance (see Sect. 10).
We rewrite \(Z\) as a time-change of a “nicer” random walk. Denote by \((X_t)_{t \geqslant 0}\) the walk speeded up by \(\tau _x\) at each \(x\), whose jump rates are thus
Note that these jump rates are symmetric: \(\omega _N(x,y) = \omega _N(y,x)\). In other words, the uniform measure is reversible for this process. We will denote the uniform probability measure over \(V_N\) by \(\mathfrak {u}_N\), and write \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}= \mathbf{P}^{\tau _N}_{\mathfrak {u}_N}\) for simplicity. We write \(\overline{\mathbb {P}}_N\) for \(\mathbb {P}\mathbf{P}^{\tau _N}_{\mathfrak {u}}\),
and \(\overline{\omega }_N=\sup _{x\in V_N}\omega _N(x)\).
We also define
General notation
If \(A\) is a set, \(|A|\) denotes its cardinality. We write \(\mathbb {N}= \{1,2,\ldots \}\), and \(a \wedge b = \min (a,b)\). For two sequences \(u_N, v_N\) with \(v_N \ne 0\) for all sufficiently large \(N\), we write \(u_N \sim v_N\) if \(u_N/v_N\) tends to \(1\) as \(N\) tends to infinity. In the spirit of the Bachmann–Landau “big O” notation, for \(\varepsilon > 0\), we write \(\pm \varepsilon \) to denote any real number whose value lies in \([-\varepsilon ,\varepsilon ]\) (this notation is convenient when we want to make explicit the fact that an estimate holds uniformly in other parameters).
3 Mixing properties
In this section, we provide various precise versions of the idea that the random walk \(X\) has good mixing properties. In the first proposition, we show that the walk started from any fixed point in \(V_N\) admits strong stationary times of order \(N\). In particular, the walk at the strong stationary time is distributed as \(\mathfrak {u}_N\) (the uniform measure on \(V_N\)), and the time scale of order \(N\) is very small compared to the exponentially growing time scales of our problem. The purpose of Propositions 3.4 and 3.5 is slightly different, since we will use them to show that the random walk has very little probability to be at a specific location after some time. Proposition 3.4 says that for almost every environment, it suffices to wait a time \(t \gg 1\) for the walk to “get lost”. Proposition 3.5 requires only \(t \gg 1/N\), but the result is averaged over the environment.
The definition of strong stationary times requires that the probability space contains additional independent randomness. Possibly enlarging the probability space, we assume that there exists infinitely many independent random variables distributed uniformly on \([0,1]\) and independent of everything else. Following [2], we say that a random variable \({\mathsf {T}} \geqslant 0\) is a randomized stopping time if the event \(\{\mathsf {T} \leqslant t\}\) depends only on \(\{X_s,\ s \leqslant t\}\) and on the value of these additional random variables.
Proposition 3.1
For any \(N\) large enough and any \(\tau _N\), there exists a randomized stopping time \({\mathsf {T}}_N\) with values in \(\{N, 2N, \ldots \}\) such that for any \(x \in V_N\), under \(\mathbf{P}^{\tau _N}_x\),
-
(1)
the distribution of \(X_{{\mathsf {T}}_N}\) is \(\mathfrak {u}_N\),
-
(2)
for any \(k \in \mathbb {N}\),
$$\begin{aligned} \mathbf{P}^{\tau _N}_x\left[{\mathsf {T}}_N \geqslant k N\right] = e^{-(k-1)}, \end{aligned}$$ -
(3)
\({\mathsf {T}}_N\) and \(X_{{\mathsf {T}}_N}\) are independent random variables.
Remark 3.2
A (randomized) stopping time satisfying (1) is called a stationary time; one satisfying (1) and (3) a strong stationary time.
Remark 3.3
Note in particular that the distribution of \(({\mathsf {T}}_N, X_{{\mathsf {T}}_N})\) does not depend on the environment \(\tau _N\).
Proof
Since the jump rates of \(X\) are symmetric, the measure \(\mathfrak {u}_N\) is reversible for this walk. Let
where \(x\sim y\) means that \(x\) and \(y\) are neighbours in \(V_N\). For a given realization of the \(E_x\)’s, the bilinear form \({\mathcal {E}}_N\) acting on \(L_2(V_N,\mathfrak {u}_N)\) is the Dirichlet form of the process \(X\). We similarly introduce
which defines the Dirichlet form of the Markov chain jumping along the edges of \(V_N\) with constant rate \(1\). We denote this Markov chain by \((X_t^\circ )_{t \geqslant 0}\).
Observe that
Let
be the spectral gap associated with the Dirichlet form \({\mathcal {E}}_N\) on \(L_2(V_N,\mathfrak {u}_N)\), and let
be the spectral gap of \({\mathcal {E}}^\circ _N\).
On the one hand, inequality (3.1) implies that \({\mathsf {SG}}_N^\circ \leqslant {\mathsf {SG}}_N\). On the other hand, \(X^\circ \) is simply the \(N\)-fold product of the simple random walk on \(V_1\), so by [33, Theorem 2.5], we have \({\mathsf {SG}}^\circ _N = {\mathsf {SG}}^\circ _1\), and in fact \({\mathsf {SG}}^\circ _1 = 2\). Therefore
Let us now introduce
where \(\Vert \cdot \Vert _{\mathsf {TV}}\) denotes the total variation distance, and
We learn from [1, Lemmas 5, 7 and 23 of Chapter 4] that, respectively,
where \(\mathfrak {u}_N^* = \min _x \mathfrak {u}_N(x)\), which in our case is \(2^{-N}\). Using (3.2), we thus derive from these three observations that, for \(N\) large enough and any \(\tau _N\),
that is, for any \(x,y \in V_N\) and any \(\tau _N\),
Following [3, Proposition (3.2)], we can thus define the strong stationary time \({\mathsf {T}}_N\), with values in \(\{N,2N,\ldots \}\). Let \((U_1, U_2, \ldots )\) be independent random variables distributed uniformly on \([0,1]\), independent of everything else. Conditionally on \(X_N= y\), we let \({\mathsf {T}}_N = N\) if
Otherwise, we define \(\mathsf {T}_N\) inductively: for every \(k \in \mathbb {N}\), conditionally on \(\mathsf {T}_N > kN\), on \(X_{kN} = z\) and on \(X_{(k+1)N} = y\), we let \(\mathsf {T}_N = (k+1)N\) if
By construction, we have
so that
Similarly, we have
By induction on \(k\), we obtain that for any \(k \in \mathbb {N}\) and any \(x,y \in V_N\),
and this finishes the proof. \(\square \)
Proposition 3.4
For any environment \(\tau _N\), any \(x,y \in V_N\) and any \(t \geqslant 0\),
Proof
The proposition is a direct consequence of [39, Corollary 2.1.5] and the spectral gap estimate (3.2). \(\square \)
Proposition 3.5
For any \(x \in V_N\) and any \(t \geqslant 0\),
Proof
We recall that \((X_t^\circ )_{t \geqslant 0}\) is the Markov chain whose jump rate from one point to any of its neighbours is equal to \(1\). Since these jump rates are dominated by the jump rates of \((X_t)\), and using [26, Lemma 2.2], we get that
Each coordinate of \((X_t^\circ )\) is a Markov chain on two states with jump rate equal to \(1\) from one state to the other, and that evolves independently of the other coordinates. This leads to the announced result. \(\square \)
Remark 3.6
As is well-known, such on-diagonal upper bound on the heat kernel automatically transfers to off-diagonal decay. In other words, for any \(x,y \in V_N\),
To see this, note that
which by the Cauchy–Schwarz inequality, is smaller than
By reversibility, the first sum above is equal to
and we arrive at
After integrating with respect to \(\mathbb {P}\), we use the Cauchy–Schwarz inequality again and Proposition 3.5 to obtain (3.3).
4 The exploration process
In this section, we define an enumeration of \(V_N\) along the trajectory of the random walk \(X\), as follows. Since the walk eventually visits every site of \(V_N\), we can let \(Y_1,\ldots , Y_{{2^N}}\) be the sequence of successive sites visited by \(X\). Consider the sequence of sites
where \((\cdot +z)_{|z| \leqslant 1}\) is enumerated in some arbitrary order. Clearly, \({\mathcal {S}}\) spans \(V_N\). We let
be the sequence \({\mathcal {S}}\) with repetitions removed (in other words, for every \(x \in V_N\), the first occurrence of \(x\) in \({\mathcal {S}}\) is kept, and all subsequent occurrences are deleted). The sequence \({\mathcal {S}}'\) spans \(V_N\) by construction, and we refer to this sequence as the exploration process.
Let \(x \in V_N\). The distribution of the trajectory up to the time when the walk is within distance \(1\) from \(x\) is independent of \(E_x\). From this observation, we can derive the following result (see [37, Proposition 4.1] for a complete proof).
Proposition 4.1
For any \(z \in V_N\), under the measure \(\mathbb {P}\mathbf{P}^{\tau _N}_z\), the random variables \((E_{x_n})_{1 \leqslant n \leqslant 2^N}\) are independent and follow a \({\mathcal {N}}_+(0,1)\) distribution. This holds also under the measure \(\overline{\mathbb {P}}_N\).
5 Law of large numbers for the number of discovered sites
For every \(t \geqslant 0\), let
and \(D_N(t) = |{\mathcal {D}}_N(t)|,\,R_N(t)=|{\mathcal {R}}_N(t)|\). We say that a site \(x \in V_N\) is discovered before time \(t\) if it belongs to \({\mathcal {D}}_N(t)\), and that it is visited before time \(t\) if it belongs to \({\mathcal {R}}_N(t)\).
More generally, for a subset \(A\subseteq V_N\), we define
and \(D^A_N(t) = |{\mathcal {D}}^A_N(t)|\). Observe that \({\mathcal {D}}^A_N(t) = {\mathcal {D}}_N(t)\) when \(A=\{z \text { s.t. } \vert z\vert \leqslant 1\}\) and \({\mathcal {D}}^A_N(t) = {\mathcal {R}}_N(t)\) when \(A\) is reduced to the identity element in \(V_N\).
The purpose of this section is to prove the following law of large numbers for the number of discovered sites.
Theorem 5.1
Let \(\overline{c} \in (0,\log (2))\) and \(({\mathsf {t}}_N)\) be such that \(\log ({\mathsf {t}}_N) \sim \overline{c} N\). Define \({\mathsf {d}}_N=\overline{\mathbb {E}}_N[D_N({\mathsf {t}}_N)]\). For any \(t>0\), we have
Moreover, \(\log ({\mathsf {d}}_N) \sim \overline{c} N\).
Before starting the proof of this result, we collect some useful bounds on \(\overline{\omega }_N = \sup _{x \in V_N} \omega _N(x)\), where we recall that \(\omega _N(x)\) was defined in (2.4).
Lemma 5.2
For every \(c > \sqrt{2 \log 2}\), we have
Moreover, for \(N\) sufficiently large,
Proof
In order to prove these results, we will rely on the fact that for \(z \geqslant 1\),
From this, it readily follows that, for every \(c > 0\),
which tends to \(0\) as soon as \(c^2 > 2 \log (2)\). On the event
we have
so (5.2) is proved.
For the second part, note that
On the event
we have
Since we assume that \(a_N \leqslant \overline{a} \sqrt{2 \log N}\), we obtain (5.3). \(\square \)
Proof of Theorem 5.1
In a first step of the proof, we shall derive estimates on the mean and variance of \(D_N(t)\) for a fixed realization of \(\tau _N\). These imply a law of large numbers as in (5.1) but with \({\mathsf {d}}_N\) replaced by \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]\) (and therefore depending on \(\tau _N\)). In a second step, we establish a concentration inequality which allows to replace \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]\) by \(\overline{\mathbb {E}}_N[D_N({\mathsf {t}}_N)]\). \(\square \)
We start by proving that \(\log ({\mathsf {d}}_N) \sim \overline{c} N\) and \(\log \mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]\sim \overline{c} N\) in \(\mathbb {P}\)-probability. It will be a consequence of the following
Lemma 5.3
We have
whenever \(t\geqslant 1\), and
whenever \(t2^{-N}\leqslant \frac{1}{2}\).
Proof
Let us fix a time \(t>0\) and consider an independent exponentially distributed random variable of mean \(t\), say \(T\). Let
be the Green function of the process \(X\).
Our estimates rely on the following formula:
Indeed it follows from the Markov property that
The reversibility of the process \(X\) implies that \(G_N^t(x,y)=G_N^t(y,x)\). Therefore, summing over \(x\), we get that
The equality \(\sum _x G_N^t(y,x)=t\) leads to (5.5).
We now explain how to derive an upper bound on \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[R_N(T)]\) from (5.5). We assume that \(t\geqslant 1\).
When starting from \(y\), the process \(X\) waits for an exponential time of parameter \(\omega _N(y)=\sum _{z\sim y} \exp (a_N(E_y+E_z))\) before leaving the point \(y\). Thus we have
Plugging this estimate in (5.5), we get that
One may now replace \(T\) by \(t\) observing that the function \(s\rightarrow \mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[R_N(s)\right]\) is increasing and therefore
Thus we have established that, for all \(t\geqslant 1\),
(we used the inequality \(\int _1^{+\infty } e^{-s}\ \mathrm {d}s=e^{-1}\geqslant 1/3\)).
We now turn to lower bounds. We assume that \(t2^{-N}\leqslant \frac{1}{2}\). Integrating the estimate from Proposition 3.4, we get that
Plugging this estimate in (5.5), we see that one has
One may now replace \(T\) by \(t\) observing that the function \(s\rightarrow \mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[R_N(s)\right]\) is subadditive and increasing; thus satisfies \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[R_N(st)\right]\leqslant (s+1)\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[R_N(t)\right]\) and
Thus we have established that, for all \(t\) such that \(t2^{-N}\leqslant 1/2\),
\(\square \)
Observe from (2.2) and (5.2) that \(\frac{1}{N} \log (1+\overline{\omega }_N) \) converges to \(0\) in \(\mathbb {P}\)-probability.
Thus, if we replace \(t\) by \({\mathsf {t}}_N\), we deduce from the bounds in Lemma 5.3 that \( \frac{1}{N} \log \mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[R_N({\mathsf {t}}_N)\right]\) converges to \(\overline{c}\) in \(\mathbb {P}\)-probability. Since \((N+1)\cdot R_N(t)\geqslant D_N(t)\geqslant R_N(t)\), we also conclude that \( \frac{1}{N} \log \mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]\) converges to \(\overline{c}\) in \(\mathbb {P}\)-probability.
Taking the expectation in the bounds in Lemma 5.3 and using (2.5), we also get that
Now we wish to show that \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]^{-1} D_N({\mathsf {t}}_N)\) converges to \(1\) in probability. This will follow from the fact that the variance of \(D_N({\mathsf {t}}_N)\) under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) is negligible compared to the square of its mean.
Lemma 5.4
For all \(p>0\), there exists a numerical constant \(c(p)\) such that
Proof
Let \(k\geqslant 1\). On the event \(R_N(t)\geqslant k+1\), the process has visited and left some point at least \(k\) times. Note that, when at \(x\), the process \(X\) waits for an exponential time of parameter \(\omega _N(x)\) before leaving the point \(x\). Thus we see that the probability \(\mathbf{P}^{\tau _N}_x\left[R_N(t)\geqslant k+1\right]\) is bounded by the probability that a sum of \(k\) i.i.d. exponential\((1)\) random variables is bounded by \(\overline{\omega }_N t\). The lemma follows from this observation. \(\square \)
We deduce from Lemma 5.4 with \(p=2\) and (5.2) that, for some constant \(c_2\),
with high \(\mathbb {P}\)-probability. Since \(D_N(t) \leqslant (N+1) R_N(t)\), a similar bound holds for \(D_N(t)\) with an extra factor \((N+1)^2\) multiplying \(c(2)\).
We also need some control on the self-intersections of the trajectories of the process \(X\). In the next lemma, we consider two Markov chains \(X\) and \(X'\) that start from the same point in \(V_N\) (which will be chosen according to the uniform probability in \(V_N\)), and then evolve independently in the same fixed environment \(\tau _N\) with jump rate from \(x\) to \(y\) given by \(\omega _N(x,y)\) (defined in (2.3)). We define
and \({\mathcal {D}}^A_N(s,t) = {\mathcal {D}}^A_N(t) \cap {\mathcal {D}}'^A_N(s),\,D^A_N(s,t) = |{\mathcal {D}}^A_N(s,t)|\).
Lemma 5.5
Let \(s\geqslant t\geqslant 1\). We have
Proof
The strategy is similar to the one we already used in the proof of the upper bound in Lemma 5.3.
Let us fix two times \(s\geqslant t\geqslant 1\) and consider independent exponentially distributed random variables \(S\) and \(T\) of mean respectively \(s\) and \(t\).
We recall that
is the Green function of the process \(X\). Let us also define
We claim that
To prove Eq. (5.8), we start by noting that
(This follows from the independence of \(X\) and \(X'\) and the Markov property: \(\frac{G_N^t(x,z+y)}{G_N^t(z+y,z+y)}\) is the probability that \(X\) hits \(z+y\) before time \(T\). We then use a union bound to control the probability of hitting the set \(z+A\) by a sum over \(y\in A\).)
From the reversibility, we get that \(G_N^t(x,z+y)=G_N^t(z+y,x)\). Observe that
and, since \(s\geqslant t\),
Summing inequality (5.9) over \(x\), we thus obtain (5.8).
As in the proof of Lemma 5.3, we plug in (5.8) crude estimates on the Green functions. We first recall that, for all \(z\) and \(z'\),
and it follows from Proposition 3.4 that
Thus we derive that
where we used the bound \(\omega _N(x)\le \overline{\omega }_N\) in the last inequality.
On the other hand, the function \((s,t)\rightarrow \mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(s,t)]\) being increasing in both arguments, we have \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(s,t)]\leqslant e^2\ \mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(S,T)]\).
Putting everything together, we conclude the proof of Lemma 5.5. \(\square \)
As a consequence of Lemma 5.5, we obtain the linearity of the expected number of discovered sites.
Lemma 5.6
For all \(t\) and \(s\) we have
and
Proof
The first inequality is obvious.
We define
to be the set of discovered sites between times \(t\) and \(t+s\), and \(D^A_N(t\rightarrow s) = |{\mathcal {D}}^A_N(t\rightarrow s)|\).
Note that, by stationarity, we have \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D^A_N(t\rightarrow s)\right]=\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D^A_N(s)\right]\). Therefore
This last quantity is bounded by the number of sites that are discovered both before time \(t\) and also between times \(t\) and \(t+s\) i.e. \(|{\mathcal {D}}^A_N(t)\cap {\mathcal {D}}^A_N(t\rightarrow s)|\).
The Markov property (applied at time \(t\)) and the reversibility of \(\mathfrak {u}_N\) imply that \(|{\mathcal {D}}^A_N(t)\cap {\mathcal {D}}^A_N(t\rightarrow s)|\) has the same law as \(D^A_N(s,t)\), thus leading to the inequality in Lemma 5.6. \(\square \)
We deduce from Lemma 5.6 that, for all \(t>0\),
The upper bound in (5.11) follows from the integrated version of the first inequality in Lemma 5.6. To get the lower bound, we also take the expectation in the second inequality in Lemma 5.6 and iterate it approximately \(t\) times. The cumulated error term is estimated with Lemma 5.5: neglecting subexponential terms, it is of order \(2^{-N}{\mathsf {t}}_N^2\). On the other hand, we already know from (5.6) that \(\overline{\mathbb {E}}_N\left[D_N({\mathsf {t}}_N)\right]\) is of the same order as \({\mathsf {t}}_N\) on the exponential scale. Since \(2^{-N}{\mathsf {t}}_N^2\) is negligible compared to \({\mathsf {t}}_N\), we can conclude the proof.
Our next task is to get estimates on the variance of \(D^A_N(t)\). We shall use the notation \(\mathbf{V}^{\tau _N}_{\mathfrak {u}}(X)\) to denote the variance of a random variable \(X\) under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) and \(\mathbf{C}^{\tau _N}_{\mathfrak {u}}(X,Y)\) to denote the covariance of the random variables \(X\) and \(Y\) under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\).
We obviously have
If we write that \(D^A_N(t+s)=D^A_N(t)+(D^A_N(t+s)-D^A_N(t))\), we are lead to estimating \(\mathbf{C}^{\tau _N}_{\mathfrak {u}}\left[D^A_N(t),D^A_N(t+s)-D^A_N(t)\right]\).
Observe that \(D^A_N(t+s)-D^A_N(t)\leqslant D^A_N(t\rightarrow s)\) (where \(D^A_N(t\rightarrow s)\) was defined in (5.10)), and therefore
On the other hand we read from Lemma 5.6 that \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(t+s)-D^A_N(t)]\geqslant \mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(s)]-\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^A_N(s,t)]\).
Thus we have
Lemma 5.7
For all \(t\) and \(s\geqslant N^2\) we have
Proof
We start observing that
(The first inequality is a consequence of Proposition 3.4, and the second one is obvious.)
Therefore
One now concludes the proof using the estimates from Lemma 5.4. \(\square \)
Lemma 5.8
Proof
We split the time interval \([0,{\mathsf {t}}_N]\) into \(K\) pieces of equal size. Here \(K\) has to be chosen properly: \(K=e^{N^{3/4}}\) would do.
We then estimate the variance of \(D^A_N({\mathsf {t}}_N)\) using (5.12) and (5.13):
Using our estimates from Lemmas 5.4, 5.7, 5.3 and 5.5, we get that
where
We iterate this computation \(K\) times. This way we get that \(\mathbf{V}^{\tau _N}_{\mathfrak {u}}\left[D^A_N({\mathsf {t}}_N)\right]\leqslant K\eta _N\).
It then follows from our assumptions on \(K\) that \(\eta _N\) is negligible compared with \({\mathsf {t}}_N^2/K\). (Observe in particular that, by Lemma 5.2, we have \((\overline{\omega }_N\frac{1}{K}{\mathsf {t}}_N)^2 \ll {\mathsf {t}}_N^2/K\).)
Thus we deduce that \(K\eta _N/{{\mathsf {t}}_N^2}\) tends to \(0\) in \(\overline{\mathbb {P}}_N\text {-probability}\).
We have checked that the variance of \(D^A_N({\mathsf {t}}_N)\) is negligible compared with the square of its mean with high \(\overline{\mathbb {P}}_N\text {-probability}\), and this ends the proof. (Recall that \({\mathsf {t}}_N\) is a lower bound for the mean of \(D^A_N({\mathsf {t}}_N)\), see Lemma 5.3.) \(\square \)
End of the proof of Theorem 5.1
In order to deduce Theorem 5.1 from Lemma 5.8, it is sufficient to show that the variance of \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]\) under \(\mathbb {P}\) (i.e. as a function of the random variable \(\tau _N\)) is negligible compared to \({\mathsf {t}}_N^2\).
We use the notation \(\mathbb {V}[\mathsf {X}]\) to denote the variance of a random variable \(\mathsf {X}\) with respect to \(\mathbb {P}\).
To estimate \(\mathbb {V}[\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]]\), we first rely on the Efron–Stein inequality (see [40]), namely
where \(\tau _N^{(x)}\) denotes an environment of traps in which we have replaced \(E_x\) by an independent copy and left all other values unchanged.
By symmetry, all the terms in this last sum have the same value, so that
Thus we see that we will be done if we prove that
is negligible with respect to \({\mathsf {t}}_N^2\). We will in fact show that, on the exponential scale, \(\mathbb {E}[(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right]-\mathbf{E}^{\tau _N^{(0)}}_{\mathfrak {u}}\left[D_N({\mathsf {t}}_N)\right])^2]\) is at most of order \((2^{-N}{\mathsf {t}}_N)^2\):
In order to show (5.14), we split the trajectory of the walk in different time intervals: let \(T_1\) be the first time the point \(0\) is discovered by the process \(X\); let \({\mathsf T}^{(1)}\) be the first strong stationary time after \(T_1\): in other words, \({\mathsf T}^{(1)}-T_1\) is a strong stationary time for the Markov chain \((X_{t+T_1})_{t\geqslant 0}\) obtained as in Proposition 3.1.
We iterate these definitions: let \(T_2\) be the first time \(0\) is discovered after time \({\mathsf T}^{(1)}\); let \({\mathsf T}^{(2)}\) be the first strong stationary time after time \(T_2\), ...We define \(I_j=[T_j,{\mathsf T}^{(j)})\) and \(I=\cup _j I_j\).
Let us also define
and \(D^{(0)}_N(t) = |{\mathcal {D}}^{(0)}_N(t)|\).
From the construction of strong stationary times, see Proposition 3.1 and its proof, it follows that the law of \({\mathcal {D}}^{(0)}_N(t)\) (and therefore the law of \(D^{(0)}_N(t)\) and its mean) is the same under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) or \(\mathbf{P}^{\tau _N^{(0)}}_{\mathfrak {u}}\). Therefore \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D^{(0)}_N({\mathsf {t}}_N)]=\mathbf{E}^{\tau _N^{(0)}}_{\mathfrak {u}}[D^{(0)}_N({\mathsf {t}}_N)]\) and
Note that these two terms have the same law under \(\mathbb {P}\). Thus we will have established (5.14) once we show that \(\mathbb {E}[\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D_N({\mathsf {t}}_N)-D^{(0)}_N({\mathsf {t}}_N)]^2]\) is at most of order \((2^{-N}{\mathsf {t}}_N)^2\) (up to subexponential factors).
Applying the Markov property and using the stationarity of the uniform probability, we have
where \({\mathsf T}_N\) is the strong stationary time considered in Proposition 3.1. It is easy to deduce from Proposition 3.1 and Lemma 5.4 that \(\max _{x\sim 0} \mathbf{E}^{\tau _N}_x [D_N({\mathsf T}_N)]\) is of subexponential order.
To bound the second term, observe that, when the process \(X\) hits a point \(x\), it then spends at \(x\) an exponentially distributed time of parameter \(\omega _N(x)\leqslant \overline{\omega }_N\). Therefore, for all \(\lambda >0\),
But, due to the stationarity of the uniform probability, we also have
Therefore
Choosing \(\lambda =1/t\) yields
Observe that \(T_k\) stochastically dominates a sum of \(k\) independent copies of \(T_1\). Therefore
which leads to
whenever \(\frac{N+1}{2^N} (1+\overline{\omega }_Nt)\leqslant \frac{1}{2}\).
On the other hand, \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D_N({\mathsf {t}}_N)-D^{(0)}_N({\mathsf {t}}_N)]\leqslant 2^N\). Thus we have
We now have to square this last inequality and take the expectation with respect to \(\mathbb {P}\). The \(\mathbb {P}\) probability of the event \(\frac{N+1}{2^N} (1+\overline{\omega }_N {\mathsf {t}}_N)\geqslant \frac{1}{2}\) decays to \(0\) at a super-exponential rate (see (5.3)), which compensates for the \(2^N\) factor in front. We already noticed that \(\max _{x\sim 0} {\mathbf{E}^{\tau _N}_x[D_N({\mathsf T}_N)]}\) is of subexponential order and the term \((1\,+\,\overline{\omega }_N {\mathsf {t}}_N)\) is of the same order as \({\mathsf {t}}_N\) (still on the exponential scale). Thus we conclude that \(\mathbb {E}[\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D_N({\mathsf {t}}_N)-D^{(0)}_N({\mathsf {t}}_N)]^2]\) is at most of order \((2^{-N}{\mathsf {t}}_N)^2\) and therefore (5.14) holds. \(\square \)
6 Defining the scales
We fix \(\overline{c} \in (0,\log (2))\), and \({\mathsf {t}}_N\) such that \(\log ({\mathsf {t}}_N) \sim \overline{c} N\). The sequence \({\mathsf {t}}_N\) is our time scale for the accelerated dynamics \(X\). Theorem 5.1 associates with it a second scale \(({\mathsf {d}}_N)\) measuring the number of sites discovered, and we recall that \(\log ({\mathsf {d}}_N) \sim \overline{c} N\).
It is a classical fact that
Letting
we obtain
Since by Proposition 4.1, the random variables \((E_{x_i})_{1 \leqslant i \leqslant {\mathsf {d}}_N}\) are i.i.d. under \(\overline{\mathbb {P}}_N\), their maxima will be of the order of \(b_N\). In order to measure the maximal values of \((\tau _{N,x_i})_{1 \leqslant i \leqslant {\mathsf {d}}_N}\), we thus define
Using the fact that \(\log ({\mathsf {d}}_N) \sim \overline{c} N\), one can show that for any \(z > 0\),
where
As a way to make it clear that the scale chosen is the correct one, we note for instance that
From now on, we assume that \( \alpha < 1, \) which corresponds to a low temperature assumption, and is the regime of aging. Actually, this assumption is only required at the very end of our proof, in order to show that the time spent out of deep traps is negligible. Specifically, we will need the following result.
Proposition 6.1
We assume \(\alpha < 1\). We have
Proof
This can be justified using the general results of [5], but we prefer to provide the reader with a more direct proof. We can write \( \mathbb {E}\left[ \tau _{N,x} \ \mathbf{1}_{\tau _{N,x} \leqslant \delta B_N} \right] \) as
The condition on the indicator function can be rewritten as
We let
Note that \(u_N\) tends to \(\sqrt{2 \overline{c}}\) as \(N\) tends to infinity. In (6.7), the term \(1/2\) can be neglected since the integral is bounded away from \(0\) (or because one can easily check that \(B_N/{\mathsf {d}}_N\) grows exponentially fast with \(N\)). By a change of variables, we can rewrite this integral as
We write \(f(u) = \beta u - u^2/2\). Since \(f\) is concave, we have
and thus the quantity appearing in (6.8) is bounded by
The term \(f'(u_N)\) converges to \(f'(\sqrt{2\overline{c}}) > 0\). We need to compute \(f(u_N)\):
Recalling that \(B_N = e^{\beta \sqrt{N} b_N}\), we obtain the right-hand side of (6.9) is bounded, up to a constant, by
The result is thus proved provided
but this is a consequence of (6.1) and (6.2). \(\square \)
Remark 6.2
A slightly more careful analysis reveals that actually,
7 The deep traps
Let \(\delta > 0\). We say that \(x \in V_N\) is a deep trap if \(\tau _{N,x} \geqslant \delta B_N\). Let \((r^{(\delta )}_N(n))_{n \in \mathbb {N}}\) be the increasing sequence spanning the set
with the convention that \(r^{(\delta )}_N(n) = +\infty \) if \(n\) exceeds the cardinality of this set. In other words, in the numbering defined by the exploration process, \(r^{(\delta )}_N(1)\) is the index of the first deep trap discovered by the random walk, \(r^{(\delta )}_N(2)\) is the index of the second one, and so on. When \(r^{(\delta )}_N(n)\) is finite, we let \(x^{(\delta )}_N(n)\) be the position of the \(n\mathrm{th}\) trap discovered, that is, \(x^{(\delta )}_N(n) = x_{r^{(\delta )}_N(n)}\). We set \(x^{(\delta )}_N(n) = \infty \) if \(r^{(\delta )}_N(n) = +\infty \). The quantity \(\tau _{N,x^{(\delta )}_N(n)}\) denotes the depth of the \(n\mathrm{th}\) trap discovered, with the understanding that \(\tau _{N,x^{(\delta )}_N(n)} = \infty \) when \(x^{(\delta )}_N(n) = \infty \). Let \(T^{(\delta )}_N(n)\) be the instant when the \(n\mathrm{th}\) trap is discovered:
Proposition 7.1
Under \(\overline{\mathbb {P}}_N\), the random variables
jointly converge in distribution as \(N\) tends to infinity (in the sense of finite-dimensional distributions). Let us write
for limiting random variables, which we assume to be defined on the \(\mathbb {P}\)-probability space for convenience. Their joint distribution is described as follows. The family \(\left(T^{(\delta )}(n)\right)_{n \in \mathbb {N}}\) is distributed as the jump instants of a Poisson process of intensity \(\delta ^{-\alpha }\), while \(\left(\tau ^{(\delta )}(n)\right)_{n \in \mathbb {N}}\) are i.i.d. with common distribution
Moreover, the two families \((T^{(\delta )}(n))_{n \in \mathbb {N}}\) and \((\tau ^{(\delta )}(n))_{n \in \mathbb {N}}\) are independent.
Proof
Recall that as given by Proposition 4.1, the random variables \((\tau _{N,x_j})_{1 \leqslant j \leqslant 2^N}\) are i.i.d. Let us write
Recall from (6.4) that
Let \(F_n\) be the \(\sigma \)-algebra generated by \((\tau _{N,x_j})_{1 \leqslant j \leqslant r^{(\delta )}_N(n)}\). For \(k \in \mathbb {N}\) and \(z \geqslant \delta \), we compute
The event \(r^{(\delta )}_N(n) = k + r^{(\delta )}_N(n-1)\) corresponds to the fact that for every \(j\) such that \(r^{(\delta )}_N(n-1) < j < k+r^{(\delta )}_N(n-1)\), the site \(x_j\) is not a deep trap, while \(x_{k+r^{(\delta )}_N(n-1)}\) is a deep trap. This event is independent of \(F_{n-1}\) and has probability
Conditionally on \(r^{(\delta )}_N(n) = k + r^{(\delta )}_N(n-1)\) and on \(F_{n-1}\), the probability of the event
is nothing but
Hence, we obtain
Summing over \(k\) leads to
In view of (7.2) and of the fact that (see (6.4))
we obtain that under \(\overline{\mathbb {P}}_N\), the random variables
jointly converge in distribution to the random variables
described in the proposition. To conclude, we need to replace the random variables \({\mathsf {d}}_N^{-1} \ r^{(\delta )}_N(n)\) by \({\mathsf {t}}_N^{-1} T^{(\delta )}_N(n)\) in the convergence of (7.3) to (7.4).
The process \(t \mapsto D_N(t)\) is increasing, right-continuous, and the size of its jumps cannot be larger than \((N+1)\) since the walk discovers at most \((N+1)\) sites at once. Hence,
Since \({\mathsf {d}}_N\) grows exponentially in \(N\), it follows that in the convergence of (7.3) to (7.4), we can replace \({\mathsf {d}}_N^{-1} \ r^{(\delta )}_N(n)\) by \({\mathsf {d}}_N^{-1} \ D_N(T^{(\delta )}_N(n))\).
Let us see that for any \(y > 0\),
For any \(\varepsilon > 0\), we have
The first term tends to \(e^{-y\delta ^{-\alpha }}\) by the preceding observation, while the second tends to \(0\) by Theorem 5.1. Hence,
Similarly, one can show that
Inequalities (7.6) and (7.7) being valid for any \(\varepsilon > 0\), they justify (7.5). The generalization of this identity to the convergence of finite-dimensional distributions is then only a matter of notation (the key point besides Theorem 5.1, as seen in the proof of (7.5), being the monotonicity of \(t \mapsto D_N(t)\)). \(\square \)
8 The environment viewed by the particle
Let \(\Omega _N = [1,+\infty )^{V_N}\), so that \(\tau _N\) is a random variable taking values in \(\Omega _N\). The group \(V_N\) naturally acts on \(\Omega _N\) by translations: for \(\tau \in \Omega _N\) and \(x \in V_N\), we let \(\theta _x \ \tau \in \Omega _N\) be such that
We define the environment viewed by the particle as the process \((\tau _N(t))_{t \geqslant 0}\), taking values in \(\Omega _N\) and such that
Proposition 8.1
For any \(\tau _N \in \Omega _N\), the process \((\tau _N(t))_{t \geqslant 0}\) is a reversible Markov chain under the measure \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\).
Proof
The Markov property of this process is inherited from \((X_t)\). A minor adaptation of the proof of [21, Lemma 4.3 (iv)] shows reversibility. \(\square \)
9 The environment around the first trap discovered
Let us write \(\tau ^{(\delta )}_N(n)\) for the environment seen at \(x^{(\delta )}_N(n)\), that is,
The main purpose of this section is to understand the distribution of \(\tau ^{(\delta )}_N(1)\). Let
The value of \(\tau ^{(\delta )}_N(1)\) at \(0\) is described by Proposition 7.1. Roughly speaking, we will show that as \(N\) tends to infinity, the distribution of \(((\tau ^{(\delta )}_N(1))_z)_{z \ne 0}\) approaches the distribution proportional to
More precisely, we will prove the following result.
Theorem 9.1
Let \({\mathcal {H}}^{(\delta )}_N(n)\) be the event that \(x^{(\delta )}_N(n)\) is visited by \(X\) during the time interval \([T^{(\delta )}_N(n),T^{(\delta )}_N(n) + N]\). For \(M > 0\), let \({\mathcal {E}}^M_N\) be the event defined by
For any \(\varepsilon > 0\), any \(M\) large enough and any \(N\) large enough, if \(h_N : \Omega _N \rightarrow \mathbb {R}\) is such that \(\Vert h_N\Vert _\infty \leqslant 1\), then
where we recall that the notation \(\pm \) was introduced at the end of Sect. 1.
Remark 9.2
In (9.3) and below, for any \(\tau \in \Omega _N\), we write \((h_N G_N)(\tau )\) as shorthand for \(h_N(\tau ) \, G_N(\tau )\).
Remark 9.3
A notable feature of Theorem 9.1 is that as soon as \(N\) is large enough, identity (9.3) holds for any function \(h_N\) satisfying \(\Vert h_N\Vert _\infty \leqslant 1\). In particular, it is not required that, for instance, \(h_N\) depend only on a finite number of \((\tau _{N,x})_{x \in V_N}\), or have any other form of spatial mixing property. This will make the proof of Theorem 9.1 slightly more involved than otherwise, but will also greatly simplify our subsequent reasoning.
We now describe our starting point for the proof of Theorem 9.1. Let us introduce \((\tilde{e}_N(n))_{n \in \mathbb {N}}\) i.i.d. exponential random variables of parameter \(1/N^2\) (under \(\mathbf{P}^{\tau _N}_x\) for any \(x\)), independent of everything else, and write \((\Theta _t)_{t \geqslant 0}\) to denote the time translations on the space of trajectories, that is, \((\Theta _t X)_s = X_{t+s}\). For any \(M \geqslant 1\), we define \(T^M_N(1) = T^{(\delta )}_N(1) \wedge (M{\mathsf {t}}_N)\) (although this depends also on \(\delta \), we keep it implicit in the notation, since \(\delta \) will be kept fixed as long as we consider this quantity). Letting \(\tilde{T}^M_N(0) = 0\), we then define recursively, for any \(n \in \mathbb {N}\),
In words, to reach time \(\tilde{T}^M_N(n)\), we wait until the end of a strong stationary time (as given by Proposition 3.1) started at time \(T^M_N(n) + \tilde{e}_N(n)\).
Proposition 9.4
Under the measure \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\), the successive pieces of trajectory
form a stationary sequence, while
are independent and identically distributed.
Proof
The first part is a consequence of the Markov property and the fact that \(X_{\tilde{T}^M_N(n)}\) is distributed uniformly over \(V_N\), itself a consequence of the fact that \({\mathsf {T}}_N\) is a stationary time.
For the second part, we only need to check the independence. Let \(F\) be a bounded function on the space of trajectories which is measurable with respect to the \(\sigma \)-algebra generated by \((X_s)_{s < T^M_N(n)+\tilde{e}_N(n)}\), and \(G\) be a bounded measurable function on the space of trajectories. We want to show that
The left-hand side above is equal to
and this leads to (9.6) by the Markov property and the fact that \({\mathsf {T}}_N\) is a stationary time. \(\square \)
Let \(h_N : \Omega _N \rightarrow \mathbb {R}\) be a function. By the ergodic theorem, for any environment \(\tau _N \in \Omega _N\),
On the other hand, the stationarity of the environment viewed by the particle (that is, Proposition 8.1) guarantees that
Combining this with (9.7), we thus obtain that, for any environment,
which we can rewrite as
where the conditional expectation is understood to be \(1\) in case the event \(\tau _{N,X_0}\geqslant \delta B_N\) has \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\)-probability equal to \(0\) (that is, when \(\tau _N\) contains no deep trap). The rest of this section is devoted to passing from (9.8) to Theorem 9.1. We want to take \(\mathbb {E}\)-expectations on both sides of this identity. Considering first the right-hand side, we prove that \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}[\tau _{N,X_0}\geqslant \delta B_N]\) and \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[\tilde{T}^M_N(1)]\) are concentrated respectively around \(\delta ^{-\alpha } {\mathsf {d}}_N^{-1}\) and \(\delta ^\alpha {\mathsf {t}}_N\).
Lemma 9.5
There exists \(c > 0\) such that for any \(\varepsilon > 0\) and any \(N\) large enough,
Proof
We can rewrite the probability \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}[\tau _{N,X_0}\geqslant \delta B_N]\) as
where \(\mathfrak {B}_{N,x} = \mathbf{1}_{\tau _{N,x}\geqslant \delta B_N}\). Under \(\mathbb {P}\), the \((\mathfrak {B}_{N,x})_{x \in V_N}\) are independent Bernoulli random variables with parameter \(p_N = \mathbb {P}[\mathfrak {B}_{N,x} = 1] \sim \delta ^{-\alpha } {\mathsf {d}}_N^{-1}\) by (6.4). Lemma 9.5 thus reflects a classical large deviation estimate. More precisely, for any \(\lambda > 0\),
with
Recalling that \(p_N \sim \delta ^{-\alpha } {\mathsf {d}}_N^{-1}\), we see that if \(\lambda > 0\) is chosen small enough and \(N\) is large, then
and since \(|V_N| {\mathsf {d}}_N^{-1} \geqslant e^{cN}\) for some \(c > 0\), we obtain, for any \(N\) large enough,
The probability
is handled in the same way. \(\square \)
Lemma 9.6
For any \(\varepsilon > 0\), any \(M\) large enough and any \(N\) large enough,
Proof
The difference \(\tilde{T}^M_N(1) - T^M_N(1)\) consists in the sum of the random variables \(\tilde{e}_N(1)\), which is exponential with mean \(N^2\), and \({\mathsf {T}}_N\), which is of order \(N\) by Proposition 3.1. Since \({\mathsf {t}}_N\) grows exponentially with \(N\), it suffices to prove that the lemma holds with \(T^M_N(1)\) instead of \(\tilde{T}^M_N(1)\).
We know from Proposition 7.1 that \({\mathsf {t}}_N^{-1} T^{(\delta )}_N(1)\) converges in law under \(\overline{\mathbb {P}}_N\) to \(T^{(\delta )}(1)\), an exponential random variable of parameter \(\delta ^{-\alpha }\). Recall that \(T^M_N(1)\) is defined as \(T^{(\delta )}_N(1) \wedge (M{\mathsf {t}}_N)\), and hence \({\mathsf {t}}_N^{-1} T^M_N(1)\) converges in law to
and is always bounded by \(M\). As a consequence, for \(M\) and \(N\) large enough, we have
In order to conclude, it thus suffices to show that the variance of \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[ {T}^M_N(1)]\) is much smaller than \({\mathsf {t}}_N^2\). The proof of this fact is similar to the end of the proof of Theorem 5.1, where we showed that the variance of \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}[D_N({\mathsf {t}}_N)]\) is much smaller than \({\mathsf {t}}_N^2\).
Recall that we write \(\mathbb {V}[{\mathsf {X}}]\) to denote the variance of the random variable \({\mathsf {X}}\) with respect to \(\mathbb {P}\). By the Efron–Stein inequality and translation invariance, we have
where \(\tau _N^{(0)}\) is the environment which coincides with \(\tau _N\), except at \(0\) where \(E_0\) has been replaced by an independent copy of it. As in the proof of Theorem 5.1, we introduce \(T_1\) the first time the point \(0\) is discovered by the process \(X\); we let \({\mathsf {T}}^{(1)}\) be the first strong stationary time after \(T_1\). We iterate these definitions: \(T_2\) is the first time \(0\) is discovered after \({\mathsf {T}}^{(1)}\), then \({\mathsf {T}}^{(2)}\) is the first strong stationary time after \(T_2\), and so on. We define \(I = \cup [T_j,{\mathsf {T}}^{(j)})\), and
This definition would coincide with that of \(T^M_N(1)\) if the constraint \(t\notin I\) was suppressed. The quantity \(T_N^{(0)}\) is convenient for our purpose, because its law under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) does not depend on \(\tau _{N,0}\), and is thus the same under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) and under \(\mathbf{P}^{\tau _N^{(0)}}_{\mathfrak {u}}\). We can thus write
Since these two terms have the same law under \(\mathbb {P}\), it suffices for our purpose to show that
Let \({\mathcal {T}}\) be the event defined by
where \({\mathsf {T}}_N\) is the strong stationary time of Proposition 3.1. By the Markov property, we have
We have seen in (5.15) that the sum on the right-hand side gives a contribution which, up to sub-exponential factors, is bounded by \({\mathsf {t}}_N/2^N\). There remains to evaluate
where in the last step, we simply bounded the square of the probability by the probability itself, and then used translation invariance. We will show that, up to sub-exponential factors, the quantity above is bounded by \({\mathsf {t}}_N^{-1}\), that is,
Assuming this to be true, we obtain that up to sub-exponential factors, the left-hand side of (9.10) is bounded by
Since \(\log ({\mathsf {t}}_N) \sim \overline{c} N\) with \(\overline{c} < \log (2)\), this proves (9.10), and thus completes the proof of the lemma.
There remains to justify (9.11). In view of Proposition 3.1, the probability that \({\mathsf {T}}_N\geqslant N^3\) is super-exponentially small. Hence, this event can be discarded, and we can focus on bounding the probability of the event
When at a site \(x\), the walk \(X\) spends an exponentially distributed time of parameter \(\omega _N(x) \leqslant \overline{\omega }_N\). As was seen in Lemma 5.2, outside of an event whose probability is super-exponentially small, we have \(\overline{\omega }_N \leqslant e^{N^{3/4}}\). It thus follows that outside of an event whose probability is super-exponentially small, the number of sites visited by the walk up to time \(N^3\) is bounded by \(e^{N^{4/5}}\). Indeed, if \(({\mathsf {e}}_k)_{k \in \mathbb {N}}\) are independent random variables of parameter \(e^{N^{3/4}}\), then a standard large deviation estimate shows that the probability of the event
decays super-exponentially fast with \(N\). As a consequence, outside of an event whose \(\overline{\mathbb {P}}_N\)-probability is super-exponentially small, the walk does not discover more than \((N+1)e^{N^{4/5}} \leqslant e^{N^{5/6}}\) sites up to time \(N^3\). But we know from Proposition 4.1 that under \(\overline{\mathbb {P}}_N\), the sequence of depths of newly discovered traps is i.i.d. As a consequence, the probability that a deep trap is found among the \(e^{N^{5/6}}\) first discovered sites is bounded by
see (6.4). Recalling that \(\log ({\mathsf {d}}_N) \sim \log ({\mathsf {t}}_N)\), we obtain (9.11), and thus complete the proof. \(\square \)
With the two lemmas above, we can now control the expectation of the right-hand side of (9.8).
Proposition 9.7
For any \(\varepsilon > 0\) and any \(N\) large enough, if \(\Vert h_N\Vert _\infty \leqslant 1\), then
Proof
Under the expectation on the left-hand side of (9.13) appears a product of three terms, which for convenience we rewrite as \({\mathsf {X}}_N {\mathsf {Y}}_N {\mathsf {Z}}_N\), with obvious identifications. We first show that
Let \(\mathfrak {E}_N\) be the event
and \(\mathfrak {E}_N^c\) be the complementary event. We decompose \(\mathbb {E}[\mathsf {X}_N \mathsf {Y}_N \mathsf {Z}_N]\) as
Since \(\mathsf {X}_N, \mathsf {Z}_N \leqslant 1\), while \(\mathsf {Y}_N \leqslant M{\mathsf {t}}_N\) and \({\mathsf {t}}_N\) grows exponentially with \(N\), the second term in the above sum is smaller than \(e^{-N}\) by Lemma 9.5. For the first term, the definition of \(\mathfrak {E}_N\) implies that
Reasoning as above, we can show that
so that
and this proves (9.14).
Let us now see how a similar reasoning enables to show that
Define the event \(\mathfrak {F}_N\) as
We decompose \(\mathbb {E}[\mathsf {Y}_N \mathsf {Z}_N]\) as
Since \(\mathsf {Y}_N \leqslant M{\mathsf {t}}_N\) and \(\mathsf {Z}_N \leqslant 1\), Lemma 9.6 ensures that
On the other hand,
with
To sum up, we have shown that for any \(\varepsilon > 0\) and any \(N\) large enough,
Recalling that \(\mathsf {Z}_N \leqslant 1\), we obtain (9.15) by suitably tuning \(\varepsilon \). Combining (9.14) and (9.15) yields
Since \(\mathsf {Z}_N\leqslant 1\) and \(\log ({\mathsf {d}}_N/{\mathsf {t}}_N) = o(N)\), this completes the proof as long as we can show that
We have
where the fraction in the right-hand side above is understood to be \(1\) when the probability in the denominator is \(0\). We have seen in Lemma 9.5 (used with (6.4)) that with probability tending to \(1\) as \(N\) tends to infinity,
while
Equation (9.16) follows from these observations, and this finishes the proof. \(\square \)
We now consider the left-hand side of (9.8), and start with the following lemma.
Lemma 9.8
Let \(\mathfrak {D}^M_N\) be the event
There exists \(c > 0\) such that for any \(M\),
Proof
Note first that the difference \(\tilde{T}^M_N(1)-T^M_N(1)\) consists in the sum of the random variable \(\tilde{e}_N(1)\) (exponentially distributed with mean \(N^2\)), and of a random variable distributed as \(\mathsf {T}_N\) appearing in Proposition 3.1. By part (2) of this Proposition, we thus know that outside of an event whose probability decays exponentially with \(N\), we have
It is thus sufficient to prove the claim with the time interval \((T^M_N(1),\tilde{T}^M_N(1)]\) replaced by \((T^M_N(1),{T}^M_N(1) + N^3]\). What we need to do then is almost the same as what we proved in the end of the proof of Lemma 9.6. Namely, the argument there ensures that outside of an event whose probability is super-exponentially small, the walk does not discover more than \(e^{N^{5/6}}\) sites during this time interval. But we know from Proposition 4.1 that under \(\overline{\mathbb {P}}_N\), the sequence of depths of newly discovered traps is i.i.d. As a consequence, the probability that a deep trap is found in no more than \(e^{N^{5/6}}\) newly discovered sites is bounded by
see (6.4). Since \({\mathsf {d}}_N\) grows exponentially with \(N\), this finishes the proof. \(\square \)
We consider the total time spent at site \(x^{(\delta )}_N(1)\) from the moment of discovery \(T^{(\delta )}_N(1)\) up to time \(T^{(\delta )}_N(1) + \tilde{e}_N(1)\), that is,
The next proposition starts the computation of the expectation of the left-hand side of (9.8).
Proposition 9.9
There exists \(c > 0\) such that for any \(M\) and any \(N\) large enough, if \(\Vert h_N\Vert _\infty \leqslant 1\), then
Proof
Note that in the integral
the interval of integration can be replaced by \([T^M_N(1), \tilde{T}^M_N(1)]\), since the walk does not discover any deep trap before \(T^M_N(1)\). Since we assume \(\Vert h_N\Vert _\infty \leqslant 1\), the integral is bounded by \(\tilde{T}^M_N(1) - {T}^M_N(1)\), which for short we simply write \(\Delta ^M_N\).
Our first step is to estimate the difference
In words, the second integral counts the time spent by the walk only on the first deep trap discovered, and only if this deep trap is discovered before time \(M{\mathsf {t}}_N\). The difference in (9.19) is thus bounded by
where \(\mathfrak {D}^M_N\) is the event defined in (9.17). The \(\mathbb {E}\)-expectation of this term is bounded by
The first term of this sum is exponentially small in \(N\) by Lemma 9.8. So is also the second term, since \(\Delta ^M_N = \tilde{e}_N(1) + \mathsf {T}_N \circ \Theta _{T^M_N(1)+\tilde{e}_N(1)}\) and \(\tilde{e}_N(1)\) is an exponential random variable with mean \(N^2\), while \(\mathsf {T}_N-N\) is stochastically dominated by an exponential random variable with mean \(N\) by Proposition 3.1. We have thus shown that there exists \(c > 0\) such that for any \(N\) large enough,
On the event \(T^{(\delta )}_N(1)\leqslant M {\mathsf {t}}_N\), we have \(T^M_N(1) = T^{(\delta )}_N(1)\). By the definition of \(l^{(\delta )}_N\) (see (9.18)), on this event, we have
We now wish to show that for some \(c > 0\) and \(N\) large enough,
Since \(\Vert h_N\Vert _\infty \leqslant 1\), we may as well show that
In order to do so, let us define \(\mathfrak {D}'_N\) the event
Arguing as above, we see that the probability of this event is exponentially small in \(N\). Proceeding as in the analysis of the term (9.20), it then follows that
In order to prove (9.23), it thus suffices to show that
Recall that as \(x^{(\delta )}_N(1)\) is discovered at time \(T^{(\delta )}_N(1)\), it must be that \(|X_{T^{(\delta )}_N(1)} - x| \leqslant 1\). Hence,
By the Markov property, the latter integral is equal to
By Proposition 3.4, the probability underbraced above is exponentially small in \(N\), uniformly over \(\tau _N,\,X_{T^{(\delta )}_N(1)}\) and \(s \geqslant N\). This completes the proof of (9.23). Equations (9.21), (9.22) and (9.23) imply the proposition. \(\square \)
We can now conclude our analysis of the expectation of the left-hand side of identity (9.8).
Proposition 9.10
There exists \(c > 0\) such that for any \(M\) and any \(N\) large enough, if \(\Vert h_N\Vert _\infty \leqslant 1\), then
where we recall that the event \({\mathcal {E}}^M_N\) was defined in (9.2).
Proof
First, note that
Indeed, the quantities under the two expectations above are equal unless the deep trap \(x^{(\delta )}_N(1)\) is visited in the time interval \((T^{(\delta )}_N(1)+N, T^{(\delta )}_N(1)+\tilde{e}_N]\). Hence (9.24) is obtained as in the proof of Proposition 9.9, noting that the probability of this event is exponentially small in \(N\) by Proposition 3.4.
Let us write \(H^{(\delta )}_N(n)\) for the first time the walk visits the site \(x^{(\delta )}_N(n)\), and define
By a similar reasoning, we have
Indeed, the quantities under the two expectations above are equal unless the deep trap \(x^{(\delta )}_N(1)\) is visited during the time interval \((T^{(\delta )}_N(1) + \tilde{e}_N(1) ; H^{(\delta )}_N(1) + \tilde{e}_N(1)]\). On the event \({\mathcal {E}}^M_N\), we have that the length of this interval is smaller than \(N\). Moreover, outside of an event whose probability is exponentially small, we have \(\tilde{e}_N(1) \geqslant N\). So we can argue as above and obtain (9.26).
Let \(e^{(\delta )}_N(n)\) be such that
Under the measure \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) and conditionally on \(x^{(\delta )}_N(1) = x\), the quantity \(\tilde{l}^{(\delta )}_N(1)\) is independent of \((X_t)_{t \leqslant H^{(\delta )}_N(1)}\). Moreover, it is the total time spent on \(x\) by the walk \(X\) starting at \(x\) and with a killing parameter equal to \(1/N^2\). In particular, \(\tilde{l}^{(\delta )}_N(1)\) is an exponential random variable with mean \(G_N(\tau ^{(\delta )}_N(1))\). In other words, we thus see that under the measure \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\) and conditionally on \(x^{(\delta )}_N(1) = x\), the quantity \({e}^{(\delta )}_N(1)\) is independent of \((X_t)_{t \leqslant H^{(\delta )}_N(1)}\) and follows an exponential distribution with parameter \(1\). Since this does not depend on \(x\), we have thus shown that under the measure \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\), the quantity \({e}^{(\delta )}_N(1)\) is independent of \((X_t)_{t \leqslant H^{(\delta )}_N(1)}\) and follows an exponential distribution with parameter \(1\). In particular, (and since \({\mathcal {E}}^M_N\) is measurable with respect to \((X_t)_{t \leqslant H^{(\delta )}_N(1)}\)), we have
Taking expectations and using (9.26), we obtain the announced result. \(\square \)
Proof of Theorem 9.1
The theorem follows by combining the identity (9.8) with Propositions 9.7, 9.9 and 9.10, and recalling that \(\log ({\mathsf {d}}_N/{\mathsf {t}}_N) = o(N)\). \(\square \)
Of course, Theorem 9.1 is truly interesting only if the probability of \({\mathcal {E}}^M_N\) can be taken arbitrarily close to \(1\). It follows from Proposition 7.1 that \(\overline{\mathbb {P}}_N[T^{(\delta )}_N(1) \leqslant M {\mathsf {t}}_N]\) can be taken arbitrarily close to \(1\) by choosing \(M\) large enough. Our goal is now to prove that \({\mathcal {H}}^{(\delta )}_N(n)\) occurs with high probability.
Proposition 9.11
For any \(n \in \mathbb {N}\),
Let us write \(x \approx y\) when \(x,y \in V_N\) are neighbours or second neighbours. The proof of the proposition relies on the following lemma.
Lemma 9.12
For any \(n \in \mathbb {N}\),
Proof
We introduce some terminology (slightly adapted from [37]). For any \(z \approx 0\), we say that \(x \in V_N\) is \(z\)-atypical if it is a deep trap (that is, \(\tau _{N,x} \geqslant \delta B_N\)) and moreover, \(E_{x+z} > 2 \sqrt{\log N}\). The lemma can be rephrased as stating that
tends to \(0\) as \(N\) tends to infinity.
We say that \(x \in V_N\) is \(z\)-uncommon if it is \(z\)-atypical, or if \((x-z)\) is \(z\)-atypical. Finally, for a subset \(\Gamma \subseteq V_N\), we say that \(x\) is \(z\)-uncommon regardless of \(\Gamma \) if one can infer that \(x\) is \(z\)-uncommon without considering sites inside \(\Gamma \), i.e. if one of the two following conditions occur:
We learn from [37, Lemma 7.2] that if \(x^{(\delta )}_N(n)\) is \(z\)-atypical, then there exists \(k \leqslant r^{(\delta )}_N(n)\) such that \(x_k\) is \(z\)-uncommon regardless of \(\{x_1,\ldots ,x_{k-1}\}\) (recall that \(x_1,x_2,\ldots \) denotes the exploration process introduced in Sect. 4). As was seen in the proof of Proposition 7.1, the random variable \({\mathsf {d}}_N^{-1} r^{(\delta )}_N(n)\) converges in distribution under \(\overline{\mathbb {P}}_N\), hence the probability
can be made as close to \(1\) as desired by choosing \(C_r\) sufficiently large. Moreover,
As was observed in the proof of [37, Proposition 7.1], the invariance of \(\mathbb {P}\) under translations implies that the probability underbraced above is bounded by
which in turn is bounded by
Using (6.4) and (6.1), we obtain that for \(N\) large enough, the probability on the left-hand side of (9.29) is bounded by
which tends to \(0\) as \(N\) tends to infinity. \(\square \)
Proof of Proposition 9.11
Let us condition on the event that the position of the walk at time \(T^{(\delta )}_N(n)\) is \(x \in V_n\). Note that in this case, \(x^{(\delta )}_N(n)\) must be a neighbour of \(x\). On this event, the \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\)-probability that the next jump of the walk is from \(x\) to \(x^{(\delta )}_N(n)\) is given by
Recall that \(a_N \geqslant 1\) and \(a_N \leqslant \overline{a}\sqrt{2 \log N}\). Since \(x^{(\delta )}_N(n)\) is a deep trap, the numerator above is larger than \(\exp (\sqrt{2\overline{c} N}(1+o(1)))\), while the sum appearing in the denominator has no more than \(N\) terms, each of them being smaller than \(\exp (2\sqrt{2} \overline{a}\log N)\) on the event that
Using Lemma 9.12, we can thus conclude that with probability tending to \(1\), the walk jumps to \(x^{(\delta )}_N(n)\) right after having discovered it. The total time spent on any site is an exponential random variable whose parameter is at least \(N\), so this jump occurs fast enough to ensure that Proposition 9.11 holds. \(\square \)
10 The Green function at the origin
Let us write \(\mathbb {P}^\circ = \mathbb {P}[ \ \cdot \ | \ \tau _{N,0} \geqslant \delta B_N]\), and \(\overline{\mathbb {P}}_N^\circ = \mathbb {P}^\circ \mathbf{P}^{\tau _N}_0\). The purpose of this section is to prove the following result.
Theorem 10.1
Under the measure \(\mathbb {P}^\circ \), the random variable \(N^2\phi (a_N)^2 \ G_N(\tau _N)\) converges to \(1\) in probability, where we recall that \(\phi \) is the function introduced in (2.5).
The idea of the proof consists in (1) computing the expected time the walk spends on the origin before exiting its \(1\)-neighbourhood, and (2) showing that once the walk has exited the \(1\)-neighbourhood of the origin, it will not come back to it sufficiently fast to be accounted in \(G_N(\tau _N)\). The most technical part is (2), which is achieved in two main steps. In the first step (Proposition 10.6), we show that there exists \(\zeta > 0\) such that with probability tending to \(1\), the walk does not come back to the origin up to time \(N^{\zeta - 1}\). In the second step, we rely on Proposition 3.5 to justify that the walk will not come back to the origin soon enough after \(N^{\zeta - 1}\). This is slightly delicate, since the statement in Proposition 3.5 is averaged over \(\mathbb {P}\), while we consider here an average over \(\mathbb {P}^\circ \).
For any \(n \in \mathbb {N}\), we let \(H_n = \inf \{ t \geqslant 0 : |X_t - X_0| = n\}\).
Lemma 10.2
Under \(\mathbf{P}^{\tau _N}_x\), the total time spent at the initial position up to \(H_2\):
follows an exponential distribution with parameter
Proof
Under \(\mathbf{P}^{\tau _N}_x\), the walk stays at the origin an exponential time of parameter
Then it jumps to some \(y \sim x\) with probability proportional to \(e^{a_N E_y}\). The fraction in (10.1) is thus the probability that the walk hits a point at distance \(2\) from \(x\) before returning to \(x\). The result then follows from these observations. \(\square \)
Proposition 10.3
Recall that we assume \(\overline{a} < 1/20\). For any \(\varepsilon > 0\), we have
Remark 10.4
For the purpose of this section, it would be sufficient to show that the probability in the left-hand side of (10.2) tends to \(1\) as \(N\) tends to infinity. The quantitative estimate given here will prove important in the next section.
Proof of Proposition 10.3
We first prove that for any \(\varepsilon > 0\),
and we do so by a computation of moments. We write \(\chi _N(x)\) for \(e^{a_N E_x} - \mathbb {E}[e^{a_N E_x}] = e^{a_N E_x} - \phi (a_N)\). For any integer \(m\), we have
The last equality is obtained in the following way. For every sequence \((x_1,\ldots , x_{2m})\), we let \(k\) be the cardinality of \(\{x_1,\ldots ,x_{2m}\}\), and write this set as \(\{y_1,\ldots , y_k\}\), where the \(y_i\)’s are pairwise distinct. Then \(e_i\) \((\geqslant \!\! 1)\) is the number of occurrences of \(y_i\) in the sequence \((x_1,\ldots , x_{2m})\). It is an interesting combinatorial exercise to see that the coefficient \(C_{e_1,\ldots ,e_k}\) appearing in the formula is the multinomial coefficient associated with \((e_1,\ldots , e_k)\) divided by \(k!\), but the only important thing for us is that this coefficient does not depend on \(N\). Since \((\chi _N(x))_{x \sim 0}\) are independent and centred random variables, the expectation \(\mathbb {E}\left[\chi _N(y_1)^{e_1} \cdots \chi _N(y_k)^{e_k}\right]\) vanishes as soon as one of the \(e_i\)’s is equal to \(1\). Hence, up to a constant (depending only on \(m\)), the left-hand side of (10.4) is bounded by
Up to a constant, the quantity above is bounded by
where we used the fact that \(\phi (\lambda ) \leqslant 2 e^{\lambda ^2/2}\) (see (2.5)). Since the square function is convex and takes value \(0\) at \(0\), we have
We have thus shown that for any \(m\), there exists a constant \(C\) such that
Choosing \(m = 4\) and recalling that \(\overline{a} < 1/20\), we obtain
An application of Chebyshev’s inequality then proves (10.3).
For any \(\varepsilon > 0\), we consider the event formed by the conjunction of
and for any \(x \sim 0\),
It follows from (10.3) that this event has probability at least \(1 - o(N^{-2})\). On this event, we have
and thus the result is proved. \(\square \)
Remark 10.5
Although the proof makes it transparent that the condition \(\overline{a} < 1/20\) is not really necessary, one can convince oneself that Proposition 10.3 does not hold for every \(\overline{a} < 1\).
Proposition 10.6
Let \(\zeta > 0\), and let \(\mathfrak {A}_N\) be the event
Recall that we assume \(\overline{a} < 1/20\). If \(\zeta \) is sufficiently small, then
Remark 10.7
Recall that \(| \cdot |\) denotes the Hamming distance on the hypercube. By a slight abuse of notation, we will simply write \(|x|\) instead of \(|x-0|\), when \(x \in V_N\).
Proof
We will first show that with probability tending to \(1\), the walk starting from time \(H_2\) keeps going away from the origin at least up to distance \(N^{1/4}\). We then show that a good proportion of these steps take time at least \(N^{-1-5\overline{a}}\), thus justifying that on this event, the walk has not returned to the \(1\)-neighbourhood of the origin up to time \(N^{-3/4-5\overline{a}}\). This will prove the proposition since we assume \(\overline{a} < 1/20\).
We begin by introducing \((\overline{Y}_n)_{n \geqslant 2}\) the discrete-time walk that records the successive positions of \((X_t)_{t \geqslant H_2}\). The indexation is chosen so that \(\overline{Y}_2\) is at distance \(2\) from the origin (that is, \(|\overline{Y}_2| = 2\)).
We also introduce a slightly modified version of the exploration process. This new exploration process, that we denote by \(\overline{x}_1, \overline{x}_2, \ldots \), is obtained from the original one by suppressing all sites belonging to \(\{x : |x| \leqslant 2\}\). Note that each step of \(\overline{Y}\) gives rise to a (possibly empty) set of newly discovered sites (for this modified exploration process). For \(n \geqslant 2\), we write \(\overline{\mathcal {X}}_{n+1}\) for the set of sites thus discovered by \(\overline{Y}\) at (discrete) time \(n\) (see Fig. 1). For \(n \in \{1,2\}\), we also define
Under \(\overline{\mathbb {P}}_N^\circ \), the distribution of \(\overline{Y}\) up to time \(n\) does not depend on \((E_{\overline{x}_i})_{i \in \overline{\mathcal {X}}_{n+1}}\), and therefore the latter is a family of independent random variables distributed as \({\mathcal {N}}_+(0,1)\), independent of \((E_{\overline{x}_i})_{i \in \bigcup _{k \leqslant n} \overline{\mathcal {X}}_{k}}\) (this is the same argument as in Proposition 4.1).
Finally, for \(n \geqslant 2\), we let
In words, to form \({\mathcal {X}}_{n+1}\), one waits until the first time the random walk visits a point at distance \(n\) from the origin, and then collects all neighbouring points that are at distance \(n+1\) from the origin. The successive sets \({\mathcal {X}}_3, {\mathcal {X}}_4,\ldots \) can be seen as yet another exploration process, this one being very far from ultimately covering the whole graph. Again, under \(\overline{\mathbb {P}}_N^\circ \), the distribution of \((X_t)_{t \leqslant H_{n}}\) does not depend on \((E_x)_{x \in {\mathcal {X}}_{n+1}}\), and the latter is thus a family of independent random variables distributed as \({\mathcal {N}}_+(0,1)\), and is independent of \((E_x)_{x \in \bigcup _{k \leqslant n} {\mathcal {X}}_k}\). Note that for \(2 \leqslant n \leqslant N\), we have \(|{\mathcal {X}}_{n+1}| = N-n\) (see Fig. 1).
Let \(\mathfrak {A}_N(n)\) be the event defined to be the conjunction of the event \(|\overline{Y}_n| = n\) and of the event
We now prove by induction on \(n\) that if (\(N\) is large enough and) \(2\leqslant n\leqslant N^{1/4}\), then
We start by checking that such is the case for \(n = 2\). The walk always satisfies \(|\overline{Y}_2|=2\), so we only need to bound the probability of the event
Noting that \(|\overline{\mathcal {X}}_1 \cup \overline{\mathcal {X}}_2| = N(N+1)/2\) and using (5.4), we can bound this probability from below by
and this justifies the base case of the induction.
Let \(2 \leqslant n < N^{1/4}\), and assume that (10.7) holds. It suffices to check that
where we write \(\mathfrak {A}_N^c(n+1)\) for the complement of \(\mathfrak {A}_N(n+1)\). The probability on the left-hand side above is smaller than the sum of
and
Since \((E_x)_{x \in \overline{\mathcal {X}}_{n+1}}\) is independent of \(\mathfrak {A}_N(n)\), we can proceed as on the base case of this induction and show that the probability in (10.9) is \(O(N^{-3})\). We now consider the probability in (10.10). Recall that \(\overline{Y}\) chooses to jump to a neighbouring site \(x\) with probability proportional to \(e^{a_N E_x}\). On the event \(|\overline{Y}_n| = n,\,\overline{Y}_n\) has \(N-n\) neighbours at level \((n+1)\) (the elements of \({\mathcal {X}}_{n+1}\)), and \(n\) other neighbours belonging to \(\bigcup _{k \leqslant n} \overline{\mathcal {X}}_k\). Since the energy on any site is always positive, on the event \(\mathfrak {A}_N(n)\), the probability that \(|\overline{Y}_{n+1}| = n+1\) is at least
Since we assume \(n < N^{1/4}\) and \(a_N \leqslant \overline{a} \sqrt{2 \log N}\), the quantity above is larger than
This justifies (10.8), and thus completes the proof of (10.7).
It follows from (10.7) that with probability tending to \(1\), the event \(\mathfrak {A}_N(N^{1/4})\) is realized. We now explain why the walk cannot run its first \(N^{1/4}\) steps too fast.
First, we recall that under \(\overline{\mathbb {P}}_N^\circ \), the values \((E_x)_{x \in {\mathcal {X}}_3}, (E_x)_{x \in {\mathcal {X}}_4}, \ldots , (E_x)_{x \in {\mathcal {X}}_{N}}\) form a family of independent random variables distributed as \({\mathcal {N}}_+(0,1)\). Since \(|{\mathcal {X}}_{n+1}| \leqslant N\) and \(\mathbb {E}[e^{a_N E_x}] = \phi (a_N) \leqslant 2 e^{a_N^2/2}\), Chebyshev’s inequality guarantees that if \(3 \leqslant k \leqslant N\), then
By independence, a classical large deviation estimate guarantees that the \(\overline{\mathbb {P}}_N^\circ \)-probability that
tends to \(1\) as \(N\) tends to infinity. Let us write \(\mathfrak {A}'_N\) for the event that both \(\mathfrak {A}_N(N^{1/4})\) and the event displayed in (10.13) are satisfied. By (10.7), we thus have
Let \(\gamma = (\gamma _n)_{2 \leqslant n \leqslant N^{1/4}}\) be a path. We write \(\overline{Y} = \gamma \) to denote the event that for any \(n,\,2\leqslant n \leqslant N^{1/4}\), we have \(\overline{Y}_n = \gamma _n\). In a given environment \(\tau _N\), the event \(\mathfrak {A}'_N\) depends only on \((\overline{Y}_n)_{2 \leqslant n \leqslant N^{1/4}}\). Let \(\Gamma (\tau _N)\) be the set of paths \(\gamma \) such that in the environment \(\tau _N\) and on \(Y = \gamma \), the event \(\mathfrak {A}'_N\) is realized.
We now fix some \(\gamma \in \Gamma (\tau _N)\). Conditionally on \(Y = \gamma \), the waiting times of the continuous-time walk \(X\) on the successive sites \((\gamma _n)_{2 \leqslant n \leqslant N^{1/4}}\) are independent exponential random variables with respective parameter
On \(Y = \gamma \), the event \(\mathfrak {A}'_N\) is realized, and in particular:
-
the set \(\{x : x \sim \gamma _n\}\) is partitioned into the \(N-n\) sites that belong to \({\mathcal {X}}_{n+1}\) on one hand, and the \(n \leqslant N^{1/4}\) sites that belong to some \(\overline{\mathcal {X}}_k\) for some \(k \leqslant n\) on the other;
-
the site \(\gamma _n\) itself belongs to \(\overline{\mathcal {X}}_k\) for some \(k \leqslant n\);
-
The events (10.6), (10.13) hold, and \(\gamma \) does not visit the origin.
These observations enable to bound the parameter in (10.15) by
and for at least \(N^{1/4}/4\) of these \(n\)’s, the latter is smaller than
To sum up, under the measure \(\mathbf{P}^{\tau _N}_0\) and conditionally on the event \(Y=\gamma \), the total time it takes for the continuous-time walk \(X\) to travel through \(\gamma \) stochastically dominates a sum of \(N^{1/4}/4\) independent exponential random variables of parameter bounded by \(5N^{1+4\overline{a}+\overline{a}^2}\). With probability tending to \(1\) (uniformly over \(\tau _N\) and \(\gamma \in \Gamma (\tau _N)\)), this sum is larger than \(N^{-3/4+5\overline{a}}\). Since we assume \(\overline{a} < 1/20\), there exists \(\zeta > 0\) such that \(-3/4+5\overline{a} = -1+\zeta \). In particular, we have shown that, uniformly over \(\tau _N\),
Moreover,
From (10.16), it thus follows that for any \(N\) large enough, uniformly over \(\tau _N\),
Integrating with respect to \(\mathbb {P}^\circ \) and using (10.14), we obtain the desired result. \(\square \)
Remark 10.8
There are many ways to improve the proof of Proposition 10.6 so that it covers higher values of \(\,\overline{a}\). Yet, we now explain why a simple optimization of exponents cannot cover every \(\overline{a} < 1\). Let \(\xi \) be an exponent such that with probability tending to \(1\), the event \(|\overline{Y}_{N^\xi }| = N^\xi \) occurs with probability tending to \(1\). We have seen in the proof of Proposition 10.6 that for \(\overline{a}\) small enough, one can take \(\xi = 1/4\). Yet, as \(\overline{a}\) is taken closer and closer to \(1\) (and fixing \(a_N = \overline{a} \sqrt{2 \log N}\)), the exponent \(\xi \) must approach \(0\), while the time spent on a site during the descent is an exponential random variable with parameter of order at least \(N \phi (a_N) \sim N^{1+\overline{a}^2}\). That is, the total time spent on the full descent is of order no more than \(N^{-\overline{\zeta }}\) with \(\overline{\zeta }\) approaching \(2\) as \(\overline{a}\) is taken closer and closer to \(1\).
Lemma 10.9
Let \(\zeta \in (0,1)\). There exists a measurable \(\Omega _N' \subseteq \Omega _N\) such that \(\mathbb {P}\left[\Omega _N'\right]\) tends to \(1\) as \(N\) tends to infinity, and moreover, for any \(\tau _N \in \Omega _N'\),
Proof
Proposition 3.5 (or more precisely, inequality (3.3)) ensures that for any \(x \in V_N\),
which, for \(N\) sufficiently large, is smaller than \(\exp (-N^{\zeta }/2)\). By a union bound and Chebyshev’s inequality, the \(\mathbb {P}\)-probability that
tends to \(0\) as \(N\) tends to infinity. We now argue that the probability appearing on the left-hand side of (10.18) is, up to some power of \(N\), an upper bound for the probability in (10.17). More precisely, let us write
and observe that
where we applied the Markov property at time \(\mathsf {H}_N\) in the last step. The total time spent on site \(y \in V_N\) is an exponential random variable with parameter
Note that
In particular, since we assume \(\overline{a} \leqslant 1\), the \(\mathbb {P}\)-probability that
tends to \(1\) as \(N\) tends to infinity. We define \(\Omega _N'\) as the set where the event displayed in (10.20) and the complement of the event displayed in (10.18) hold. We have seen that \(\mathbb {P}\left[ \Omega _N'\right]\) tends to \(1\) as \(N\) tends to infinity. For \(\tau _N \in \Omega '_N\), we have
In view of (10.19), it thus follows that
Using also the fact that when \(\tau _N \in \Omega _N'\), the complement of the event displayed in (10.18) holds, we obtain that for any \(x \approx 0\),
and this finishes the proof. \(\square \)
Proof of Theorem 10.1
Let us write
where we recall that \(\tilde{e}_N(1)\) is an exponential random variable of parameter \(1/N^2\), independent of everything else. It follows from (9.1) that
Since \(\ell \) is the total time spent at a point by a certain transient Markov chain (the chain \(X\) with killing parameter \(1/N^2\)), we have in fact that under \(\mathbf{P}^{\tau _N}_0\), the random variable \(\ell \) follows an exponential distribution with mean \(G_N(\tau _N)\).
We decompose \(\mathbf{E}^{\tau _N}_0\left[\ell \right]\) as
Our first step is to show that the second term in this sum is small. Note that it may be negative, since it may happen that \(H_2 > \tilde{e}_N(1)\). We bound its absolute value by
The random variable \(H_2\) can be decomposed as the time spent on the origin before reaching level \(2\), say \(H_2^{(0)}\), plus the time spent at level \(1\) before reaching level \(2\), say \(H_2^{(1)}\). The random variable \(H_2^{(0)}\) follows an exponential distribution whose parameter is displayed in (10.1) (with \(x = 0\)), and takes the form
where
and
As a consequence, we have
which implies that \(H_2^{(0)}\) is dominated by an exponential random variable of parameter \(N/2\). As for \(H_2^{(1)}\), at any site at distance \(1\) from the origin, the walk may jump at distance \(2\) with a transition rate at least \((N-1)\). Hence, \(H_2^{(1)}\) is dominated by an exponential random variable of parameter \((N-1)\). The first term in (10.22) is smaller than
and this is exponentially small in \(N\) (uniformly over \(\tau _N\)) by the above observations, since
We now turn to the second term appearing in (10.22). Recalling the definition of \(\mathfrak {A}_N\) given in (10.5), we define \(\tilde{\mathfrak {A}}_N\) as the conjunction of \(\mathfrak {A}_N,\,H_2 \leqslant \tilde{e}_N(1)\) and \(\tilde{e}_N(1) \leqslant N^3\). By Proposition 10.6 and the above observations, we know that
It thus follows from Chebyshev’s inequality that there exists a sequence \(\varepsilon _N\) tending to \(0\) and such that
(indeed, it suffices to take \(\varepsilon _N^2 = {\overline{\mathbb {P}}_N^\circ [ \tilde{\mathfrak {A}}_N]}\)). From now on, we only consider environments \(\tau _N\) such that
We rewrite the second term appearing in (10.22) as
The second term in the right-hand side above is bounded by
by the Cauchy–Schwarz inequality. Since \(\ell \) is an exponential random variable of parameter \(G_N(\tau _N)\), the quantity underbraced above is equal to \(\sqrt{2} \ G_N(\tau _N)\). Since we also focus on environments such that (10.23) holds, we thus have
We now consider the first term in the right-hand side of (10.24). By the definition of the event \(\tilde{\mathfrak {A}}_N\), this term is (positive and) smaller than
By the Markov property at time \(H_2\), this is smaller than
where we wrote \(\overline{\mathfrak {A}}_N\) for the event
Since \(X_{H_2}\) is at distance \(2\) from the origin, the quantity in (10.26) is bounded by
The crucial point of our argument is that the event underbraced above does not depend on \(\tau _{N,0}\). Indeed, it is a function of the trajectory up to hitting the \(1\)-neighbourhood of \(0\), and this does not depend on \(\tau _{N,0}\). Let us write
where \(\Omega _N'\) is given by Lemma 10.9. The probability \(\mathbb {P}^\circ \left[\Omega _N^\circ \right]\) tends to \(1\) as \(N\) tends to infinity. For any \(\tau _N \in \Omega _N^\circ \), there exists \(\tau _N' \in \Omega _N'\) that agrees with \(\tau _N\) outside of \(0\). For such \(\tau _N\) and \(\tau _N'\), we have
and the right-hand side is smaller than \(\exp (-N^{\zeta /2})\) for any \(x \approx 0\) by Lemma 10.9. As a consequence, the left-hand side of (10.27), and thus also the first term in the right-hand side of (10.24), is smaller than \(\exp (-N^{\zeta /4})\) for all \(N\) sufficiently large, provided \(\tau _N \in \Omega _N^\circ \).
To sum up, we have shown that there exists \(\overline{\Omega }_N\) such that \(\mathbb {P}^\circ [\overline{\Omega }_N]\) tends to \(1\) as \(N\) tends to infinity, and moreover, for any \(\tau _N \in \overline{\Omega }_N\),
where \(c > 0\) and we recall that \(\lim \varepsilon _N = 0\). To conclude, it thus suffices to see that for any \(\varepsilon > 0\), there exists \(\Omega _N''\) such that \(\mathbb {P}^\circ [{\Omega }_N'']\) tends to \(1\) as \(N\) tends to infinity, and moreover, for any \(\tau _N \in \Omega _N''\),
as soon as \(N\) is sufficiently large. Indeed, showing (10.29) is sufficient in view of (10.28) since
which is much larger than \(\exp (-N^{\zeta /4})\). By Lemma 10.2, the left-hand side of (10.29) is equal to the inverse of the quantity in (10.1) taken for \(x = 0\). By Proposition 10.3, there exists \(\Omega _N''\) such that \(\mathbb {P}^\circ [{\Omega }_N'']\) tends to \(1\) as \(N\) tends to infinity, and moreover, for any \(\tau _N \in \Omega _N''\),
since the second condition does not involve \(\tau _{N,0}\). Under the conditions displayed in (10.30), the parameter in (10.1), taken for \(x = 0\), is equal to
Recalling that \(B_N = \exp (\beta \sqrt{N} b_N)\) and \(b_N \sim \sqrt{2\overline{c} N}\), it comes that if \(\tau _{N,0}= \exp (\beta \sqrt{N} E_0) \geqslant \delta B_N\), then it must be that \(E_0 \geqslant N^{1/4}\). Since \(a_N \geqslant 1\), we also have \(\exp (a_N E_0) \geqslant \exp (N^{1/4})\). Consequently, the fraction appearing in the right-hand side of (10.31) tends to \(1\) as \(N\) tends to infinity, and this completes the proof. \(\square \)
11 The Green function at the deep trap
The purpose of this section is to show the following theorem.
Theorem 11.1
Under the measure \(\overline{\mathbb {P}}_N\), the random variable \(N^2\phi (a_N)^2 \ G_N(\tau ^{(\delta )}_N(1))\) converges to \(1\) in probability.
What Theorem 9.1 says about \(G_N(\tau ^{(\delta )}_N(1))\) is that its size-biased distribution is the distribution of \(G_N(\tau _N)\) under \(\mathbb {P}^\circ \). In the previous section, we characterized the latter distribution. Knowing the size-bias of a distribution is however not sufficient to characterize it in general. Indeed, consider that any non-identically zero Bernoulli random variable has the constant \(1\) as its size-biased distribution. As we will now see, in order to conclude, it suffices to show that \(G_N(\tau ^{(\delta )}_N(1))\) cannot take values smaller than \(c N^{-2}\phi (a_N)^{-2}\) for some \(c > 0\). This is the content of the next proposition.
Proposition 11.2
We have
Proof
The quantity \(G_N(\tau ^{(\delta )}_N(1))\) is bounded from below by the inverse of the parameter displayed in (10.1) with \(x\) replaced by \(x^{(\delta )}_N(1)\). Moreover, since \(x^{(\delta )}_N(1)\) is a deep trap, we have \(E_{x^{(\delta )}_N(1)} \geqslant N^{1/4}\). Reasoning as in (10.31) and the few lines below, we see that in order to prove Proposition 11.2, it suffices to show that
We will prove (11.1) with the same technique as the one used to prove Lemma 9.12. Let \(\varepsilon \in (0,1)\), and
We say that a site \(z \in V_N\) is atypical if both
and
hold (where we write \(\mathsf {a} \ne (1\pm \varepsilon ) \mathsf {b}\) to mean that there exists no \(\mathsf {c}\) such that \(|\mathsf {c}|\leqslant \varepsilon \) and \(\mathsf {a} = (1+\mathsf {c})\mathsf {b}\)). We say that \(z \in V_N\) is uncommon if
For \(\Gamma \subseteq V_N\), we say that \(z\) is uncommon regardless of \(\Gamma \) if
In order to show that (11.1) holds, it suffices to show that
As in the proof of Lemma 9.12, if \(x^{(\delta )}_N(n)\) is atypical, then there must exist \(k \leqslant r^{(\delta )}_N(n)\) such that \(x_k\) is uncommon regardless of \(\{x_1,\ldots , x_{k-1}\}\). We recall that the probability in (9.28) can be made arbitrarily small if \(C_r\) is taken large enough. It thus suffices to show that
tends to \(0\) with \(N\). This probability is bounded by
Every summand is bounded by \({\mathbb {P}}\left[0 \text { is uncommon}\right]\), which itself is smaller than
The conclusion follows since \({\mathsf {d}}_N\mathbb {P}\left[\tau _{N,0} \geqslant \delta B_N\right]\) converges to a constant, \(\left|{\mathcal {W}}_N\right| = O(N^2)\), while the last probability in (11.3) is \(o(N^{-2})\) by Proposition 10.3. \(\square \)
Proof of Theorem 11.1
Let \(A\) be a measurable subset of \([1/2,+\infty )\), and define
We have \(\Vert h_N\Vert _\infty \leqslant 1\). Let \(\varepsilon > 0\). We learn from Theorem 9.1 that for any \(M\) large enough and any \(N\) large enough,
Moreover, if \(M\) is chosen sufficiently large, then for any \(N\) large enough, we have
as follows from Propositions 7.1 and 9.11. From now on, we fix \(M\) such that this and identity (11.4) hold. Identity (11.4) can be rewritten as
We first choose \(A = [1/2,+\infty )\). In this case, we know from (11.5) and Proposition 11.2 that the left-hand side is equal to \((1\pm 2\varepsilon )\) for all large enough \(N\). On the other hand, the expectation on the right-hand side converges to \(1\) by Theorem 10.1, since the integrand is bounded by \(2\). We thus get that for any \(N\) large enough,
Since this holds for arbitrary \(\varepsilon > 0\), we have shown that
Using this information, we can rewrite (11.6) as
We now choose \(A = [1-\eta ,1+\eta ]\). In this case, the expectation in the right-hand side above still converges to \(1\) by Theorem 10.1. Hence,
Using also (11.5), we thus obtain that for any \(N\) large enough,
Since this holds for arbitrary \(\varepsilon > 0\), we have shown that
The parameter \(\eta > 0\) being arbitrary, this completes the proof of Theorem 11.1. \(\square \)
As a by-product, we also proved (11.7), which we restate as a proposition for later usage.
Proposition 11.3
We have
12 Handling several traps
What we have achieved up to now is a thorough understanding of the asymptotic behaviour of certain quantities associated with the first deep trap visited, and most importantly, of the behaviour of the Green function at this location.
The purpose of this section is to show that from this information, we can infer the asymptotic joint behaviour of all the relevant quantities at all the deep traps visited, namely: the time it takes to find them, their depths, the value of the Green function there, and the total time actually spent on the sites. In order to do so, we use again the times introduced in (9.4) and (9.5), which we can slightly simplify by taking \(M = \infty \). In particular, we have \(T^\infty _N(1) = T^{(\delta )}_N(1)\). For values of \(n\) other than \(1\), it is not true in general that \(T^\infty _N(n) = T^{(\delta )}_N(n)\): the difference is that \(T^{(\delta )}_N(2)\) is the time when we discover a second deep trap, while \(T^\infty _N(2)\) is the time for the process to discover a first deep trap, then wait \(\tilde{e}_N(1)\), then wait a strong stationary time, and then come at distance \(1\) to a deep trap again (possibly the same as the first). The interest of these times is that they carry a lot of independence, as given by Proposition 9.4.
We define \(x^\infty _N(n)\) as the deep trap discovered at time \(T^\infty _N(n)\) (with some tie-breaking rule if multiple deep traps are simultaneously discovered). We write \(\tau ^\infty _N(n)\) for the environment around site \(x^\infty _N(n)\), that is, \(\tau ^\infty _N(n) = \theta _{x^\infty _N(n)} \ \tau _{N}\).
Proposition 12.1
Under the measure \(\overline{\mathbb {P}}_N\) and for every \(n,\,N^2\phi (a_N)^2 \ G_N(\tau ^\infty _N(n))\) converges to \(1\) in probability.
Proof
We know from Proposition 9.4 that the random variables \((\tau ^\infty _N(n))_{n \in \mathbb {N}}\) are independent and identically distributed under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\). In particular, for every \(\varepsilon > 0\) and every \(n\),
Recall that \(\tau ^\infty _N(1) = \tau ^{(\delta )}_N(1)\). Taking \(\mathbb {E}\)-expectation above and using Theorem 11.1, we obtain the result. \(\square \)
Recall that \({\mathcal {H}}^{(\delta )}_N(n)\) is the event that \(x^{(\delta )}_N(n)\) is visited during the time interval \([T^{(\delta )}_N(n), T^{(\delta )}_N(n)+N]\). For a time interval \(I\), we say that \(x\) is seen during \(I\) if there exists \(s \in I\) such that \(|X_s - x| \leqslant 1\). This differs slightly from the notion of being discovered during \(I\) in that it may happen that a point is seen during \(I\), but had already been discovered at a time prior to \(I\).
For \(k\in \mathbb {N}\), let \({\mathcal {F}}^{(\delta )}_N(k)\) be the event defined as the conjunction of
Proposition 12.2
-
(1)
For every \(k\), we have
$$\begin{aligned} \overline{\mathbb {P}}_N\left[ {\mathcal {F}}^{(\delta )}_N(k) \right] \xrightarrow [N \rightarrow \infty ]{} 1. \end{aligned}$$ -
(2)
For every \(k\in \mathbb {N}\), with \(\overline{\mathbb {P}}_N\)-probability tending to \(1\), we have \(x^\infty _N(n) = x^{(\delta )}_N(n)\) for every \(n \leqslant k\).
Proof
It was seen in Proposition 9.11 and Lemma 9.8 that the event described in (12.1) has \(\overline{\mathbb {P}}_N\)-probability tending to \(1\). We will show that for every \(n \in \mathbb {N}\) and every \(M > 0\),
Let us see why this would imply part (2) of the Proposition, and then part (1) as well. First, outside of an event whose probability can be made as close to \(0\) as desired, we can always assume that \(T^{(\delta )}_N(k+1) \leqslant M {\mathsf {t}}_N\), provided we choose \(M\) large enough. Hence, (12.3) implies
Note that we always have \(x^{(\delta )}_N(1) = x^\infty _N(1)\) and \(T^\infty _N(1) = T^{(\delta )}_N(1)\). Next, two things could make \(x^{(\delta )}_N(2)\) and \(x^\infty _N(2)\) different: (i) that a deep trap is discovered during the time interval \((T^\infty _N(1), \tilde{T}^\infty _N(1)]\) (this would define \(x^{(\delta )}_N(2)\), but not \(x^\infty _N(2)\)); or (ii) that the site \(x^{(\delta )}_N(1)\) is seen during the time \((\tilde{T}^\infty _N(1), T^{(\delta )}_N(2)]\) (this would define \(x^\infty _N(2)\), but not \(x^{(\delta )}_N(2)\)). With probability tending to \(1\), we have
so the \(\overline{\mathbb {P}}_N\)-probability of the events described in (i)–(ii) tend to \(0\) as \(N\) tends to infinity by (12.4) and the fact that the event in (12.1) has \(\overline{\mathbb {P}}_N\)-probability tending to \(1\). The reasoning can be continued to cover any \(n\), and we obtain part (2) of the proposition. Finally, part (2) and (12.4) ensure that indeed, the event in (12.2) has \(\overline{\mathbb {P}}_N\)-probability tending to \(1\) as well, and the proof is complete. \(\square \)
There remains to prove (12.3). In order to do so, we will need the following lemma.
Lemma 12.3
We write \(t_+ = \max (t,0)\). For every \(M > 0\) and \(x \in V_N\), we have
Proof of Lemma 12.3
Let
The left-hand side of (12.5) is larger than
Applying the Markov property at time \({\mathcal {T}}\), we obtain the result. \(\square \)
For \(M > 0\), we study the \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\)-probability that the site \(x^\infty _N(n)\) is visited during the time interval \([T^\infty _N(n)+N,T^\infty _N(n)+M{\mathsf {t}}_N]\). Using the Markov property at time \(T^\infty _N(n)\) and the Lemma at \(x = X_{T^\infty _N(n)}\), we see that this probability is bounded by the \(\mathbf{E}^{\tau _N}_{\mathfrak {u}}\)-expectation of
By Proposition 3.4, we know that the expectation in (12.6) is bounded by
and the latter decays exponentially fast since \(\log ({\mathsf {t}}_N) \sim \overline{c} N\) with \(\overline{c} < \log 2\). Note that the bound thus obtained on the expectation in (12.6) holds uniformly over the environment.
On the other hand, by Proposition 12.1, outside of an event whose \(\overline{\mathbb {P}}_N\)-probability tends to \(0\) with \(N\), we have
Since \(N^2 \phi (a_N)^2 \leqslant 2N^2a_N^2\) grows sub-expoentially with \(N\) (recall (2.5) and (2.2)), we have shown that
We now explain how we can infer (12.3) from this. We begin by proceeding as in the proof of Lemma 9.12 to show that
Since on the other hand, \(E_{x^\infty _N(n)} \geqslant b_n \sim \sqrt{2 \overline{c} N}\), any visit to a neighbour of \(x^\infty _N(n)\) is followed (with probability close to \(1\)) by a visit to \(x^\infty _N(n)\) itself shortly afterwards (by a reasoning identical to the proof of Proposition 9.11), so that the probability in (12.3) cannot be large if the one in (12.7) is close to \(0\).
Proposition 12.4
Under the measure \(\overline{\mathbb {P}}_N\) and for every \(n,\,N^2\phi (a_N)^2 \ G_N(\tau ^{(\delta )}_N(n))\) converges to \(1\) in probability.
Proof
This is a direct consequence of Proposition 12.1 and part (2) of Proposition 12.2. \(\square \)
Recall that we write \(H^{(\delta )}_N(n)\) for the first time the site \(x^{(\delta )}_N(n)\) is visited, and that we defined \(e^{(\delta )}_N(n)\) by the relation
Proposition 12.5
Under \(\overline{\mathbb {P}}_N\), the family of triples
jointly converges in distribution as \(N\) tends to infinity (where we let \(T^{(\delta )}_N(0) = 0\)). Let us write
for limiting random variables, which we assume to be defined on the \(\mathbb {P}\)-probability space for convenience (these can be taken as an extension of those appearing in Proposition 7.1, and we fix \(T_N^{(\delta )}(0) = 0\)). Their joint distribution is described as follows. The family of triples in (12.8) is a family of i.i.d. triples. Moreover, the random variables \(T^{(\delta )}(1),\,\tau ^{(\delta )}(1)\) and \(e^{(\delta )}(1)\) are themselves independent. Their respective distributions are exponential of parameter \(\delta ^{-\alpha }\), the law displayed in (7.1), and exponential of parameter \(1\).
Proof
We introduce a slight modification of the sequence of times \((T^\infty _N(n))_{n \in \mathbb {N}}\), that we write \((T'_N(n))_{n \in \mathbb {N}}\). They are constructed in the same way as the sequence \((T^\infty _N(n))_{n \in \mathbb {N}}\), except for the fact that \(T'_N(1)\) is the first time when a deep trap is visited by the walk (instead of being simply discovered). More precisely, we let \(\tilde{T}'_N(0) = 0\),
and then define recursively, for any \(n \in \mathbb {N}\),
We let \(x'_N(n)\) be the deep trap visited at time \(T'_N(n),\,\tau '_N(n) = \theta _{x'_N(n)} \ \tau _N\), and \(e'_N(n)\) be such that
As follows from Proposition 9.4, the family of triples
is a family of i.i.d. triples under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\). Moreover, the pair \((T'_N(n)-\tilde{T}'_N(n-1),\tau _{N,x'_N(n)})\) is independent of \(e'_N(n)\) under \(\mathbf{P}^{\tau _N}_{\mathfrak {u}}\). Indeed, conditionally on the history of the walk up to time \(T'_N(n)\), the random variable \(e'_N(n)\) is simply an exponential random variable of parameter \(1\) (the argument is the same as the one in the paragraph below (9.27)). Note that from this argument, we also learn what is the distribution of \(e'_N(n)\).
For definiteness, let us look at the Fourier transform of the random variables under consideration. By the above remarks, for every \(\tau _N\), every \(k \in \mathbb {N}\) and every \(x_n, y_n,z_n \in \mathbb {R}\), we have
We now argue that as \(N\) tends to infinity, the difference between
and
tends to \(0\) as \(N\) tends to infinity. First, we have already seen several times that the difference between \(\tilde{T}'_N(n-1)\) and \(T'_N(n-1)\) is subexponential (to be precise, it is smaller than \(N^3\) with probability tending to \(1\)). So in (12.10), if we replace \(\tilde{T}'_N(n-1)\) by \(T'_N(n-1)\), then the error made tends to \(0\) as \(N\) tends to infinity.
The first deep trap discovered is \(x^{(\delta )}_N(1)\) and at time \(T^{(\delta )}_N(1)\). On the event \({\mathcal {F}}_N(k)\), we know that \(x^{(\delta )}_N(1)\) is actually visited during the time interval \([T^{(\delta )}_N(1), T^{(\delta )}_N(1)+N]\), and that no other deep trap is discovered during this time interval, so \(x^{(\delta )}_N(1) = x'_N(1)\). On this event, the definitions of \(e'_N(1)\) and \(e^{(\delta )}_N(1)\) actually coincide. Note also that on this event, we have \(T^{(\delta )}_N(1) \leqslant T'_N(1) \leqslant T^{(\delta )}_N(1)+N\). Proceeding as in the proof of part (2) of Proposition 12.2, we can iterate this reasoning to subsequent sites, and obtain that indeed the difference between (12.10) and (12.11) converges to \(0\) as \(N\) tends to infinity.
In order to conclude, we want to take the \(\mathbb {E}\)-expectation in (12.9) and identify the right-hand side as the “correct answer”. Note that the limit of
as \(N\) tends to infinity is known and given by Proposition 7.1. By the fact that we can replace \(x^{(\delta )}_N(n)\) by \(x'_N(n)\) above, we obtain explicitly the limit of the \(\mathbb {E}\)-expectation of the right-hand side of (12.9), and this is also the limit of
so the proof is complete. \(\square \)
13 Convergence of the clock process
In this section, we obtain the scaling limit of the clock process.
Theorem 13.1
Recall that we assume (2.2) with \(\,\overline{a} < 1/20\), that \(({\mathsf {t}}_N)\) satisfies \(\log {\mathsf {t}}_N\sim \overline{c} N\) with \(\overline{c} \in (0,\log (2))\), that \({\mathsf {d}}_N\) is given by Theorem 5.1 (and satisfies (11.8)), that \(\phi ,\,B_N\) and \(\alpha \) are defined in (2.5), (6.3) and (6.5) respectively, and that we assume \(\alpha < 1\). Let
be the clock process, and let \(C_N\) be its rescaled version, defined by
Under \(\overline{\mathbb {P}}_N\), the rescaled clock process \(C_N\) converges in distribution, for the \(M_1\) topology and as \(N\) tends to infinity, to the \(\alpha \)-stable subordinator \({\mathcal {C}}\) whose Lévy measure is given by
where \(\Gamma \) is Euler’s Gamma function.
Remark 13.2
The \(M_1\) topology is defined on the space of càd-làg functions, we refer to [41, (3.3.4)] for a precise definition. Here, we understand the convergence of processes defined on \(\mathbb {R}_+\) as the convergence of the restriction of the processes on \([0,t]\) for every \(t > 0\).
Remark 13.3
In Theorem 13.1, as in the next Theorem 14.1 on the convergence of the age process, we impose two restrictions on the time scale, namely \(\overline{c}\in (0,\log (2))\) and \(\overline{c}<\beta ^2/2\), this last inequality being equivalent to the condition \(\alpha <1\).
We note that we thus cover all possible time scales. Indeed, the convergence of the clock process towards a subordinator can only hold on time scales that are shorter than the equilibrium time. Let us see that this corresponds to the restriction \(\overline{c} < \log (2) \wedge \beta ^2/2\).
Remember that we are considering a time \({\mathsf {t}}_N\) such that \(\log ({\mathsf {t}}_N) \sim \overline{c} N\). This corresponds to looking at the process \(Z\) up to times of order \(B_N\) (on the exponential scale), with \(\log (B_N)\sim \beta \sqrt{2\overline{c}}N\).
On the one hand, by general facts on reversible Markov chains, we know that the process \(Z\) has reached equilibrium for times larger than the inverse spectral gap. As can be seen by adapting the arguments in [24], the inverse spectral gap of the dynamics we are considering here is of order \(\exp (\beta \beta _c N)\), where \(\beta _c=\sqrt{2\log (2)}\). Imposing that \(B_N\) is smaller than \(\exp (\beta \beta _c N)\) on the exponential scale leads to the condition \(\overline{c}<\log (2)\).
It turns out that the time needed for the dynamics to reach equilibrium when started with the uniform law (as here) may be much shorter than the inverse spectral gap in the high-temperature regime, and should then rather be estimated using generalized Poincaré inequalities. Results in [35] show that if \(\beta >\beta _c\), then \(Z\) is indeed close to equilibrium by a time of order \(\exp (\beta ^2N)\). (The dynamics considered in [35] is rather the RHT dynamics, but the argument readily generalizes to the dynamics considered here as soon as we assume \(a_N=o(\sqrt{N})\), and in particular under our present assumption (2.2).) Requiring that, on the exponential scale, \(B_N\) is smaller than \(\exp {\beta ^2N}\) is equivalent to assuming that \(\overline{c}<\beta ^2/2\).
Let
Proposition 13.4
For every \(T > 0\), the \(M_1\) distance on between the processes \(C^{(\delta )}_N\) and \({\mathcal {C}}^{(\delta )}_N\) restricted to \([0,T]\) converges to \(0\) in \(\overline{\mathbb {P}}_N\)-probability.
Proof
The proof is similar to that of [37, Proposition 6.3]. Let \(\tilde{B}_N^{-1}\) be such that
Let \(T > 0\). We begin by showing that for every \(n \leqslant k\), the \(M_1\) distance between
and
as processes defined on \([0,T]\), tends to \(0\) in \(\overline{\mathbb {P}}_N\)-probability. These two processes are increasing. The second process is equal to \(0\) until time \({\mathsf {t}}_N^{-1} T^{(\delta )}_N(n)\), when it jumps to the value
(recall that \(H^{(\delta )}_N(n)\) is the first time the walk visits \(x^{(\delta )}_N(n)\)). We know from Proposition 12.2 that with probability close to \(1\), the site \(x^{(\delta )}_N(n)\) is not visited by the walk during the time interval \([0,T^{(\delta )}_N(k+1)] {\setminus } [T^{(\delta )}_N(n), T^{(\delta )}_N(n)+N]\). We know also from Proposition 9.11 that with probability tending to \(1,\,x^{(\delta )}_N(n)\) is indeed visited by the walk during the time interval \([T^{(\delta )}_N(n),T^{(\delta )}_N(n) + N]\). As a consequence, with probability tending to \(1\), the process in (13.4) is \(0\) until time \({\mathsf {t}}_N^{-1} T^{(\delta )}_N(n)\), and then is constant equal to
for all times greater than \({\mathsf {t}}_N^{-1} (T^{(\delta )}_N(n)+N)\).
In words, we have argued that the two increasing processes in (13.4) and (13.5) are with high probability equal at every time, except possibly during a time interval whose length \({\mathsf {t}}_N^{-1} N\) tends to \(0\) as \(N\) tends to infinity. Moreover, by Proposition 12.5, their final value is bounded in probability. These observations ensure that we can construct a parametrization of the completed graphs of these processes, as defined in [41, (3.3.3)], showing that the \(M_1\) distance between the two processes tends to \(0\), provided we can guarantee that the interval where the two processes possibly differ is far from the endpoints of the interval \([0,T]\). But this is so with high probability, since we know that the limiting distribution of \({\mathsf {t}}_N^{-1} \ T^{(\delta )}_N(n)\) has a density, see Proposition 12.5.
From this observation, we obtain that the \(M_1\) distance between
and
(as processes defined on \([0,T]\)) converges to \(0\) in \(\overline{\mathbb {P}}_N\)-probability. Proposition 12.2 ensures that with probability tending to \(1\), no other deep trap than \((x^{(\delta )}_N(n))_{n \leqslant k}\) is visited up to time \(T^{(\delta )}_N(k+1)\). Hence, we obtain that the \(M_1\) distance between
and (13.6) converges to \(0\) as well. Now, in view of Proposition 12.5, this is sufficient, since by choosing \(k\) very large, we can make the probability that \({\mathsf {t}}_N^{-1} \ T^{(\delta )}_N(k+1) \leqslant T\) as close to \(0\) as we wish. \(\square \)
Proposition 13.5
Under \(\overline{\mathbb {P}}_N\), the process \(C^{(\delta )}_N\) converges in distribution, for the \(M_1\) topology and as \(N\) tends to infinity, to the process \({\mathcal {C}}^{(\delta )}\) defined by
where the random variables
are distributed as in Proposition 12.5.
Proof
By Proposition 13.4, it suffices to study the convergence of \({\mathcal {C}}^{(\delta )}_N\). The proof is then identical to the proof of [37, Proposition 7.5], so we only briefly mention the argument. We use Skorokhod’s representation theorem, so that we can assume that the convergence described in Proposition 12.5 is actually an almost sure convergence. With this in hand, the convergence of the (Skorokhod representation of the) process \({\mathcal {C}}^{(\delta )}_N\) to the (Skorokhod representation of the) process \({\mathcal {C}}^{(\delta )}\) for the \(M_1\) topology easily follows. \(\square \)
Proof of Theorem 13.1
Let \(T > 0\). We begin by showing that
The process \(C^{(\delta )}_N-C_N\) is increasing, and thus
where we recall that \(\tilde{B}_N\) was defined in (13.3). Note that the integrand above is simply a function of the environment viewed by the particle, which is stationary under \(\overline{\mathbb {P}}_N\). The right-hand side of (13.8) is thus equal to
Finally, note that by Proposition 11.3, we have
Identity (13.7) then follows from (6.6).
We want to complete the following diagram
where each arrow represents convergence in distribution for the \(M_1\) topology. Since the uniform norm controls the \(M_1\) distance, we have proved with (13.7) that the convergence of \(C^{(\delta )}_N\) to \(C_N\) holds uniformly in \(\delta \). Proposition 13.5 ensures that \(C^{(\delta )}_N\) converges indeed to \({\mathcal {C}}^{(\delta )}\). In order to deduce the convergence of \(C_N\) to \({\mathcal {C}}\) (and thus prove Theorem 13.1), what remains to be done (see [14, Theorem 4.2]) is thus simply to show that \({\mathcal {C}}^{(\delta )}\xrightarrow [\delta \rightarrow 0]{} {\mathcal {C}}\).
Let us thus proceed to prove this convergence. It is clear from its definition that \({\mathcal {C}}^{(\delta )}\) is a subordinator. Let \(\psi _\delta (\lambda )\) be the Laplace exponent of \({\mathcal {C}}^{(\delta )}\), defined by
Using the definition of \({\mathcal {C}}^{(\delta )}(t)\), we obtain that
and thus
Given the definition of \((e^{(\delta )}(1),\tau ^{(\delta )}(1))\) (see Proposition 12.5), we can compute this last quantity. It is equal to
The \(\delta ^{\alpha }\) cancel out. We make the change of variables \(u = yz\) to obtain
By the monotone convergence theorem, we thus obtain that \(\psi _\delta (\lambda )\) converges to
and this is indeed the Laplace exponent of \({\mathcal {C}}\). We have thus shown that for every \(t\geqslant 0\), the Laplace transform of \({\mathcal {C}}^{(\delta )}(t)\) converges to the Laplace transform of \({\mathcal {C}}(t)\). Since both processes are subordinators, this ensures convergence in the sense of finite-dimensional distributions. Finally, for increasing processes, convergence of finite-dimensional distributions imply convergence for the \(M_1\) topology (this can be seen directly from the definition, or using the tightness criterion for increasing processes given in [37, Lemma 8.4]; in our case, we could also easily show that the difference between \({\mathcal {C}}\) and \({\mathcal {C}}^{(\delta )}\) converges to \(0\) in probability for the supremum norm). This finishes the proof. \(\square \)
14 Convergence of the age process
We now discuss the convergence of the age process. In the next theorem as in the next proofs, we use the following definitions. Let \(f\) be a càd-làg function. Its inverse is defined as
Let \(f\) be a continuous non-decreasing function. Then \(f'(t)\) denotes the right derivative of \(f\) at point \(t\) (provided it exists).
Theorem 14.1
Recall the definitions and assumptions summarized in Theorem 13.1. Let
be the age process, and let \(A_N\) be its rescaled version, defined by
Under \(\overline{\mathbb {P}}_N\), the rescaled age process \(A_N\) converges in distribution, for the \(L_{1,loc}\) topology and as \(N\) tends to infinity.
The limiting process, say \({\mathcal {Z}}\), may be constructed as follows: let \(\Upsilon \) be an \(\alpha \)-stable subordinator, and let
where \((T_t)_{t\geqslant 0}\) is an independent family of independent, mean \(1\), exponential random variables. Let \({\mathcal {W}}={\mathcal {V}}^{-1}\) be the inverse of \(V\). Finally, define the process
In the sequel, we use the notation \(\mu (x)=\Upsilon (x)-\Upsilon (x-)\). Note that \({\mathcal Z}\) is a self-similar process of index \(1\). One checks that its law does not depend on the choice of the \(\alpha \)-stable subordinator \(\Upsilon \).
The following proof follows a similar strategy as in [28]. Recall that the processes \(X\) and \(Z\) are connected through
where \(C\) is the clock process
Observe that
As in the proof of Theorem 13.1, we consider the rescaled version of the clock process
(Recall that \(\tilde{B}_N\) was defined in (13.3)). Note that
and thus
Also, as in the proof of Theorem 13.1, we introduce truncated processes :
and
and
The proof of the next proposition is similar to the proof of Proposition 13.4.
Proposition 14.2
For every \(T > 0\), the \(M_1\) distance between the processes \(A^{(\delta )}_N\) and \({\mathcal {A}}^{(\delta )}_N\) restricted to \([0,T]\) converges to \(0\) in \(\overline{\mathbb {P}}_N\)-probability.
Proof
Let \(k\geqslant 1\). Observe that on the event \({\mathcal {F}}^{(\delta )}_N(k)\), the two processes \(A^{(\delta )}_N\) and \({\mathcal {A}}^{(\delta )}_N\) take the same values (namely the sequence \(B_N^{-1}\tau _{N,x^{(\delta )}_N(n)}\), for \(n\in \{1,\ldots ,k\}\)) in the same order. The holding time at a trap \(x^{(\delta )}_N(n)\) is then \(C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(n+1))-C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(n))\) for the process \(A^{(\delta )}_N\), and \({\mathcal {C}}^{(\delta )}_N({\mathsf {t}}_N^{-1}T^{(\delta )}_N(n+1)-)-{\mathcal {C}}^{(\delta )}_N(T^{(\delta )}_N(n)-) =\tilde{B}_N^{-1} \ e^{(\delta )}_N(n) \ G_N(\tau ^{(\delta )}_N(n)) \tau _{N,x^{(\delta )}_N(n)}\) for the process \({\mathcal {A}}^{(\delta )}_N\). On \({\mathcal {F}}^{(\delta )}_N(k)\), these two quantities coincide. Therefore the two processes \(A^{(\delta )}_N\) and \({\mathcal {A}}^{(\delta )}_N\) coincide on \({\mathcal {F}}^{(\delta )}_N(k)\) until time \(C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(k+1))\).
Proposition 12.2 implies that the \(\overline{\mathbb {P}}_N\)-probability of \({\mathcal {F}}^{(\delta )}_N(k)\) tends to \(1\). Therefore we also have that the \(M_1\) distance between the processes \((A^{(\delta )}_N(t))_{t\leqslant C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(k+1)) }\) and \(({\mathcal {A}}^{(\delta )}_N(t))_{t\leqslant C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(k+1))}\) tends to \(0\).
Finally we argue that, by choosing \(k\) large, we can make the probability that \(C^{(\delta )}_N({\mathsf {t}}_N^{-1}H^{(\delta )}_N(k+1))\leqslant T\) as small as we want. On the one hand, the process \(C^{(\delta )}_N\) has a limit in distribution, \({\mathcal {C}}^{(\delta )}\), which is described in Proposition 13.5 and satisfies \(\lim _{t\rightarrow +\infty }{\mathcal {C}}^{(\delta )}(t)=+\infty \). On the other hand, we observe that \({\mathsf {t}}_N^{-1}H^{(\delta )}_N(k+1)\) is larger than \({\mathsf {t}}_N^{-1}T^{(\delta )}_N(k+1)\) and that the limit in distribution of \({\mathsf {t}}_N^{-1}T^{(\delta )}_N(k+1)\), which we wrote \(T^{(\delta )}(k+1)\), was obtained in Proposition 12.5 and satisfies \(\lim _{k\rightarrow +\infty } T^{(\delta )}(k+1)=+\infty \). \(\square \)
Next, we introduce truncated versions of the process \({\mathcal {Z}}\): enumerate the set \(\{x\ge 0:\, \mu (x)>\delta \}\) as
and let
where an empty sum equals \(0\). We consider the inverse of \({\mathcal {V}}^{(\delta )}\), say \({\mathcal {W}}^{(\delta )}\), and let
Remark 14.3
The trajectories of the process \({\mathcal {Z}}^{(\delta )}\) are easy to describe: \({\mathcal {Z}}^{(\delta )}\) successively takes the values \(\mu (\upsilon _1), \mu (\upsilon _2), \ldots \) with respective holding times \(\mu (\upsilon _1)T_{\upsilon _1},\,\mu (\upsilon _2)T_{\upsilon _2}, \ldots \)
Proposition 14.4
Under \(\overline{\mathbb {P}}_N\), the process \(A^{(\delta )}_N\) converges in distribution, for the \(M_1\) topology and as \(N\) tends to infinity, to the process \({\mathcal {Z}}^{(\delta )}\).
Proof
By Proposition 14.2, it suffices to study the convergence of \({\mathcal {A}}^{(\delta )}_N\). We use Skorokhod’s representation theorem, so that we can assume that the convergence described in Proposition 12.5 is actually an almost sure convergence. With this in hand, the convergence of the (Skorokhod representation of the) process \({\mathcal {A}}^{(\delta )}_N\) to the (Skorokhod representation of the) process \({\mathcal {Z}}^{(\delta )}\) for the \(M_1\) topology easily follows: compare the description of the process \({\mathcal {A}}^{(\delta )}_N\) in the proof of Proposition 14.2 with the description of \({\mathcal {Z}}^{(\delta )}\) in Remark 14.3. \(\square \)
Proof of Theorem 14.1
Recall that the space of trajectories is equipped with the \(L_{1,loc}\) topology. Note that the \(M_1\) is stronger so that Proposition 14.4 also holds for the \(L_{1,loc}\) topology.
Let \(T>0\). Let us first show that, for any \(\eta > 0\),
It will be convenient to consider the following process: let \({\bar{A}}^{(\delta )}_N(t)\) be given by \(B_N^{-1}\tau _{N,x}\), where \(x\) is the last deep trap visited by the process \(X({\mathsf {t}}_NC_N^{-1}(.))\) before time \(t\), with the convention \({\bar{A}}^{(\delta )}_N(t)=0\) if no deep trap has been visited by time \(t\).
On the one hand, we observe that both processes \(A^{(\delta )}_N\) and \({\bar{A}}^{(\delta )}_N\) take the same values in the same order (namely the sequence of values of the successive deep traps visited by \(X\)). The holding times differ: for \(A^{(\delta )}_N\), it is given by the occupation time of a deep trap; for \({\bar{A}}^{(\delta )}_N\) it is given by the increment of the clock process \(C_N\) between visits of deep traps. Thus we see that
where the \(\sup \) is computed on the set of deep traps that have been visited by \(X\) before time \({\mathsf {t}}_NC_N^{-1}(T)\) and \(K^{(\delta )}_N(T)\) is the number of times \({\bar{A}}^{(\delta )}_N(t)\) changes its value before time \({\mathsf {t}}_NC_N^{-1}(T)\).
(Let \(f\) and \(g\) be two functions defined on \([0,T]\) with values in a finite subset of \(\mathbb {R}_+\). Assume both \(f\) and \(g\) are piecewise constant, taking the values \(x_1,\ldots ,x_K\) in that order and with respective sojourn times \(w_1,\ldots ,w_K\) for \(f\) and \(\zeta _1,\ldots ,\zeta _K\) for \(g\). Assume \(\zeta _i\geqslant w_i\) for all \(i\). Then \(f\) and \(g\) coincide except possibly on a set whose Lebesgue measure is smaller than \((\zeta _1-w_1)+(\zeta _2+\zeta _1-w_2+w_1)+\cdots \leqslant K\sum _i(\zeta _i-w_i)\). Therefore the \(L_1\) distance between \(f\) and \(g\) is bounded by \(2\sup _i\vert x_i\vert K\) \(\sum _i(\zeta _i-w_i)\).)
Both \(B_N^{-1} \sup \tau _{N,x}\) and \(K^{(\delta )}_N(T)\) are of order \(1\). To justify this last claim, we let \(L^{(\delta )}_N(T)\) be the number of deep traps visited until time \({\mathsf {t}}_NC_N^{-1}(T)\). On the one hand, we already know that \(L^{(\delta )}_N(T)\) is of order \(1\). On the other hand \(K^{(\delta )}_N(T) \mathbf{1}_{L^{(\delta )}_N(T)\leqslant k}\mathbf{1}_{{\mathcal {F}}^{(\delta )}_N(k)}=L^{(\delta )}_N(T) \mathbf{1}_{L^{(\delta )}_N(T)\leqslant k}\mathbf{1}_{{\mathcal {F}}^{(\delta )}_N(k)}\) for all \(k\). As the probability of \({\mathcal {F}}^{(\delta )}_N(k)\) tends to \(1\), we deduce that \(K^{(\delta )}_N(T)\) is of order \(1\) in probability.
As we already saw in the proof of (13.7), the expected value of
tends to \(0\). Thus we conclude that
Let us now consider the difference \(A_N-{\bar{A}}^{(\delta )}_N\): these processes do not necessarily take the same values, but they coincide when the process \(X\) is visiting a deep trap, i.e. \(A_N(t)={\bar{A}}_N(t)\) whenever \(A_N(t)\geqslant \delta \). Therefore
Combining this with (14.4), we get (14.3).
To conclude the proof of Theorem 14.1, it then only remains to show that \({\mathcal {Z}}^{(\delta )}\) converges to \({\mathcal Z}\) as \(\delta \) tends to \(0\). This is elementary.\(\square \)
References
Aldous, D., Fill, J. : Reversible Markov chains and random walks on graphs. Online version available at www.stat.berkeley.edu/aldous/RWG/book.html (in preparation)
Aldous, D., Diaconis, P.: Shuffling cards and stopping times. Am. Math. Mon. 93(5), 333–348 (1986)
Aldous, D., Diaconis, P.: Strong uniform times and finite random walks. Adv. Appl. Math. 8(1), 69–97 (1987)
Barlow, M.T., Černý, J.: Convergence to fractional kinetics for random walks associated with unbounded conductances. Probab. Theory Relat. Fields 149(3–4), 639–673 (2011) [Erratum. Probab. Theory Relat. Fields 149(3–4), 675–677 (2011)]
Ben Arous, G., Bogachev, L.V., Molchanov, S.A.: Limit theorems for sums of random exponentials. Probab. Theory Relat. Fields 132(4), 579–612 (2005)
Ben Arous, G., Bovier, A., Černý, J.: Universality of random energy model-like ageing in mean field spin glasses. J. Stat. Mech. Theory Exp. (4), L04003 (2008)
Ben Arous, G., Bovier, A., Černý, J.: Universality of the REM for dynamics of mean-field spin glasses. Commun. Math. Phys. 282(3), 663–695 (2008)
Ben Arous, G., Bovier, A., Gayrard, V.: Glauber dynamics of the random energy model. I. Metastable motion on the extreme states. Commun. Math. Phys. 235(3), 379–425 (2003)
Ben Arous, G., Bovier, A., Gayrard, V.: Glauber dynamics of the random energy model. II. Aging below the critical temperature. Commun. Math. Phys. 236(1), 1–54 (2003)
Ben Arous, G., Černý, J.: Dynamics of trap models. Mathematical statistical physics. Papers from the 83rd session of the summer school in physics, Les Houches, pp. 331–394 (2006)
Ben Arous, G., Černý, J.: The arcsine law as a universal aging scheme for trap models. Commun. Pure Appl. Math. 61(3), 289–329 (2008)
Ben Arous, G., Gün, O.: Universality and extremal aging for dynamics of spin glasses on subexponential time scales. Commun. Pure Appl. Math. 65(1), 77–127 (2012)
Bezerra, S.C., Fontes, L.R.G., Gava, R.J., Gayrard, V., Mathieu, P.: Scaling limits and aging for asymmetric trap models on the complete graph and K processes. ALEA Lat. Am. J Probab. Math. Stat. 9, 303–321 (2012)
Billingsley, P.: Convergence of Probability Measures, Wiley Series in Probability and Statistics, 2nd edn. Wiley, New York (1968)
Bouchaud, J.-P.: Weak ergodicity breaking and aging in disordered systems. J. Phys. I (France) 2, 1705–1713 (1992)
Bouchaud, J.-P., Dean, D.S.: Aging on Parisi’s tree. J. Phys. I (France) 5, 265–286 (1995)
Bovier, A., Faggionato, A.: Spectral characterization of aging: the REM-like trap model. Ann. Appl. Probab. 15(3), 1997–2037 (2005)
Bovier, A., Gayrard, V.: Convergence of clock processes in random environments and ageing in the p-spin SK model. Ann. Probab. 41(2), 817–847 (2013)
Bovier, A., Gayrard, V., Švejda, A.: Convergence to extremal processes in random environments and extremal ageing in SK models. Probab. Theory Relat. Fields 157(1–2), 251–283 (2013)
Černý, J.: On two-dimensional random walk among heavy-tailed conductances. Electron. J. Probab. 16(10), 293–313 (2011)
De Masi, A., Ferrari, P.A., Goldstein, S., Wick, W.D.: An invariance principle for reversible Markov processes. Applications to random motions in random environments. J. Stat. Phys. 55(3–4), 787–855 (1989)
Derrida, B.: Random-energy model: limit of a family of disordered models. Phys. Rev. Lett. 45, 79–82 (1980)
Derrida, B.: Random-energy model: an exactly solvable model of disordered systems. Phys. Rev. B 24, 2613–2626 (1981)
Fontes, L.R.G., Isopi, M., Kohayakawa, Y., Picco, P.: The spectral gap of the REM under Metropolis dynamics. Ann. Appl. Probab. 8(3), 917–943 (1998)
Fontes, L.R.G., Lima, P.H.S.: Convergence of symmetric trap models in the hypercube. New trends in mathematical physics, pp. 285–297 (2009)
Fontes, L.R.G., Mathieu, P.: On symmetric random walks with random conductances on \({\mathbb{Z}}^d\). Probab. Theory Relat. Fields 134, 565–602 (2006)
Fontes, L.R.G., Mathieu, P.: \(K\)-processes, scaling limit and aging for the trap model in the complete graph. Ann. Probab. 36(4), 1322–1358 (2008)
Fontes, L.R.G., Mathieu, P.: On the dynamics of trap models in \({\mathbb{Z}}^d\). Proc. Lond. Math. Soc. (to appear)
Galves, A., Martínez, S., Picco, P.: Fluctuations in Derrida’s random energy and generalized random energy models. J. Stat. Phys. 54(1–2), 515–529 (1989)
Gayrard, V.: Aging in reversible dynamics of disordered systems. II. Emergence of the arcsine law in the random hopping time dynamics of the REM. arXiv:1008.3849v1 (2010, preprint)
Gayrard, V.: Convergence of clock process in random environments and aging in Bouchaud’s asymmetric trap model on the complete graph. Electron. J. Probab. 17(58), 33 (2012)
Gayrard, V., Švejda, A.: Convergence of clock processes on infinite graphs and aging in Bouchaud’s asymmetric trap model on \({\mathbb{Z}}^d\). arXiv:1309.3066 (2013, preprint)
Guionnet, A., Zegarlinski, B.: Lectures on logarithmic Sobolev inequalities. Séminaire de Probabilités XXXVI, Lecture Notes in Mathematics, vol. 1801, pp. 1–134. Springer, Berlin (2003)
Jara, M., Landim, C., Teixeira, A.: Universality of trap models in the ergodic time scale. Ann. Probab. (to appear)
Mathieu, P.: Convergence to equilibrium for spin glasses. Commun. Math. Phys. 215(1), 57–68 (2000)
Mathieu, P., Picco, P.: Convergence to equilibrium for finite Markov processes, with application to the Random Energy Model. Unpublished manuscript (2003). arXiv:0307148v1
Mourrat, J.-C.: Scaling limit of the random walk among random traps on \({\mathbb{Z}}^d\). Ann. Inst. Henri Poincaré Probab. Stat. 47(3), 813–849 (2011)
Olivieri, E., Picco, P.: On the existence of thermodynamics for the random energy model. Commun. Math. Phys. 96(1), 125–144 (1984)
Saloff-Coste, L.: Lectures on Finite Markov Chains. Lectures on Probability Theory and Statistics (Saint-Flour 1996), pp. 301–413. Springer, Berlin (1997)
Steele, J.: An Efron–Stein inequality for non symmetric statistics. Ann. Stat. 14(2), 753–758 (1986)
Whitt, W.: Stochastic-Process Limits, Springer Series in Operations Research. Springer, Berlin (2002)
Acknowledgments
We wish to thank G. Ben Arous for insightful discussions at an early stage of this paper, and for encouraging us to complete this work. P.M. acknowledges his kind hospitality at the Courant Institute.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Mathieu, P., Mourrat, JC. Aging of asymmetric dynamics on the random energy model. Probab. Theory Relat. Fields 161, 351–427 (2015). https://doi.org/10.1007/s00440-014-0551-2
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00440-014-0551-2