1 Introduction
The (1+1) dimensional Kardar-Parisi-Zhang (KPZ) equation [Reference Kardar, Parisi and ZhangKPZ86] is a stochastic partial differential equation (SPDE) given by
where
$\gamma>0$
and
$\xi (t,x)$
is a standard Gaussian space-time white noise on
${\mathbb {R}}_+\times {\mathbb {R}}$
. As an SPDE, (KPZ) is ill-posed due to the presence of the nonlinear term
$(\partial _xH)^2$
. One way to make sense of the equation is to consider
$\mathcal {U}:= e^H$
which formally solves the stochastic heat equation (SHE) with multiplicative noise:
The SHE is known to be well-posed and has a well-developed solution theory based on the Itô integral and chaos expansions [Reference WalshWal86, Reference Bertini and CancriniBC95, Reference QuastelQua11, Reference CorwinCor12]. In this paper, we will consider the solution to (SHE) started with Dirac delta initial data
$\mathcal {U}(0,x) = \delta _0(x)$
. For this initial data, [Reference Moreno FloresFlo14] established that
$\mathcal {U}(t,x)> 0$
for all
$t> 0$
and
$x \in \mathbb {R}$
almost surely (see also [Reference MuellerMue91]). Thus
$H=\log \mathcal {U}$
is well-defined and is called the Hopf-Cole solution of the KPZ equation. This is the notion of solution that we will work with in this paper, and it coincides with other existing notions of solutions using regularity structures [Reference HairerHai13, Reference HairerHai14], energy solutions [Reference Gubinelli and JaraGJ13, Reference Gonçalves and JaraGJ14, Reference Gubinelli and PerkowskiGP18], paracontrolled products [Reference Gubinelli, Imkeller and PerkowskiGIP15, Reference Gubinelli and PerkowskiGP17, Reference Perkowski and Cornelis RosatiPR19], and the Polchinski flow [Reference Kupiainen and MarcozziKM17, Reference DuchDuc25, Reference Chandra and FerdinandCF24].
The KPZ equation has been studied intensively over the past three decades in both the mathematics and physics literature. It first gained attention in the physics community in [Reference Kardar, Parisi and ZhangKPZ86], and since then a precise mathematical understanding of the equation and its well-posedness theory has gradually evolved. In the mathematical physics literature, the relevance of the equation is that it appears as the fluctuation of a number of physically complex systems such as interacting particle systems [Reference Bertini and GiacominBG97] and directed polymers [Reference Alberts, Khanin and QuastelAKQ14]. We refer to [Reference Ferrari and SpohnFS10, Reference QuastelQua11, Reference CorwinCor12, Reference Quastel and SpohnQS15, Reference Chandra and WeberCW17, Reference Corwin and ShenCS20] for some surveys of the mathematical studies of the KPZ equation, and some of the models in which it arises.
The present work will be concerned with how the KPZ equation arises in a family of probabilistic models called stochastic flows of kernels. Roughly speaking, a discrete-time stochastic flow of kernels is a family of random probability measures indexed by each site of space-time (
$\mathbb N\times {\mathbb {Z}}$
or
$\mathbb N\times {\mathbb {R}}$
depending on the desired model), where spatial correlations between the randomized transitions are allowed but temporal correlations are not. The notion of stochastic flows initially gained importance for their relation to geometry and PDEs with random coefficients, see the monograph [Reference KunitaKun97]. A resurgence of interest and a broadening of the scope of stochastic flow theory was introduced by work of Tsirelson-Vershik on continuous products of probability spaces [Reference Tsirelson and VershikTV98, Reference TsirelsonTsi04], and further work of Le Jan-Raimond on their relation to projective families of Markov chains [Reference Le Jan and RaimondLJR02, Reference Le Jan and RaimondLJR04b, Reference Le Jan and RaimondLJR04a].
In a recent work [Reference Das, Drillick and ParekhDDP24b], we were able to show that (SHE) arises as a continuum limit in the following discrete stochastic flow model. Consider a family of IID
$[0,1]$
-valued random variables
$\omega _{r,y}$
with
$r \in \mathbb {Z}_{\ge 0}$
and
$y\in {\mathbb {Z}}$
, drawn from a common law
$\nu $
. A random walk
$(R(r))_{r\ge 0}$
in the environment
$\omega $
starts from
$(r,y)=(0,0)$
and at each time step goes from
$(r,y)\to (r+1,y+1)$
with probability
$\omega _{r,y}$
and from
$(r,y)\to (r+1,y-1)$
with probability
$1-\omega _{r,y}$
. Consider the (quenched) random transition probabilities
A series of physics papers [Reference Thiery and Le DoussalTLD16, Reference Le Doussal and ThieryLDT17, Reference Barraquand and Le DoussalBLD20] conjectured that the fluctuations of the random field
$P^\omega (r,y)$
should be described by the KPZ equation at some point within the tail of that probability measure at spatial distance of order
$r^{3/4}$
, when r is very large. We thus define for
$t\ge 0$
and
$x\in {\mathbb {R}}$
and
$N\in \mathbb N$
the random field
The main results of [Reference Das, Drillick and ParekhDDP24b] proved the conjecture of [Reference Thiery and Le DoussalTLD16, Reference Le Doussal and ThieryLDT17, Reference Barraquand and Le DoussalBLD20], by showing that the random field
$\mathscr U_N(t,x)$
converges in law (in some topology of tempered distributions) to the solution of (SHE) with
$\gamma ^2= \frac {8\sigma ^2}{1-4\sigma ^2}$
, where
$\sigma ^2 = \mathrm {Var}(\omega _{t,x})$
. It turns out that the spatial distance of order
$N^{3/4}$
is the unique location within the tail of
$P^\omega (Nt,\cdot )$
where one expects to see (SHE) appear: at smaller exponents one observes a Gaussian limit and at larger exponents one sees Tracy-Widom limits (see Section 1.2 below). This established a “KPZ crossover” at spatial location
$N^{3/4}$
when time is of order N, with a somewhat unexpected value for the noise coefficient in the limiting SPDE. Thus we say that the crossover exponent for this model is
$3/4$
.
A number of follow-up questions may then be asked, such as whether the methods of proving the convergence to (SHE) could be generalized to more complex models in which there is non-nearest-neighbor random walk in a random environment, or correlations between the transition kernels at each lattice site. Numerical work of J. Hass and coauthors in the physics community [Reference Hass, Carroll-Godfrey, Corwin and CorwinHCGCC23, Reference Hass, Corwin and CorwinHCC23, Reference Hass, Drillick, Corwin and CorwinHDCC24, Reference HassHas25] observed that there are non-nearest-neighbor models where one should observe KPZ fluctuations at spatial location
$N^{7/8}$
rather than
$N^{3/4}.$
More precisely, at each lattice site
$(r,y) \in {\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}}$
sample IID Uniform[0,1] random variables. Then impose that the random walker goes from
$(r,y) \to (r+1,y+1)$
with probability
$(1-U_{r,y})/2$
, from
$(r,y)\to (r+1,y)$
with probability
$U_{r,y}$
and from
$(r,y) \to (r+1,y-1)$
with probability
$(1-U_{r,y})/2$
. The physicists observed that imposing this “first-order symmetry” of the transition kernels changes the crossover exponent from
$3/4$
to
$7/8$
, but the KPZ equation is still expected to arise at the new exponent.
A completely different type of model posed to us by I. Corwin is the following. Let
$(\omega _{r,x})_{r\in \mathbb N, x\in {\mathbb {Z}}}$
be an IID collection of random variables. Let
$b: {\mathbb {Z}}\to [0,1]$
be any deterministic function, nonzero on at least two sites, such that
$\sum _{x\in {\mathbb {Z}}} b(x)=1$
. Define the kernels
$$ \begin{align*}K_r(x,y):= \frac{b(y-x) e^{\omega_{r,y}}}{\sum_{y'\in {\mathbb{Z}}} b(y'-x) e^{\omega_{r,y'}}}.\end{align*} $$
Consider the “random landscape model” of random walk in a random environment on
${\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}}$
in which one goes from site
$(r,x) \to (r+1,y)$
with probability
$K_r(x,y)$
. This model is substantially different from the random walk models considered above, since the transitions of nearby particles can be strongly correlated via depending on the same weights. A natural question is to investigate whether or not one also finds the KPZ equation at location of order
$N^{3/4}$
away from the expected displacement, if one probes the quenched transition density of the random walker at time of order N.
All of these questions then led to the desire to prove a “super-universality theorem” or “generalized invariance principle” for the equation (SHE) arising in stochastic flows of kernels, that is, a minimal set of conditions for determining when and where (SHE) is expected to arise in any family of stochastic kernels, and moreover to test the robustness of the methods developed in [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b]. This is exactly the subject of the present work. Our main result will be Theorem 1.5 below, where we find a collection of reasonably sharp conditions under which one expects to see (SHE) arise in a family of stochastic kernels. Moreover this theorem will give the precise location where (SHE) can be found (analogous to
$N^{3/4}$
above), as well as identify the noise coefficient of the limiting equation (analogous to
$\frac {8\sigma ^2}{1-4\sigma ^2}$
above).
1.1 Generalized model and main results
Definition 1.1. Let
$I={\mathbb {R}}$
or
$I=c {\mathbb {Z}}$
where
$c>0$
is fixed henceforth. Let
$\mathcal B(I)$
denote the Borel
$\sigma $
-algebra on I. A Markov kernel on I is a continuous function from
$I\to \mathcal P(I)$
, where
$\mathcal P(I)$
is the space of probability measures on
$(I,\mathcal B(I)),$
equipped with the topology of weak convergence of probability measures.
Any Markov kernel K on I is written as
$K(x,A)$
, which is identified with the continuous function from
$I\to \mathcal P(I)$
given by
$x\mapsto K(x,\cdot )$
. We denote by
$\mathcal M(I)$
the space of all Markov kernels on I.
We equip
$\mathcal M(I)$
with the weakest topology under which the maps
$T_{f,x}:\mathcal M(I)\to {\mathbb {R}}$
given by
are continuous, where one varies over all
$x\in I$
and all bounded continuous functions
$f:I\to {\mathbb {R}}$
. This topology endows
$\mathcal M(I)$
with a Borel
$\sigma $
-algebra which allows us to talk about random variables taking values in the space of Markov kernels. For
$a\in I$
define the translation operator
$\tau _a:\mathcal M(I)\to \mathcal M(I)$
by
$(\tau _a K)(x,A):= K(x+a,A+a)$
which is a Borel-measurable map on this space.
Assumption 1.2 (Assumptions for the main result)
Assume we have a family
$\{K_n\}_{n\ge 1}$
of random variables in the space
$\mathcal M(I)$
, defined on some probability space
$(\boldsymbol \Omega , \mathcal F^\omega , \mathbb P)$
such that the following hold true.
-
1. (Stochastic flow increments)
$K_1,K_2,K_3,...$
are independent and identically distributed under
$\mathbb P$
. -
2. (Spatial translational invariance)
$K_1$
has the same law as
$\tau _a K_1$
for all
$a\in I,$
under
$\mathbb P$
. -
3. (Exponential moments for the annealed law) Letting
$\mu (A) := {\mathbb {E}}[ K_1(0,A)],$
we have that the moment generating function exists, that is,
$\int _I e^{\eta |x|}\mu ({\mathrm {d}} x)<\infty $
for some
$\eta>0$
. Letting
$m_k:= \int _I x^k \mu ({\mathrm {d}} x)$
denote the sequence of moments of
$\mu $
, we also impose that the variance of
$\mu $
is 1, that is,
$m_2-m_1^2=1$
. -
4. (The first
$p-1$
moments are deterministic) Define the k-point correlation kernel There exists some
$$ \begin{align*}\boldsymbol p^{(k)}\big((x_1,...,x_k),({\mathrm{d}} y_1,...,{\mathrm{d}} y_k)\big):= {\mathbb{E}}[ K_1(x_1,{\mathrm{d}} y_1)\cdots K_1(x_k,{\mathrm{d}} y_k)].\end{align*} $$
$p\in \mathbb N$
such that
$\int _I y^kK_1(0,{\mathrm {d}} y)=m_k \mathbb P$
-almost surely for
$1\leq k \leq p-1$
, or equivalently Furthermore
$$ \begin{align*}\int_{I^2} (x-x_0)^k(y-y_0)^k \; \boldsymbol p^{(2)}((x_0,y_0),({\mathrm{d}} x,{\mathrm{d}} y))=m_k^2,\;\;\;\;\;\; x_0,y_0\in I, \;\;\;1\leq k \leq p-1.\end{align*} $$
$p-1$
is the largest such value, that is,
$\int _I y^{p}K_1(0,{\mathrm {d}} y)$
is nondeterministic.
-
5. (Strong decay of correlations up to
$(4p)^{th}$
joint moments) Consider a bounded function
$\mathrm {F_{decay}}:\Bbb [0,\infty )\to [0,\infty )$
such that
$x\mapsto x\mathrm {F_{decay}}(x)$
is decreasing for large x, and
$\int _0^\infty x \mathrm {F_{decay}}(x){\mathrm {d}} x<\infty $
. Then assume that the following spatial decay of correlations holds for the moments of the Markov kernel
$K_1$
: for all
$k\in \{2p,...,4p\}$
and all
$x_1,...,x_k\in I$
we have where
$$ \begin{align*}\max_{\substack{0\le r_1,...,r_k\le 4p-1\\2p\leq r_1+...+r_k\leq 4p}}\bigg| \int_{I^k} \prod_{j=1}^k (y_j-x_j)^{r_j} \boldsymbol p^{(k)} \big( {\mathbf{x}}, {\mathrm{d}}{\mathbf{y}}\big)- \int_{I^k} \prod_{j=1}^k y_j^{r_j}\mu^{\otimes k}({\mathrm{d}}{\mathbf{y}})\bigg| \leq \mathrm{F_{decay}} \big(\min_{1\le i<j\le k}|x_i-x_j|\big)\end{align*} $$
${\mathbf {x}}:= (x_1,...,x_k)$
and
${\mathbf {y}}=(y_1,...,y_k).$
-
6. (Nondegeneracy) Define the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}(x,A):=\int _I \mathbf {1}_{\{y_1-y_2\in A\}} \boldsymbol p^{(2)} \big ((x,0),({\mathrm {d}} y_1,{\mathrm {d}} y_2)\big )$
indexed by
$x\in I$
and
$A\in \mathcal B(I)$
. We impose:-
• (Regularity) The family of transition laws
$x\mapsto \boldsymbol {p}_{\mathbf {dif}}(x,\bullet )$
are continuous in total variation norm. -
• (Irreducibility) For any two points
$x,y \in I$
and any
$\varepsilon>0$
there exists
$m\in \mathbb N$
such that
${\boldsymbol {p}_{\mathbf {dif}}^m\big (x,(y-\varepsilon ,y+\varepsilon )\big )>0,}$
where
$\boldsymbol {p}_{\mathbf {dif}}^m$
is the
$m^{th}$
power of this Markov kernel.
-
Condition (4) says that the model exhibits “symmetry up to order p.” This integer p will play an extremely important role in all of the theorems and calculations throughout the rest of the paper.
To clarify an important detail about condition (1), we are not assuming that
$K_i(x,\cdot ),K_i(x',\cdot )$
are independent for distinct
$x,x'.$
Rather we are assuming that the family
$\{K_1(x,\cdot )\}_{x\in I}$
is independent of
$\{K_2(x,\cdot )\}_{x\in I}$
and has the same distribution as that family, and so on. But this says nothing about how
$K_i(x,\cdot ),K_i(x',\cdot )$
might be correlated for different
$x,x'$
. There are many interesting examples where there are indeed spatial correlations. This question of correlation of distinct spatial locations is covered by condition (5).
Indeed, condition (5) is the most important part of the assumption, whereas the other parts are more definitional. Notice that if
$r_1+...+r_k<2p$
then the left side in Item (5) vanishes identically thanks to Item (4), thus the maximum is really over all
$0\leq r_1+...+r_k \leq 4p$
. We remark that the finite-range models, in which
$\mathrm {F_{decay}}$
is compactly supported, already constitute many interesting examples. Additionally, note that all bounds in (5) are automatically satisfied if there exists
$\mathrm {C_{len}}>0$
so that for all
$x_1,...,x_{4p}\in I$
, one has the total variation bound
Here the first measure denotes the translate of
$\mu ^{\otimes k}$
by the vector
$(x_1,...,x_k)$
. Indeed if (1.1) holds, then one may verify using a coupling argument that the inequality of Item (5) holds with
$\mathrm {F_{decay}}(x)$
being some multiple of
$e^{-\frac {x}{2\mathrm {C_{len}}}}$
. The constant
$\mathrm {C_{len}}$
can be thought of as the correlation length of the microscopic model of stochastic kernels. In most specific models of interest, it is not difficult to show (1.1) for any fixed
$k\in \mathbb {N}$
(see Section 6). Nonetheless, we have stated the more general condition (5) above for interest, as it only requires control on a finite number of observables (which is much weaker than total variation bounds).
Condition (6) is just irreducibility of the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
in the usual sense when
$I=c \mathbb {Z}$
. In particular there is no loss of generality as long as the chain is merely recurrent, because one can always replace I by the communicating class of
$0$
throughout all of the assumptions and results. For
$I={\mathbb {R}}$
, (6) is a technical condition that ensures that the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
is sufficiently regularizing. In models of interest, both conditions in Item (6) will be quite easy to check. Thanks to Assumption (6), we will prove the following, see Theorem 3.5 and Proposition 3.8.
Proposition 1.3. Let
$(X_k)_{k\ge 0}$
denote the Markov chain on I associated to the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
from Assumption 1.2 Item (6). There exists an invariant measure
$\pi ^{\mathrm {inv}}$
for
$\boldsymbol {p}_{\mathbf {dif}}$
, unique up to scalar multiple. This measure
$\pi ^{\mathrm {inv}}$
necessarily has full support. For all measurable functions
$f:I\to {\mathbb {R}}$
such that
$|f(x)| \leq F(|x|)$
for some decreasing
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
, one has that
$\int _I |f|d\pi ^{\mathrm {inv}}<\infty $
. Furthermore for all
$f,g\in L^1(\pi ^{\mathrm {inv}})$
such that
$\int _I g\;d\pi ^{\mathrm {inv}}\ne 0$
, we have that
$$ \begin{align*}\lim_{r\to \infty} \frac{\sum_{k=0}^r f(X_k)}{\sum_{k=0}^r g(X_k)} = \frac{\int_I f\;d\pi^{\mathrm{inv}}}{\int_I g\;d\pi^{\mathrm{inv}}}\end{align*} $$
almost surely starting from any point
$X_0\in I$
.
The measure
$\pi ^{\mathrm {inv}}$
will always be infinite under Assumption 1.2, typically resembling Lebesgue or counting measure on I but with extra mass near the origin. With these preliminaries in place, we can state our main result. For
$r\in \mathbb N$
one may define the probability measure
where the product of two Markov kernels is another Markov kernel that is defined by
$K_1K_2(x,A) := \int _I K_2(y,A)K_1(x,{\mathrm {d}} y)$
, and onward by associativity. Intuitively,
$P^\omega (r,\cdot )$
is the quenched transition probability at time r of a random walker starting from the origin, such that the walker at position x at time
$n-1$
uses the kernel
$K_n(x,\cdot )$
to determine what the position will be at time n, for every
$n\ge 0$
and
$x\in I$
. With these definitions and notations in place, we now define the drift constants
Then define the family of deterministic space-time functions
Definition 1.4. Let
$P^\omega (r,\cdot )$
be as in (1.2) and let
$d_N$
be as in (1.3). Let
$D_{N,t,x}$
be the deterministic sequence of space-time functions defined just above. For
$t\in N^{-1}{\mathbb {Z}}$
we define the rescaled density field
Thus for each
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
we note that
$\mathfrak H^N(t,\cdot )$
is a nonnegative Borel measure on
${\mathbb {R}}$
, which is informally equal to the function given by
$x\mapsto N^{1/2}D_{N,t,x} P^\omega (Nt,d_Nt +N^{1/2}x)$
. The definition of
$\mathfrak H^N(t,\cdot )$
is extended to
$t\in {\mathbb {R}}_+$
by linearly interpolating, that is, taking an appropriate convex combination of the two measures at the two nearest points of
$N^{-1}{\mathbb {Z}}_{\ge 0}$
.
An illustration of our generalized model of random walk in random environment. Throughout the paper,
$r \in {\mathbb {Z}}_{\ge 0}$
will be used to denote the microscopic time variable. The random walker starts at position
$x=0$
at time
$r=0$
, then samples
$x_1$
from the random probability measure
$K_1(0,{\mathrm {d}} x)$
. After landing at position
$x_1$
at time
$r=1$
, the random walker then samples
$x_2$
from the probability measure
$K_2(x_1,{\mathrm {d}} x)$
, and continues in this fashion. Each vertical line represents a copy of the underlying lattice
$I= {\mathbb {Z}}$
or
$I={\mathbb {R}}$
, and the Markov kernel
$K_i$
should be thought of as the entire collection of probability measures
$\{K_i(x,\cdot )\}_{x\in I}$
which lives on the vertical line
$r=i-1$
. The random environment is then the whole collection of Markov kernels
$\{K_i\}_{i=1}^\infty $
which in this paper are assumed to be independent of one another for distinct values of i, translationally invariant in law, and sampled from the same distribution on Markov kernels on I. The dotted arrows in the above picture represent any one of many possible jumps that were not actually executed by the random walker in this random environment, while the solid arrows represent the jumps that were executed.

Theorem 1.5 (Main result)
Let
$\{K_n\}_{n\ge 1}$
denote an environment satisfying Assumption 1.2, fix
${T>0}$
, and let
$\mathfrak H^N$
be as in (1.4). There is an explicit Banach space X of distributions on
${\mathbb {R}}$
, which is continuously embedded in
$\mathcal S'({\mathbb {R}})$
, such that the collection
$\{\mathfrak H^N\}_{N \ge 1}$
is tight with respect to the topology of
$C([0,T], X)$
. Furthermore any limit point as
$N\to \infty $
lies in
$C((0,T],C({\mathbb {R}}))$
and coincides with the law of the Itô solution of the SPDE
started with initial data
$\mathcal U(0,x)=\delta _0(x)$
. Here
$\xi $
is a standard space-time white noise on
${\mathbb {R}}_+\times {\mathbb {R}},$
and
$$ \begin{align*}\gamma_{\mathrm{ext}}^2:= \frac{ \frac{(-1)^{p+1}}{(2p)!}\int_{I} \bigg[ \int_{I^2}(x-y)^{2p}\mu({\mathrm{d}} x)\mu({\mathrm{d}} y)-\int_I (a-z)^{2p} \boldsymbol{p}_{\mathbf{dif}}(z,\mathrm da)\bigg] \pi^{\mathrm{inv}}(\mathrm dz)}{\int_{I} \bigg[- |z|+\int_{I} |a| \; \boldsymbol{p}_{\mathbf{dif}}(z,\mathrm da)\bigg]\pi^{\mathrm{inv}}(\mathrm dz)}.\end{align*} $$
The explicit Banach space X is given by a weighted Hölder space with negative exponent and polynomial weight (neither will be optimized). See Section 5 for more details.
The identification of the coefficient
$\gamma _{\mathrm {ext}}^2$
for this general context is one of the main technical challenges and novelties of this work. As remarked earlier, identifying this constant was a challenge even for very simple examples of models driven by IID weights. We remark that (by using that the first
$p-1$
moments are deterministic, and the binomial formula to expand the
$(2p)^{th}$
power) another way to write the integrand in the numerator is
$$ \begin{align*}\frac{(-1)^{p+1}}{(2p)!}\bigg[ &\int_{I^2}(x-y)^{2p}\mu({\mathrm{d}} x)\mu({\mathrm{d}} y)-\int_I (a-z)^{2p} \boldsymbol{p}_{\mathbf{dif}}(z,\mathrm da)\bigg]\\&= - (p!)^{-2} \bigg[m_{p}^2-\int_{I^2} (x-x_0)^p(y-y_0)^p \; \boldsymbol p^{(2)}((x_0,y_0),({\mathrm{d}} x,{\mathrm{d}} y))\bigg] \\ &= (p!)^{-2} \mathrm{Cov}\bigg( \int_I (x-x_0)^p K_1(x_0,{\mathrm{d}} x), \int_I (y-y_0)^p K_1(y_0,{\mathrm{d}} y)\bigg) ,\;\;\;\;\;\; \text{ where } x_0-y_0=z. \end{align*} $$
In the second line, we use the fact that all terms except the middle terms are zero by the deterministic moments in Assumption 1.2 Item (4). We will give for each
$p\in \mathbb N$
a nontrivial example of a model where Theorem 1.5 is applicable. In [Reference Das, Drillick and ParekhDDP24b] we had coined the term “environmental variance coefficient” for
$\gamma _{\mathrm {ext}}$
, which in that paper was
$\frac {8\sigma ^2}{1-4\sigma ^2}$
for the specific choice of model discussed earlier. In the expression for
$\gamma _{\mathrm {ext}}^2$
, one may show from Assumption 1.2 that the integrand in both the numerator and denominator are exponentially decaying functions of the variable z, thus by Proposition 1.3 they are both finite quantities. Jensen’s inequality and the full support of
$\pi ^{\mathrm {inv}}$
imply that the denominator is nonvanishing, thus
$\gamma _{\mathrm {ext}}^2$
is always a finite value. It will be an artifact of the proof of the above theorem that the numerator of the expression for
$\gamma _{\mathrm {ext}}^2$
cannot be a negative value (hence there is no contradiction); see, for example, Proposition 5.18(4) below. With this said, we are not sure if
$\gamma _{\mathrm {ext}}^2$
can be zero under the above assumptions, though one may directly show
$\gamma _{\mathrm {ext}}^2>0$
for many models of interest.
1.2 Context and discussion about the result
Besides the fact that the limit is given by the multiplicative-noise stochastic heat equation (1.5), there are two other aspects of the above result that are worth noting.
-
1. As (1.4) suggests, these KPZ fluctuations are obtained by probing the tail of the probability distribution
$P^\omega (r,{\mathrm {d}} y)$
near spatial location
$y= d_N t+N^{1/2}x$
at time
$r = Nt$
, where N is very large. It is not so surprising that the spatial variable x needs a factor
$N^{1/2}$
at time of order N, since this scaling respects the parabolic structure of the limiting equation. However the factor
$d_N$
in front of t is more mysterious and eluded a rigorous understanding until recently. This drift term of course begs the question of whether or not there are other locations in the tail of
$P^\omega (Nt,\cdot )$
where nontrivial fluctuation behavior can be observed.It is now known that
$d_Nt $
is the unique spatial location at which one expects the KPZ equation to arise. If one were to probe the tail of
$P^\omega (Nt,\cdot )$
at a spatial location of order that is strictly smaller than
$d_Nt$
, one will instead obtain a Gaussian field as the limit, as evidenced by works [Reference Balázs, Rassoul-Agha and SeppäläinenBRAS06, Reference YuYu16, Reference Joseph, Rassoul-Agha and SeppäläinenJRAS19]. These Gaussian fluctuations can be proved for the present model using similar techniques to the ones developed in the present work, see [Reference Drillick and ParekhDP25].On the other hand, when
$d_Nt$
is replaced by a larger quantity (i.e., probing even deeper into the tail of
$P^\omega (Nt,\cdot )$
) it was shown in [Reference Barraquand and CorwinBC17, Reference Barraquand and RychnovskyBR20] that under certain exactly solvable stochastic kernels
$K_i$
, one obtains the Tracy-Widom GUE distribution as the one-point fluctuations. The Tracy-Widom GUE distribution is the one-point marginal of a space-time process called the directed landscape [Reference Matetski, Quastel and RemenikMQR21, Reference Dauvergne, Ortmann and ViragDOV22], a recently constructed universal scaling limit of models in the KPZ universality class. We expect that for all models of stochastic kernels as above and all locations in the tail of
$P^\omega (Nt,\cdot )$
that are strictly larger than
$d_N t$
, we have the directed landscape as the scaling limit. However, this is still a difficult open problem and far beyond the scope of the present work.Consequently, the spatial location
$d_N$
is where one observes the “crossover” from Gaussian to non-Gaussian fluctuation behavior, and the Hopf-Cole solution of the KPZ equation describes this crossover. -
2. There is a nontrivial environmental variance coefficient
$\gamma _{\mathrm {ext}}$
appearing in front of the noise, which depends on the weights only through their two point motions
$\boldsymbol p^{(2)}$
. We shall see later in our proofs that this coefficient arises through interaction of the two particles of the two-point motion generated by the stochastic kernels. Theorem 1.5 thus implies that under the scaling (1.4), the information about the kernels that survives in the limit is only up to the two point motion. Even then, only certain specific observables of the two-point motion are detected by the limiting equation (1.5). But any finer information about higher-order particle interactions will vanish in the limit. This question is related to physics literature of extreme diffusion theory [Reference Le Doussal and ThieryLDT17, Reference Thiery and Le DoussalTLD16, Reference Barraquand and Le DoussalBLD20, Reference Hass, Carroll-Godfrey, Corwin and CorwinHCGCC23, Reference Hartmann, Krajenbrink and Le DoussalHKLD23, Reference Hass, Corwin and CorwinHCC23, Reference Le DoussalLD23, Reference Hass, Drillick, Corwin and CorwinHDCC24, Reference HassHas25].
Works such as [Reference Brockington and WarrenBW22, Reference Hass, Drillick, Corwin and CorwinHDCC24, Reference Das, Drillick and ParekhDDP24b] suggest that for particular models the Markov chain
$\boldsymbol {p}_{\mathbf {dif}}$
is reversible with respect to
$\pi ^{\mathrm {inv}}$
, and that the coefficient
$\gamma _{\mathrm {ext}}^2$
can be simplified further into nicer forms. For example, for the model of [Reference Das, Drillick and ParekhDDP24b] explained earlier, the Markov chain
$\boldsymbol {p}_{\mathbf {dif}}$
just evolves as simple symmetric random walk on
${\mathbb {Z}}$
except that the origin is more attractive, hence
$\pi ^{\mathrm {inv}}$
is counting measure on
${\mathbb {Z}}$
with some extra mass at the origin which can be calculated explicitly by solving the detailed balance equations. In the present work we do not explore the general question of trying to massage the expression for
$\gamma _{\mathrm {ext}}^2$
into alternative forms.
One of the novel aspects of our result is that we have not assumed very strong spatial mixing conditions on the kernel
$K_1$
, only the much weaker condition (5) in Assumption 1.2, which states that fast spatial mixing holds only up to the first
$4p$
joint moments. In particular, we allow the model to have “infinite range of interaction.” As we will see in Section 6, Theorem 1.5 includes as special cases our main results from [Reference Das, Drillick and ParekhDDP24b, Reference Das, Drillick and ParekhDDP24a], with the caveat that we could prove additional interesting results in those papers thanks to the specific structure (see Remark 1.8 below). As noted above, these generalized stochastic kernel models may see a nontrivial interaction among nearby particles, which would include models such as the transport SPDE from [Reference Brockington and WarrenBW22]. The main purpose of the present work is to illustrate the scope and generality of our methods from [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b], showing that they can still be applied in the above setting. It would be difficult to implement, for example, a polynomial chaos in this setting because we do not assume spatial independence, only fast decay of correlations up to (4
$p)^{th}$
-order joint moments of
$K_1(x,\cdot )$
, as in Assumption 1.2 Item (5). But even more crucially, it is unclear what the chaos variables would be under the minimal nature of Assumption 1.2.
Next, we give some further remarks about how to understand the result, and further problems about the quenched density that still remain open.
Remark 1.6. (Drift term and crossover exponents) Let us comment on the drift term
$d_N$
from (1.3). By Taylor expansion at the origin of the function
$\lambda \mapsto \int _I xe^{x\lambda -\log M(\lambda )}\mu ({\mathrm {d}} x)$
we see that to leading order,
The terms in the expansion are just the cumulants of
$\mu $
. In Theorem 1.5 it is actually allowable to replace
$d_N$
by
$\tilde d_N$
, where the latter is given by truncating the above expansion once it reaches scales smaller than
$N^{1/2}$
, as that is the spatial fluctuation scale in Theorem 1.5. If
$p=1$
and
$m_1=0$
, then
$\tilde d_N=N^{3/4}+\frac 12 m_3 N^{1/2},$
because these are the only terms of order
$N^{1/2}$
and higher, and subleading order terms may thus be disregarded. However if
$p=2$
and
$m_1=0$
the leading order term is
$N^{7/8}$
but there are still additional relevant terms of order
$N^{3/4}$
and then
$N^{5/8}$
which cannot be disregarded as these surpass the spatial scaling of order
$N^{1/2}.$
Likewise, for larger values of p, there will be increasingly many subleading terms that contribute to the correct drift term. For instance when
$p=20$
there may be as many as 39 terms of subleading order beyond the leading order term
$N^{79/80}$
that are still relevant in the recentering, and the coefficients of these terms would depend in some complicated way on the first 41 moments of
$\mu $
. This is why we have simply chosen to write the simpler nonexpanded form of
$d_N,$
as opposed to previous special cases [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b]. This idea of going out to larger scales to see the KPZ equation arising is very much reminiscent of [Reference Hairer and QuastelHQ18, Theorem 1.1], where the authors observed a similar hierarchy of nonequivalent scalings leading to the KPZ equation, albeit in a very different context than stochastic flow models. It is also reminiscent of [Reference Hairer and ShenHS17], where the authors similarly observe cumulant-based techniques for certain models of non-Gaussian polymers to obtain the KPZ equation as the scaling limit.
In this sense, Theorem 1.5 shows that
$\frac {4p-1}{4p}$
is the crossover exponent of this model, that is, the unique exponent where one observes the crossover from Gaussian to non-Gaussian behavior (since
$d_N-m_1N$
is given by
$N^{\frac {4p-1}{4p}}$
to leading order). The question of calculating these crossover exponents is physically interesting, and Theorem 1.5 thus provides a fairly complete picture of finding the crossover exponent in one-dimensional random walks in random media, as well as the question of calculating the environmental variance coefficient in full generality. Interestingly, the possible crossover exponents form a discrete set
$\{3/4, 7/8, 11/12, 15/16,...\}$
. While discussing the subject of the drift term and the crossover exponents, let us now acknowledge two concurrent physics works that are strongly related to this work, and provided much of the motivation for it.
While we were initially only aware of the case
$p=1$
, it was observed by J. Hass [Reference HassHas25] that for certain types of models satisfying “symmetry up to order p” one expects different locations than
$N^{3/4}$
at which the KPZ equation appears in the tail, for instance
$N^{7/8}$
or
$N^{11/12}.$
That paper observes that the fluctuations would approach the deterministic heat equation if, for instance,
$p=2$
but one probes into location
$N^{3/4}$
within the tail of the quenched density
$P^\omega (Nt,\cdot )$
rather than
$N^{7/8}$
as in Theorem 1.5. That paper furthermore establishes the conjecture of super-universality of KPZ arising for different values of p, based on a more physical argument (in the spirit of earlier work of [Reference Le Doussal and ThieryLDT17, Reference Barraquand and Le DoussalBLD20] on the 3/4 case). For the 7/8 case, [Reference HassHas25] is also able to numerically validate his predictions and show they hold to reasonable size of N. This paper of Hass provided the inspiration for us to consider the
$p>1$
case in the present work.
Another physics paper [Reference Hass, Drillick, Corwin and CorwinHDCC24] argues based on convergence of second moments to those of the SHE for the convergence to the KPZ equation just in the 3/4 exponent case
$(p=1)$
. That paper also includes a very nice interpretation for the coefficient
$\gamma _{\mathrm {ext}}$
, translates the convergence result into the language of extreme diffusion theory, and numerically validates the asymptotic predictions down to rather small values of N.
Remark 1.7 (Optimality questions)
In Condition (3) the normalization condition
$m_2-m_1^2=1$
is not crucial, one can always change the lattice spacing to make this condition true (e.g., replace
$I={\mathbb {Z}}$
by
$I=a{\mathbb {Z}}$
for the appropriate value of
$a>0$
, and rescale the kernels accordingly). For the limiting equation, this rescaling just corresponds to changing the diffusion coefficient
$\frac 12$
to some other value in (1.5).
On the other hand, we are not sure if the exponential moments condition on
$\mu $
can be relaxed without changing the outcome of Theorem 1.5, for example, in polymers just six moments is enough [Reference Alberts, Khanin and QuastelAKQ14, Reference Dey and ZygourasDZ16]. However in our model we rely on the moment generating function more crucially and we are unsure if there are more optimal conditions. Certainly at least
$4p$
moments is necessary since this many moments appear in all of the cumulant expansions and the statement of Theorem 1.5, but we have no conjecture about optimality.
Another question is whether it is possible to upgrade the convergence in Theorem 1.5 to this topology, since the law of the limiting SPDE is supported on
$C((0,T],C(\Bbb R))$
. We showed in [Reference Das, Drillick and ParekhDDP24b] that this is impossible. This is intimately related to the failure of the chaos expansion technique for this model and the failure of the noise coefficient of the limiting SPDE to match the noise coefficient of the prelimiting model (both of which were also explained there). See the end of Subsection 6.3 for more discussion on this. Thus, we conclude that weak convergence in a topology of tempered distributions is the best that one may hope for in Theorem 1.5.
Another interesting question is if the decay of
$\mathrm {F_{decay}}$
in Assumption 1.2 Item (5) can be replaced by a weaker decay such as just being decreasing and integrable, for example,
$\int _0^\infty \mathrm {F_{decay}}(x){\mathrm {d}} x<\infty $
. Although we have not tried to optimize this assumption, we believe that the latter condition is the truly optimal one. The only place we use the stronger condition is the proof of Proposition 4.9, where it is needed to control one of the error terms coming from the martingale equation.
An even more interesting question would be to introduce some weak temporal correlation between the stochastic kernels
$K_i$
, rather than assuming they are independent as we do here. This would require a completely different proof than the one given here, as the k-point motion would lose the Markov property which is crucial in all of our calculations. This may be explored in future work.
Remark 1.8 (Quenched tail field)
While we have defined the rescaled field (1.4) as a distribution-valued field (i.e., only well-defined when integrated against spatial test functions), one might ask about function-valued counterparts. This leads us to define the quenched tail field. For
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
and
$x\in {\mathbb {R}}$
we define the quenched tail field by
with
$P^\omega $
as in (1.2). Then we may prove the following result.
Fix
$m\in \mathbb {N}$
, and
$\phi _1,\ldots ,\phi _m\in C_c^{\infty }(\mathbb {R})$
. Consider sequences
$t_{N,1},\ldots ,t_{N,m} \in N^{-1}{\mathbb {Z}}_{\ge 0}$
such that
$t_{N,i} \to t_i>0$
for all
$1\le i \le m$
as
$N\to \infty .$
Then
$$ \begin{align*} \bigg( \int_{\mathbb{R}} \phi_i(x) F_N(t_{N,i},x) {\mathrm{d}} x\bigg)_{i=1}^m \stackrel{d}{\to} \bigg( \int_{\mathbb{R}} \phi_i(x) \mathcal U_{t_i}(x) {\mathrm{d}} x\bigg)_{i=1}^m. \end{align*} $$
Here
$(t,x)\mapsto \mathcal U_t(x)$
is the solution of (1.5).
The proof of this follows using the same exact arguments used in Sections 7 of [Reference Das, Drillick and ParekhDDP24a] and [Reference Das, Drillick and ParekhDDP24b], first proving that
${\mathbb {E}}[F_N(t,x)] \leq 4(\pi t)^{-1/2}$
and then proceeding exactly as we did in those papers using the result of Theorem 1.5 and integrating by parts.
While proving convergence with test functions is interesting, one might ask about multi-point convergence in law of
$F_N$
at some finite collection of individual points
$\{(t_{N,i},x_{N,i})\}_{i=1}^m$
converging as
$N\to \infty $
to
$\{(t_i,x_i)\}_{i=1}^m$
. This remains an open problem, as we do not have the estimates necessary to prove this result in the present generalized context. In [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b] we were actually able to prove this multi-point convergence from the above test function result by using additional structure of those specific models. This is because we were able to exploit the exact solvability of the 2-point motion of those models, which is a very powerful tool when used in conjunction with the more analytic ideas here. This illustrates the payoff of the generality of our result. We can obtain Theorem 1.5 but the additional interesting results proved in [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b] (e.g., extremal particle limit theorems and creation of independent noise) remain out of reach at this level of generality.
1.3 Main ideas of the proof
Previous results about the fluctuations of
$P^\omega $
were proved either using cluster expansion techniques or by using the environment seen from the particle [Reference Boldrighini, Minlos and PellegrinottiBMP97, Reference Boldrighini and PellegrinottiBP01, Reference Balázs, Rassoul-Agha and SeppäläinenBRAS06, Reference YuYu16]. The method of the present work will be different, instead capitalizing on the discrete SPDE technique. More specifically, we show that the prelimiting model (1.4) satisfies a discretized version of the martingale problem for (1.5). This discretized martingale problem is in turn based on Girsanov’s formula. This Girsanov technique was first discovered independently in the papers [Reference Brockington and WarrenBW22, Reference Das, Drillick and ParekhDDP24a], but in the present work we have numerous additional difficulties coming from the general assumptions.
We now explain the discrete Girsanov formula. We illustrate this idea by computing the first moment of
$\mathfrak H^N$
as defined in (1.4) and showing that it converges to the first moment of the SHE. This computation contains the core idea that will be used in many of the later proofs. Recall the measure
$\mu $
from Assumption 1.2 Item (3) and its moment generating function
$M(\lambda )$
from (1.3). For a random walk
$(R_r)_{r\ge 0}$
on I starting from the origin with increment distribution
$\mu $
, the process
$$ \begin{align*}e^{\lambda R_r - r\log M(\lambda)} = \frac{e^{\lambda R_r}}{{\mathbb{E}}[e^{\lambda R_r}]}\end{align*} $$
is a martingale in the variable
$r\ge 0$
, simply because it is a product of r IID strictly positive mean-one random variables
$e^{\lambda \xi _i}/{\mathbb {E}}[e^{\lambda \xi _i}]$
where the
$\xi _i$
are the increments of the process
$(R_r)_{r\ge 0}$
. Given this fact, let us now fix some Schwartz function
$\phi $
and compute the first moment of the pairing
$(\mathfrak H^N(t,\cdot ),\phi )_{L^2({\mathbb {R}})}=: \mathfrak H^N(t,\phi ).$
Note that for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
where
$E^\omega $
denotes a quenched expectation operator given the realization of the environment kernels
$\omega = (K_r)_{r=0}^\infty $
. Thus after taking the annealed expectation (that is, averaging over all possible environments
$\omega $
) we have
where now the expectation on the right is with respect to a random walk path on I with increment distribution
$\mu $
. Notice that
We already know that
$e^{N^{-\frac 1{4p}}R_{Nt} - Nt\log M(N^{-\frac 1{4p}})}$
is a martingale in the
$Nt$
variable and thus has mean 1, so it can be interpreted as a change of measure where the annealed law of the increments of
$R_r$
has changed from the usual law
$\mu $
to the new law given by
$\mu ^{N^{-\frac 1{4p}}}({\mathrm {d}} x) = e^{N^{-\frac 1{4p}} x - \log M(N^{-\frac 1{4p}})} \mu ({\mathrm {d}} x)$
. This new law has mean equal to
$N^{-1}d_N$
with
$d_N$
defined by (1.3). Consequently, under the new law the process
$N^{-1/2}(R_{Nt}-d_Nt)$
is centered, and by Donsker’s principle will converge in law to a standard Brownian motion. Denoting
$\widetilde {\mathbf E}$
as the tilted expectation on the path space of the random walk, we have that
where
$\mathbf E_{\mathrm {BM}}$
denotes expectation with respect to a standard Brownian motion B. Here
$p_t(x) = (2\pi t)^{-1/2}e^{-x^2/2t}$
is the standard heat kernel. The right-hand side equals
${\mathbb {E}} \big [ \int _{\mathbb {R}} \mathcal U_t(x)\phi (x){\mathrm {d}} x \big ],$
where
$(t,x)\mapsto \mathcal U_t(x)$
solves (1.5) with initial condition
$\delta _0$
. This shows that the first moment of
$\mathfrak H^N(t,\cdot )$
converges to the first moment of the SHE as desired. While the calculation for higher moments is more complicated and involves analyzing the so-called k-point motion, the core idea of using a “discrete Girsanov transform” is rather similar, and these higher moments will be computed in Section 3.
This k-point motion will play an extremely important role in the analysis, that is, the Markov chain on
$I^k$
with transition kernel
$\boldsymbol p^{(k)}$
from Assumption 1.2. One of the key observations of the paper is that even after applying the Girsanov transform to this Markov chain, the tilted process
$\boldsymbol q^{(k)}_\beta $
is still a time-homogeneous Markov chain, see Corollary 2.7. This was not observed in our previous works [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b], which instead used a workaround method. Thus, much of the bulk of the paper consists of fairly standard estimates for these tilted Markov chains, extracting information and estimates from the associated Dynkin martingales.
Once the moments are computed using this method, in Section 4 we derive a discrete martingale-driven SPDE for the field in (1.4). The martingale observables arising from this SPDE can also be analyzed using the aforementioned Girsanov-type transforms. In particular, this Girsanov trick yields moments and regularity estimates for our observables which eventually lead to tightness estimates for our field. Sufficiently strong estimates using this method will eventually lead us to the proof of Theorem 1.5, by showing that any limit point must satisfy the martingale problem for (1.5). Remarkably, this model has the property that the error terms appearing in the discrete martingale equation behave very nicely in relation to the original object itself, which is rare among KPZ-related models where martingale characterizations have been used, see, for example, [Reference Bertini and GiacominBG97, Reference Dembo and TsaiDT16, Reference Gonçalves and JaraGJ14, Reference YangYan23] where an extremely careful analysis was needed to show vanishing of error terms for exclusion-type models.
Let us briefly comment on similarities and differences of this paper with previous works [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b] in this area. Both of those papers were concerned with a particular special case of a model satisfying Assumption 1.2 with
$p=1$
. In the case
$p=1$
, there is a trick to control the tilted processes arising from the Girsanov transform, in which one can write the tilted measure in terms of an auxiliary process that is very easy to control, modulo a Radon-Nikodym derivative that stays tight in
$C[0,T]$
as
$N\to \infty $
. This trick simplifies many calculations. For
$p>2$
this trick fails completely as the Radon-Nikodym derivatives with respect to any candidate of an auxiliary process will simply fail to be tight. Consequently we cannot apply this trick, and instead we need to directly study the tilted processes arising from the k-point motion, using the fact that they are Markovian as mentioned above. This will be done in Sections 2 and 3 and is one of the main technical novelties in the paper. These sections are also the places where the identification of this coefficient
$\gamma _{\mathrm {ext}}^2$
will be done. Once this is done, the remainder of the proof of convergence will follow the approach of [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b], although finding a discrete martingale-driven SPDE for the general model is of course more involved than for any particular model. Many of the arguments need to be adapted to the case of general p-values which sometimes leads to additional calculations necessary in the proofs. As we already explained in Remark 1.8, a further difference with those papers is that the specific models studied there had nice properties that often led to additional interesting results that we do not pursue here. One key difference of the present model with those of [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b] is the lack of exact solvability of the two-point motion, which is enjoyed by the specific models of those works.
Outline: In Section 2 we will define the k-point motion, a Markov chain on
$I^k$
associated to the kernels
$(K_r)_{r\ge 0} $
that is instrumental in proving Theorem 1.5. In Section 3 we will continue to study the k-point motion and in particular prove convergence theorems about it under diffusive scaling. In Section 4 we will begin studying the field (1.4) and in particular show that it satisfies a discrete version of (1.5). Section 5 will then use the results of Section 4 to complete the proof of Theorem 1.5. In Section 6 we will introduce a number of models, both discrete and continuous, and verify Assumption 1.2 for these models.
Notations: We use several different notions for probability and expectation in this paper. We will use
${\mathbb {E}}$
for annealed expectations with respect to the environment kernels
$K_i$
of Assumption 1.2. We will use
$E^\omega _{(k)}$
for quenched expectation of k independent random walkers, each of which uses the kernels
$K_i$
to transition to the next step. Thus these are random measures that are a function of the kernels
$K_i$
. We will use a boldface
$\mathbf E$
for expectations with respect to stochastic processes on the path space
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
. Typically these measures will arise from annealed expectations over the quenched path measures, for example,
$\mathbf E_{{RW}^{(k)}} = \mathbb EE^\omega _{(k)}$
in the notation of Section 2 below. These will be deterministic measures.
2 The k-point motion and its Girsanov tilts
In this section, we introduce and study the k-point motion of the stochastic flow introduced in Assumption 1.2, and we prove various properties about it under certain exponential Girsanov transforms that are relevant for the proof of Theorem 1.5.
Definition 2.1 (k-point motion)
Fix
$k\in \Bbb N$
. We define the path measure
${\mathbf {P}}_{\mathrm {RW}^{(k)}}$
on the canonical space
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
to be the annealed law of k independent walks (all started from the origin) sampled from kernels
$K_n$
satisfying Assumption 1.2.
We remark that the canonical process
$\mathbf R= (\mathbf R_r)_{r\ge 0}$
is a Markov chain on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
under
${\mathbf {P}}_{\mathrm {RW}^{(k)}}$
, whose one-step transition probabilities are precisely
$\boldsymbol p^{(k)}({\mathbf {x}},A) := \int _A \boldsymbol p^{(k)}({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})$
where the latter was defined in Assumption 1.2 Item (4). This Markov chain is called the
$ {k}$
-point motion. The most basic property of the k-point motion is that it is projective: the temporal evolution of any deterministic subset consisting of
$\ell <k$
coordinates behaves as the
$\ell $
-point motion, which is immediate from the definition. In particular, each coordinate
$R^j$
of the k-point motion is marginally distributed as a random walk on I (i.e., has independent increments) with increment distribution
$\mu $
as defined in Assumption 1.2(3). However, the marginal law of a pair
$( R^i, R^j)$
of distinct coordinates will not be as simple, for instance, it may not have independent increments. Nonetheless, one can say the following.
Lemma 2.2. For any
$1\le i,j\le k$
with
$i\neq j$
, the process
$R^i-R^j$
under
${\mathbf {P}}_{\mathrm {RW}^{(k)}}$
is a Markov chain with transition kernel
$\boldsymbol {p}_{\mathbf {dif}}$
as defined in Assumption 1.2 Item (6).
The proof is straightforward from the definitions.
Definition 2.3. Let
$m,k\in \mathbb N$
and
$\mathbf j=(j_1,...,j_m) \in \mathbb \{1,...,k\}^m$
. For
${\mathbf {x}}, {\mathbf {y}}\in I^k$
define the joint cumulants
$$ \begin{align*}\kappa^{(m;k)}(\mathbf j;{\mathbf{x}}) := \frac{\partial^m}{\partial\beta_{j_1} \cdots \partial \beta_{j_m}} \log \int_{I^k} \exp \bigg( \sum_{i=1}^k \beta_i (y_i-x_i) \bigg) \boldsymbol p^{(k)} ({\mathbf{x}}, {\mathrm{d}} {\mathbf{y}})\;\;\bigg|_{(\beta_1,...,\beta_k)=(0,...,0)},\end{align*} $$
where the vertical bar denotes evaluation of the function at the origin.
When taking the above partial derivatives, we remark for clarity that the indices
$j_i$
may be repeated as many times as desired.
Lemma 2.4. Suppose that
$m,k\geq 2$
and
$\mathbf j=(j_1,...,j_m) \in \mathbb \{1,...,k\}^m$
such that all
$j_i$
are not the same index, that is,
$(j_1,...,j_m)$
is not of the form
$(j,j,...,j)$
for some
$j=1,...,k$
. Let
$p\in \mathbb N$
be as in Assumption 1.2 Item (4).
-
1. If
$m< 2p$
then
$\kappa ^{(m;k)}(\mathbf j;{\mathbf {x}})=0$
-
2. If
$m= 2p$
then
$\kappa ^{(m;k)}(\mathbf j;{\mathbf {x}})=0$
, unless the set of indices
$\{j_1,...,j_m\}=\{a,b\}$
has cardinality exactly two where exactly p of the indices
$j_i$
are equal to a (and thus the other p indices are equal to b). -
3. If
$2p\leq m\leq 4p$
, then there exists
$C>0$
such that
$$ \begin{align*}|\kappa^{(m;k)}(\mathbf j;{\mathbf{x}})| \leq C \cdot \mathrm{F_{decay}} \big(\min_{1\leq i <j \leq k} |x_i-x_j|\big).\end{align*} $$
Proof. For any multi-index
$\mathbf j$
, note that
$\kappa ^{(m;k)}(\mathbf j,{\mathbf {x}})$
is actually just a finite linear combination of products of quantities of the form
$\int _{I^k} \prod _{j=1}^k (y_j-x_j)^{r_j} \boldsymbol p^{(k)} \big ( {\mathbf {x}}, d{\mathbf {y}}\big ), $
where
$0\leq r_1+...+r_k\leq m$
.
Moreover, it is a very special linear combinationFootnote
1
with the following property: if the random variables
$y_i-x_i (1\le i \le k)$
were independent under
$\boldsymbol p^{(k)}({\mathbf {x}},d{\mathbf {y}})$
then in fact this linear combination would vanish identically, simply by the definition of the cumulants and the condition that all indices are not the same (i.e., joint cumulants of independent variables are always zero).
But for
$m<2p$
the moments
$\int _{I^k} \prod _{j=1}^k (y_j-x_j)^{r_j} \boldsymbol p^{(k)} \big ( {\mathbf {x}}, d{\mathbf {y}}\big ), $
agree precisely with those of the independent ones given by
$\int _{I^k} \prod _{j=1}^k u_j^{r_j}\mu ^{\otimes k} ({\mathrm {d}}\mathbf u),$
by the deterministic moments in Item (4) of Assumption 1.2. Therefore we immediately conclude Item (1). The proof of Item (2) is similar, noting that the moments agree precisely with those of the independent family unless the given condition on the indices is satisfied.
For
$2p\leq m\leq 4p$
, the
$m^{th}$
order joint moments of
$y_i-x_i$
under
$\boldsymbol p^{(k)}({\mathbf {x}},d{\mathbf {y}})$
will not necessarily agree precisely with the independent version under
$\mu ^{\otimes k}$
, hence they may not vanish outright. However the moment bound in Item (5) of Assumption 1.2 guarantees that all of the moments are close to those of the independent ones as measured by the function
$\mathrm {F_{decay}}$
, therefore we can immediately conclude Item (3).
Note that if
$(X_r)_{r\ge 0}$
is any sequence of random variables with finite exponential moments, defined on some probability space
$(\Omega ,\mathcal F,{\mathbf {P}})$
and adapted to a filtration
$(\mathcal F_r)_{r\ge 0}$
on this space, then the process
$$ \begin{align}\mathcal E_r(X):=\exp\bigg( X_r- \sum_{s=1}^{r} \log \mathbf E[e^{X_{s}-X_{s-1}}|\mathcal F_{s-1}]\bigg)\end{align} $$
is a martingale in
$r\ge 0$
, with respect to the same filtration. Henceforth, our probability space will always be
$\Omega =(I^k)^{{\mathbb {Z}}_{\ge 0}},$
equipped with its canonical filtration. For
$\beta \in [-\eta /k,\eta /k]$
with
$\eta $
defined in Assumption 1.2 Item (3), if we apply (2.1) to the process
$X:=\beta (R^1+\cdots +R^k)$
, we see that

is a
${\mathbf {P}}_{\mathrm {RW}^{(k)}}$
-martingale for the k-point motion where
$f_\ell ^{\beta ,k} := f^{\beta ,k}(R_\ell ^1,\ldots ,R_\ell ^k)$
, and the latter function
$f^{\beta ,k}$
is given by
Here
$\mathbf E_{\mathrm {RW}^{(k)}}^{(x^1,\ldots , x^k)}$
denotes expectation of the Markov chain when started from
$(x^1,\ldots ,x^k).$
We then use the martingale (2.2) to define the following measure.
Definition 2.5. For
$\beta \in [-\eta /k,\eta /k$
] with
$\eta $
defined in Assumption 1.2 Item (3), define the tilted path measure
${\mathbf {P}}^{(\beta ,k)}$
on the canonical space
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
to be given by

where
$\mathcal F_r$
is the
$\sigma $
-algebra on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
generated by the first r coordinate maps, and
is the martingale given by (2.2). We will denote by
$\mathbf E^{(\beta ,k)}$
the expectation operator associated to
${\mathbf {P}}^{(\beta ,k)}.$
An obvious but important fact is that
${\mathbf {P}}^{(0,k)}={\mathbf {P}}_{\mathrm {RW}^{(k)}}$
which will implicitly be used many times below.
Lemma 2.6. Let
$\mathfrak H^N(t,\cdot )$
be as defined in (1.4) in the introduction. For
$t\in N^{-1}{\mathbb {Z}}$
, the annealed expectation
$ {\mathbb {E}}[\mathfrak H^N(t,\phi )^k]$
is given by
$$ \begin{align*} \mathbf E^{(N^{-\frac1{4p}},k)}\bigg [ \exp \bigg( \sum_{s=1}^{Nt} \bigg\{ \log \mathbf E_{\mathrm{RW}^{(k)}} \big[ e^{N^{-\frac1{4p}} \sum_{j=1}^k (R^j_s -R^j_{s-1})} \big| \mathcal F_{s-1}\big] \; - k \log M(N^{-\frac1{4p}})\bigg\}\bigg) \prod_{j=1}^k \phi\big(N^{-1/2}(R^j_{Nt}- d_Nt)\big)\bigg] \end{align*} $$
where we recall from (1.3) that M is the moment generating function of the measure
$\mu $
from Assumption 1.2 Item (3).
Proof. Since
$P^\omega (Nt,\cdot )$
as defined in (1.2) is a probability measure we can write
$$ \begin{align}\mathfrak H^N(t,\phi)^k &= E^\omega_{(k)} \bigg[\prod_{j=1}^kD_{N,t,N^{-1/2}(R^j_{Nt}- d_N t)}\phi\big(N^{-1/2}(R^j_{Nt}- d_Nt)\big)\bigg], \end{align} $$
where the latter is a quenched expectation of k independent particles
$(R^1,...,R^k)$
started at 0 and sampled from a fixed realization of the environment kernels
$\omega = \{K_n\}_{n\ge 1}.$
Therefore the claim just follows by taking the annealed expectation over (2.3) and using the definitions of the measures.
2.1 Studying the tilted measures
${\mathbf {P}}^{(\beta ,k)}$
Proposition 2.7. Take
$|\beta |\leq \eta /k$
where
$\eta $
is as in Assumption 1.2 Item (3) and
$k\in \mathbb N$
. The canonical process
$(\mathbf R_r)_{r\ge 0}$
on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
under the tilted measure
${\mathbf {P}}^{(\beta ,k)}$
of Definition 2.5 is a Markov chain on
$I^k$
, with one-step transition law given by
$$ \begin{align}\boldsymbol q^{(k)}_\beta ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) := \frac{e^{\beta \sum_{i=1}^k(y_i-x_i)} \boldsymbol p^{(k)}({\mathbf{x}},{\mathrm{d}}{\mathbf{y}})}{\int_{I^k}e^{\beta \sum_{i=1}^k(a_i-x_i)} \boldsymbol p^{(k)}({\mathbf{x}},{\mathrm{d}}{\mathbf{a}})}.\end{align} $$
Proof. Take any Borel set
$A\subset I^k$
. By the definition of the measures
${\mathbf {P}}^{(\beta ,k)}$
, we have

thus proving the claim. In the third line we applied the Markov property of
${\mathbf {P}}_{\mathrm {RW}^{(k)}}.$
The condition
$|\beta |\leq \eta /k$
ensures that all moment generating functions actually exist, that is, the integrals are convergent for all
${\mathbf {x}}\in I^k$
, for example, by Hölder’s inequality which says that
$\int _{I^k}e^{\beta \sum _{i=1}^k (a_i-x_i)} \boldsymbol p^{(k)}({\mathbf {x}},{\mathrm {d}}{\mathbf {a}}) \leq \prod _{i=1}^k \big (\int _{I^k}e^{k\beta (a_i-x_i)} \boldsymbol p^{(k)}({\mathbf {x}},{\mathrm {d}}{\mathbf {a}})\big )^{1/k} = M(\beta k)$
, where M as always denotes the moment generating function of the annealed one-step law as in (1.3).
Given the explicit expression (2.4), we thus make the following definition of the tilted k-point Markov chain, generalized to start from arbitrary points. These Markov chains will be of fundamental importance.
Definition 2.8. Define the measure
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
to be the law on the canonical space
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
of the Markov chain associated to (2.4) started from
${\mathbf {x}} \in I^k.$
We also let
$\mathbf E_{\mathbf {x}}^{(\beta ,k)}$
be the associated expectation operator.
Immediately from Proposition 2.7 we see that the path measures
${\mathbf {P}}^{(\beta ,k)}$
from Definition 2.5 are a special case of the measures
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
with
${\mathbf {x}}= (0,0,...,0).$
In particular
${\mathbf {P}}_{\mathrm {RW}^{(k)}} = {\mathbf {P}}^{(0,k)}_{(0,...,0)}.$
Lemma 2.9. Fix
$k\in \mathbb N$
. Let
$\eta $
be as in Assumption 1.2 Item (3). With
$\boldsymbol q^{(k)}_\beta $
defined in (2.4), we have
$$ \begin{align*}\sup_{|\beta|\leq \eta/(2k)}\;\; \sup_{{\mathbf{x}}\in I^k} \max_{1\le i\le k} \int_{I^k} e^{\frac{\eta}2 |y_i-x_i|} \boldsymbol q^{(k)}_\beta ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) <\infty.\end{align*} $$
In particular, we have that
$$ \begin{align*}\sup_{|\beta|\leq \eta/(2k)}\; \; \sup_{{\mathbf{x}}\in I^k} \int_{I^k} e^{\frac{\eta}{2k} \sum_{i=1}^k |y_i-x_i|} \boldsymbol q^{(k)}_\beta ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) <\infty.\end{align*} $$
Proof. The second bound follows immediately from the first, since by Hölder’s inequality we have that
$\int _{I^k} e^{\lambda \sum _{i=1}^k |y_i-x_i|} q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}}) \le \prod _{i=1}^k \big ( \int _{I^k} e^{k\lambda |y_i-x_i|} q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}}) \big )^{1/k}.$
Thus we will prove the first bound. We find for
$\lambda \ge 0$
and
$\beta \in [-\eta /(2k),\eta /2k]$
and
$1\le i \le k$
that
$$ \begin{align*} \int_{I^k} e^{\lambda |y_i-x_i|} \boldsymbol q^{(k)}_\beta ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) &= \frac{\int_{I^k} e^{\lambda |y_i-x_i|} e^{\beta \sum_{j=1}^k (y_j-x_j)}\boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) }{\int_{I^k} e^{\beta \sum_{j=1}^k (y_j-x_j)} \boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) } \\ &\leq \frac{\big(\int_{I^k} e^{2\lambda |y_i-x_i|} \boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) \big)^{1/2}\prod_{j=1}^k \big(\int_{I^k} e^{2k|\beta| |y_j-x_j|} \boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) \big)^{1/(2k)}}{e^{\int_{I^k}\beta \sum_{j=1}^k (y_j-x_j) \boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}})} } \\ &= \frac{\big(\int_I e^{2\lambda |a|} \mu(\mathrm da)\big)^{1/2}\big(\int_I e^{2k |\beta||a|} \mu(\mathrm da)\big)^{1/2}}{e^{\beta k m_1}}. \end{align*} $$
In the first line we simply applied the definition of the measures
$\boldsymbol q^{(k)}_\beta $
. In the second line we used Hölder’s inequality for the numerator and Jensen for the denominator. In the last line we used the fact that the marginal law of each
$y_j-x_j$
under
$\boldsymbol p^{(k)} ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})$
is just
$\mu $
, and
$m_1:= \int _I a\mu (\mathrm da).$
By Assumption 1.2 Item (3) the last expression is clearly finite if
$\lambda =\eta /2$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
, thus completing the proof.
Corollary 2.10. Let
$\eta $
be as in Assumption 1.2 Item (3), and let
$\beta \in [-\eta /(2k),\eta /(2k)]$
. For any
$f:I^k\to {\mathbb {R}}$
such that
$|f({\mathbf {x}})| \leq Ce^{\frac {\eta }{2k}\sum _{i=1}^k |x_i|}$
, define
$\mathscr Q_k^\beta f({\mathbf {x}}) := \int _{I^k} f({\mathbf {y}}) q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})$
. Then the process
$$ \begin{align*}\mathcal M_r^{f,\beta}(\mathbf R):= f(\mathbf R_k)- \sum_{s=0}^{r-1} (\mathscr Q_k^\beta-\mathrm{Id}) f(\mathbf R_s)\end{align*} $$
is a
${\mathbf {P}}^{(\beta ,k)}_{\mathbf {x}}$
-martingale in the canonical filtration
$(\mathcal F_r)_{r\ge 0},$
for every
${\mathbf {x}}\in I^k$
.
Proof. The proof of martingality is immediate from the fact that
$\mathbf E_{\mathbf {x}}^{(\beta ,k)} [f(\mathbf R_{r+1})|\mathcal F_r] = \mathscr Q_k^\beta f(\mathbf R_r)$
by Proposition 2.7. The growth condition on f ensures that all integrals and expectations are finite, for example, by Lemma 2.9.
Definition 2.11. Henceforth we define the function
$\boldsymbol {\zeta }:I\to {\mathbb {R}}$
by
$$ \begin{align*} \boldsymbol{\zeta}(y_1-y_2) := (p!)^{-2} \int_{I^2} \prod_{j=1}^2 (x_j-y_j)^p \bigg(\boldsymbol p^{(2)} \big( (y_1,y_2),({\mathrm{d}} x_1,{\mathrm{d}} x_2)\big) - \mu({\mathrm{d}} x_1-y_1)\mu({\mathrm{d}} x_2-y_2)\bigg). \end{align*} $$
The fact that the expression on the right side defines a function of
$y_1-y_2$
follows from the translation invariance in Assumption 1.2 Item (2). This function
$\boldsymbol {\zeta }$
will play a crucial role throughout the remainder of the paper, as it is precisely the function appearing in the numerator of the noise coefficient
$\gamma _{\mathrm {ext}}$
in Theorem 1.5. We remark that
$\boldsymbol {\zeta }$
decays to
$0$
quite fast at infinity thanks to Assumption 1.2 Item (5).
Note that the relevant observable inside Lemma 2.6 is given by the cumulant generating function of a certain sum minus the sum of individual cumulant generating functions (latter being the
$k\log M(N^{-\frac 1{4p}})$
term). The natural thing to do here is to Taylor expand everything in the variable
$N^{-\frac 1{4p}}$
, thus a study of joint cumulants will be necessary going forward.
Proposition 2.12. Fix
$k\in \mathbb N$
. There exists
$C>0$
such that the following bounds hold uniformly over
$\beta _1,...,\beta _m \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}\in I^k$
and multi-indices
$\mathbf j =(j_1,...,j_m)\in \bigcup _{m\ge 0} \{1,...,k\}^m $
$$ \begin{align*}\bigg|\frac{\partial^m}{\partial \beta_{j_1}\cdots\partial\beta_{j_m}}&\bigg( \log \int_{I^k} e^{\sum_{i=1}^k \beta_i (a_i-x_i)} \boldsymbol p^{(k)}({\mathbf{x}},{\mathrm{d}}{\mathbf{a}}) - \sum_{i=1}^k\log M(\beta_i) - \sum_{1\le i<j\le k} \beta_i^p\beta_j^p\boldsymbol{\zeta} (x_i-x_j)\bigg)\bigg| \\&\leq C\bigg( \bigg|\sum_{i=1}^k|\beta_i|\bigg|^{2p+1-m} \mathrm{F_{decay}}\big( \min_{1\le i<j\le k}|x_i-x_j|\big)\;\;+\;\;\bigg|\sum_{i=1}^k|\beta_i|\bigg|^{4p+1-m}\bigg). \end{align*} $$
Here
$\mathrm {F_{decay}}$
is the function from Assumption 1.2 Item (5).
Proof. Let us define
$$ \begin{align}\mathcal K({\mathbf{x}},\boldsymbol\beta):= \log \int_{I^k} e^{\sum_{i=1}^k \beta_i (a_i-x_i)} \boldsymbol p^{(k)}({\mathbf{x}},{\mathrm{d}}{\mathbf{a}})\;\;\;-\;\;\;\sum_{i=1}^k\log M(\beta_i), \end{align} $$
so that by Definition 2.3 the
$(4p)^{th}$
order Taylor expansion of
$\mathcal K$
in the
$\beta _i$
variables is given by
$$ \begin{align*}\mathcal K_{4p}({\mathbf{x}},\boldsymbol\beta)&:= \sum_{m=1}^{4p} \frac1{m!} \bigg(\sum_{j_1,...,j_m=1}^k \beta_{j_1}\cdots \beta_{j_m}\kappa^{(m;k)} (\mathbf j,{\mathbf{x}})\;\;\;-\;\;\; \sum_{\substack{1\leq j_1,...,j_m\leq k\\ j_i \text{ all same}}} \beta_{j_1}\cdots \beta_{j_m} \kappa^{(m;k)} (\mathbf j,{\mathbf{x}})\bigg) \\ &= \sum_{m=1}^{4p} \frac1{m!}\sum_{\substack{1\leq j_1,...,j_m\leq k\\ j_i \text{ not all same}}}\beta_{j_1}\cdots \beta_{j_m} \kappa^{(m,k)} (\mathbf j; {\mathbf{x}})\\&=\sum_{m=2p}^{4p} \frac1{m!}\sum_{\substack{1\leq j_1,...,j_m\leq k\\ j_i \text{ not all same}}}\beta_{j_1}\cdots \beta_{j_m} \kappa^{(m,k)} (\mathbf j; {\mathbf{x}})\end{align*} $$
In the first equality, the terms with all
$j_i$
being the same comes from the
$\log M(\beta _i)$
contribution to
$\mathcal K$
, and uses the fact that the marginal law of
$a_i-x_i$
under
$\boldsymbol p^{(k)}({\mathbf {x}},{\mathrm {d}}{\mathbf {a}})$
is precisely
$\mu $
for every
${\mathbf {x}}\in I^k$
. In the third line, we used Item (1) in Lemma 2.4 which guarantees that all of the joint cumulants up to order
$2p-1$
vanish.
By Taylor’s theorem and the uniform exponential moment bounds in Lemma 2.9, for any fixed indices
$\ell _1,...,\ell _m,$
the difference
$\mathcal K-\mathcal K_{4p}$
satisfies the bound
$$ \begin{align*}\sup_{{\mathbf{x}}\in I^k} \bigg|\frac{\partial^m}{\partial \beta_{\ell_1}\cdots\partial\beta_{\ell_m}}\bigg( \mathcal K({\mathbf{x}},\boldsymbol\beta)-\mathcal K_{4p}({\mathbf{x}},\boldsymbol\beta) \bigg)\bigg|\le C\bigg|\sum_{i=1}^k|\beta_i|\bigg|^{4p+1-m} ,\end{align*} $$
uniformly over
$\beta _1,...,\beta _m \in [-\eta /(2k),\eta /(2k)].$
Furthermore by Item (3) in Lemma 2.4 we have that
$$ \begin{align*}\sup_{{\mathbf{x}}\in I^k} \bigg|\frac{\partial^m}{\partial \beta_{\ell_1}\cdots\partial\beta_{\ell_m}}\bigg( \sum_{m=2p+1 }^{4p} \frac1{m!}&\sum_{\substack{1\leq j_1,...,j_m\leq k\\ j_i \text{ not all same}}}\beta_{j_1}\cdots \beta_{j_m} \kappa^{(m,k)} (\mathbf j; {\mathbf{x}})\bigg)\bigg| \\&\le C\bigg|\sum_{i=1}^k|\beta_i|\bigg|^{2p+1-m} \mathrm{F_{decay}}\big( \min_{1\le i<j\le k}|x_i-x_j|\big),\end{align*} $$
uniformly over
$\beta _1,...,\beta _m \in [-\eta /(2k),\eta /(2k)].$
On the other hand, since the joint cumulant of two random variables is just their covariance, the second bullet point in Lemma 2.4 (together with Definition 2.11) implies that when
$m=2p$
we have that
$$ \begin{align*}\frac1{m!}\sum_{\substack{1\leq j_1,...,j_m\leq k\\ j_i \text{ not all same}}}\beta_{j_1}\cdots \beta_{j_m} \kappa^{(m,k)} (\mathbf j; {\mathbf{x}}) = \sum_{1\le i<j\le k} \beta_i^p\beta_j^p\boldsymbol{\zeta} (x_i-x_j),\end{align*} $$
which is enough to complete the proof.
Corollary 2.13. Define functions
by
Here
$\mu $
is as defined in Assumption 1.2 Item (3). There exists
$C>0$
such that the following bounds hold uniformly over
$\beta \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}\in I^k$
and
$1\le i<j\le k$
:

Proof. Let
$\mathcal K({\mathbf {x}},\boldsymbol \beta )$
be as defined in (2.5) so that
Proposition 2.12 with
$m=1$
clearly implies that this quantity is bounded above by
thus proving the first bound (actually Proposition 2.12 seems to give a better bound with factor
$|\beta |^{2p}$
as opposed to
$|\beta |^{2p-1}$
, but this would ignore the term of order
$2p$
that had already been subtracted there, which we must take into account here). Next, note that

Proposition 2.12 with
$m=2$
clearly implies that this quantity is bounded above by
But the first bound readily implies that

with C independent of
${\mathbf {x}}\in I^k$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
. Together with the previous expression, this is enough to imply the second and third bounds.
2.2 Tightness estimates for the tilted processes
Here we prove tightness under diffusive scaling of various processes related to the path measures from Definition 2.8. These will be crucial for the remainder of the paper.
Definition 2.14. Let
$k \in \mathbb N$
and
$1\leq i\neq j \leq k$
. Let
$F:[0,\infty )\to \Bbb R_+$
be any decreasing function such that
$\sum _{k=1}^\infty F(k)<\infty $
. We define a process
$(V^{ij}(F;r))_{r\ge 0}$
on the underlying canonical probability space
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
by
$$ \begin{align} V^{ij}(F;r):= \sum_{s=0}^{r-1} F(|R^i_{s}-R^j_{s}|). \end{align} $$
Note that the processes
$V^{ij}(F;\bullet )$
are predictable with respect to the canonical filtration on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
.
These processes
$V^{ij}$
will play a crucial role going forward. The main results of this subsection will be to control the
$q^{th}$
moments of the additive functional processes
$V^{ij}$
and the difference processes
$R^i-R^j$
under the tilted measures
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
of Definition 2.8. This will be done in Theorem 2.18 and the subsequent corollaries, but first we will need three lemmas.
Lemma 2.15. Fix
$k\in \mathbb N$
and
$1\leq i,j\leq k$
. Let
$u_{\mathbf {x}}^{ij}({\mathbf {y}}):=|y_i-y_j-(x_i-x_j)|$
, for
${\mathbf {x}},{\mathbf {y}}\in I^k$
. Then there exists
$C>0$
such that uniformly over all
$\beta \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}},{\mathbf {y}}\in I^k$
one has
where
$\mathrm {F_{decay}}$
was defined in Assumption 1.2 Item (5), and
$\mathscr Q^\beta _k$
is the Markov operator from Corollary 2.10. Here
$\mathscr Q^\beta _k$
acts in the
${\mathbf {y}}$
variable.
Proof. For the upper bound, we have that
$$ \begin{align*} (\mathscr Q^\beta_k-\mathrm{Id}) u_{\mathbf{x}}^{ij}({\mathbf{y}})&= \int_{I^k} \bigg(\big|a_i-a_j-(x_i-x_j)\big|-\big|y_i-y_j-(x_i-x_j)\big|\bigg) \boldsymbol q^{(k)}_\beta ({\mathbf{y}},{\mathrm{d}}{\mathbf{a}}) \\& \leq \int_{I^k} |a_i-a_j - (y_i-y_j)| \boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}} {\mathbf{a}}). \end{align*} $$
From here the uniform bound follows from the uniform exponential moment bounds on the increments, see Lemma 2.9.
For the lower bound, use Jensen’s inequality to say
$$ \begin{align*} \mathscr Q^\beta_k u_{{\mathbf{x}}}({\mathbf{y}}) & = \int_{I^k} |a_i-a_j-(x_i-x_j)| \boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}}{\mathbf{a}}) \\&\ge \bigg| \int_{I^k} (a_i-a_j)\boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}}{\mathbf{a}}) - (x_i-x_j)\bigg| \\&\ge |y_i-y_j - (x_i-x_j)| - \bigg| y_i-y_j - \int_{I^k} (a_i-a_j)\boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}}{\mathbf{a}})\bigg| \\&= u_{{\mathbf{x}}}({\mathbf{y}}) - \bigg| y_i-y_j - \int_{I^k} (a_i-a_j)\boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}}{\mathbf{a}})\bigg|. \end{align*} $$
Thus we just need to show that
$$ \begin{align}\bigg| y_i-y_j - \int_{I^k} (a_i-a_j)\boldsymbol q^{(k)}_\beta({\mathbf{y}},{\mathrm{d}}{\mathbf{a}})\bigg|\leq C|\beta|^{4p} +C|\beta|^{2p-1}\mathrm{F_{decay}}\big(\min_{1\le i'<j'\le k}|y_{i'}-y_{j'}|\big). \end{align} $$
For this, first note that
$y_i-y_j = \int _{I^k} (a_i-a_j) \boldsymbol p^{(k)}({\mathbf {y}},{\mathrm {d}}{\mathbf {a}})$
where we recall that
$\boldsymbol p^{(k)} = \boldsymbol q_0^{(k)}$
. Thus use the definition (2.4) of
$\boldsymbol q^{(k)}_\beta $
and Taylor expand the exponentials to order
$4p-1$
around
$\beta =0$
, and note that all terms of order less than
$2p-1$
vanish identically thanks to the deterministic moments condition in Assumption 1.2 Item (4), and the fact that
$\mu ^{\otimes k}$
is a permutation-invariant measure. Then note that all terms of order between
$2p-1$
and
$4p-1$
contribute (at worst) a term of the form
$\mathrm {F_{decay}}\big ( \min _{1\le i<j\le k} |x_i-x_j|\big )$
thanks to the moment mixing condition in Assumption 1.2 Item (5). Finally note that the remainder in the Taylor expansion is at worst of order
$C|\beta |^{4p},$
which when combined with the exponential moment bounds of Lemma 2.9 completes the proof of (2.7).
Lemma 2.16. Recall
$\boldsymbol {p}_{\mathbf {dif}}$
from Assumption 1.2. For
$x\in I$
, we have that
$$ \begin{align}\bigg| 2- \bigg(\int_I a^2 \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da) -x^2\bigg) \bigg| \leq 2 \mathrm{F_{decay}}( |x|). \end{align} $$
Moreover, we have that
Proof. Note by the definition of
$\boldsymbol {p}_{\mathbf {dif}}$
that
$$ \begin{align*} \int_I a^2 \boldsymbol{p}_{\mathbf{dif}}(r,\mathrm da)-r^2 &=\int_I (a-r)^2 \boldsymbol{p}_{\mathbf{dif}}(r,\mathrm da) \\&= \int_{I^2} (x-r-y)^2 \boldsymbol p^{(2)}((r,0),({\mathrm{d}} x,{\mathrm{d}} y))\\&=2m_2-2\int_{I^2} (x-r)y\;\boldsymbol p^{(2)}((r,0),({\mathrm{d}} x,{\mathrm{d}} y)), \end{align*} $$
therefore we have that
$$ \begin{align*}\bigg| 2(m_2-m_1^2)- \bigg(\int_I a^2 \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da) -x^2\bigg) \bigg| = 2\bigg|\int_{I^2} (x-r)y\;\boldsymbol p^{(2)}((r,0),({\mathrm{d}} x,{\mathrm{d}} y))-m_1^2\bigg| \leq 2\mathrm{F_{decay}}(|x|), \end{align*} $$
where we used Assumption 1.2 Item (5). But
$m_2-m_1^2 =1$
by Assumption 1.2 Item (3), thus proving (2.8). To prove the second statement, note by the first bullet point of Assumption 1.2 Item (6) and the uniform exponential moment bounds of Lemma 2.9 that
$y\mapsto \int _I (y-a)^2 \boldsymbol {p}_{\mathbf {dif}}(y,{\mathrm {d}} a)$
is a continuous function of y. In particular it achieves its minimum on any compact interval. If the minimum was
$0$
, then
$\int _I (y_0-a)^2 \boldsymbol {p}_{\mathbf {dif}}(y_0,{\mathrm {d}} a)=0$
for some
$y\in I$
. Thus using the fact that
$\boldsymbol {p}_{\mathbf {dif}}(y_0,\cdot )$
has mean
$y_0$
, it would follow that
$y_0$
is an absorbing state for the chain. But this would contradict the second bullet point of Assumption 1.2 Item (6).
Lemma 2.17. Consider a martingale M defined on any filtered probability space
$(\Omega , \mathcal F, P),$
and suppose that the increments of M satisfy a uniform
$q^{th}$
moment bound under the probability measure P:
where
$q\ge 2$
. Then
$E [ |M_r-M_s|^q ] \leq C_q (r-s)^{\frac {q}2} \mathcal S_q(M;P),$
for some absolute constant
$C_q$
that is uniform over all martingales M and probability spaces
$(\Omega , \mathcal F,P).$
Proof. We may apply Burkholder-Davis-Gundy, then Jensen, then the definition of
$\mathcal S_q(M;P)$
in that order to obtain uniformly over
$r\ge s\ge 0$
$$ \begin{align*} E [ |M_r-M_s|^q ] &\leq C_q E \bigg[ \bigg( \sum_{n=s}^{r-1} (M_{n+1}-M_n)^2\bigg)^{q/2}\bigg] \\ &\leq C_q E \bigg[ (r-s)^{\frac{q}2 - 1} \sum_{n=s}^{r-1} |M_{n+1}-M_n|^q \bigg] \\ &\leq C_q (r-s)^{\frac{q}2} \mathcal S_q(M;P). \end{align*} $$
Since the constant
$C_q$
in Burkholder-Davis-Gundy is known to be univeral, the claim follows.
The above bound will be quite useful in showing tightness of all of the relevant processes in the topology of
$C[0,T]$
. The following estimate is the main result of this section.
Theorem 2.18. Fix
$k\in \mathbb N$
and
$q\ge 2$
and a decreasing function
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
. There exists some
$C=C(k,F ,q)>0$
and
$\beta _0=\beta _0(k,F,q)>0$
such that uniformly over all
$\beta \in [-\beta _0,\beta _0]$
, and all integers
$r\ge r'\ge 0$
and
$1\le i,j\le k$
we have the bounds
Before proving the theorem, let us give some intuition as to how the proof proceeds. Both processes will be shown to be martingales modulo some drift terms that will be controlled using Lemma 2.15. Establishing this martingality will use a sort of discrete-time version of the Itô-Tanaka formula.
Proof. By the Markov property and the fact that there is a supremum over
${\mathbf {x}}$
, it suffices to prove the claim when
$r'=0$
, so that
$R^i_{r'}-R^j_{r'} = x_i-x_j$
where
${\mathbf {x}}=(x_1,...,x_k)$
in coordinates. We can also assume without any loss of generality that
$F\geq \mathrm {F_{decay}}$
because otherwise one may replace F with
$\tilde F:=\max \{F,\mathrm {F_{decay}}\}$
, note that
$\tilde F$
is still decreasing (for large x) and summable, then prove the claim for
$\tilde F$
, and note that
$V^{ij}(\tilde F;r)- V^{ij}(\tilde F;r')\ge V^{ij}(F;r)-V^{ij}( F;r')$
to immediately obtain the claim for F.
We now break the proof into five steps.
Step 1. In this step we show that the Markov kernels
$\boldsymbol q^{(k)}_\beta $
satisfy a certain coercivity condition on test functions given by absolute values of differences of coordinates, specifically (2.11) below. This will be extremely useful later. For
${\mathbf {x}},{\mathbf {y}}\in I^k$
, let
so by Lemma 2.15 we have that
$v_{\mathbf {x}}^\beta ({\mathbf {y}})\ge -C|\beta |^{4p} -C|\beta |^{2p-1}\mathrm {F_{decay}}\big (\min _{1\le i'<j'\le k}|y_{i'}-y_{j'}|\big )$
.
We now show that
$\inf _{\beta \in [-\eta /(2k),\eta /(2k)]} \inf _{{\mathbf {x}}\in I^k} \mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {x}})>0.$
To prove this, first note that the measure
$\boldsymbol q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})$
is never supported on the set
$\{{\mathbf {y}}\in I^k:y_i-y_j=x_i-x_j\}$
, because this would then imply by mutual absolute continuity that
$\boldsymbol p^{(k)}({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})$
is also supported on that set. By Lemma 2.2, this would mean that
$\boldsymbol {p}_{\mathbf {dif}}(x_i-x_j,\cdot )$
would be a Dirac mass. But
$\boldsymbol {p}_{\mathbf {dif}}(x,\cdot )$
is never a Dirac mass, simply because it has mean x and Assumption 1.2 Item (6) implies that there cannot be absorbing states. Furthermore
$\mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {x}})$
is a continuous function of
${\mathbf {x}}$
, simply by the assumption of weak continuity of all of the kernels in the paper. So far, this argument shows that
$\inf _{\beta \in [-\eta /(2k),\eta /(2k)]} \inf _{{\mathbf {x}}\in I^k\cap K}\mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {x}})>0$
for all compact
$K\subset I^k$
. Now if we had that
$\mathscr Q^{\beta _n}_k u_{{\mathbf {x}}_n}({\mathbf {x}}_n)\to 0$
for some sequence of values
$|{\mathbf {x}}_n|\to \infty $
and
$\beta _n\in [-\eta /(2k),\eta /(2k)]$
, this would imply (by mutual absolute continuity and the uniform moment bounds of Lemma 2.9) that the measures
$\boldsymbol {p}_{\mathbf {dif}}(x^i_n-x^j_n,x^i_n-x^j_n+\cdot )$
would be converging weakly to a Dirac mass at
$0$
. But the exponential moments of these measures are uniformly bounded by Lemma 2.9 (applied with
$\beta =0$
). Thus the variances of
$\boldsymbol {p}_{\mathbf {dif}}(x^i_n-x^j_n,\cdot )$
would be converging to zero, which is impossible since Lemma 2.16 shows that they converge to
$2>0$
if
$|x^i_n-x^j_n|\to \infty $
. Thus we have shown that indeed
$\inf _{\beta \in [-\eta /(2k),\eta /(2k)]} \inf _{{\mathbf {x}}\in I^k} \mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {x}})>0.$
We now claim that there exists
$\delta>0$
so that for all
${\mathbf {x}},{\mathbf {y}}\in I^k$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
one has that
To prove this, note that
$$ \begin{align*} |\mathscr Q^\beta_k u_{\mathbf{x}}({\mathbf{y}}) - \mathscr Q^\beta_k u_{\mathbf{y}}({\mathbf{y}})| &= \bigg| \int_{I^k} |a_i-a_j-(x_i-x_j)|\boldsymbol q^{(k)}_\beta ({\mathbf{y}}, {\mathrm{d}}{\mathbf{a}})-\int_{I^k} |a_i-a_j-(y_i-y_j)|\boldsymbol q^{(k)}_\beta ({\mathbf{y}}, {\mathrm{d}}{\mathbf{a}})\bigg| \\ &\le \int_{I^k} \big| |a_i-a_j-(x_i-x_j)|-|a_i-a_j-(y_i-y_j)| \big| \boldsymbol q^{(k)}_\beta ({\mathbf{y}}, {\mathrm{d}}{\mathbf{a}}) \\ &\leq \int_{I^k}|x_i-x_j-(y_i-y_j)| \boldsymbol q^{(k)}_\beta ({\mathbf{y}}, {\mathrm{d}}{\mathbf{a}})=|x_i-x_j-(y_i-y_j)|. \end{align*} $$
Recall that
$c:=\inf _{\beta \in [-\eta /(2k),\eta /(2k)]} \inf _{{\mathbf {x}}\in I^k} \mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {x}})$
has already been shown to be strictly positive. Thus for all
${\mathbf {x}},{\mathbf {y}}\in I^k$
we have
$\mathscr Q^\beta _k u_{\mathbf {y}}({\mathbf {y}})\ge c$
, and from the above bound it then follows that
$\mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {y}})> 2c/3$
if
$|x_i-x_j-(y_i-y_j)|<c/3.$
Thus we see that
$v_{\mathbf {x}}^\beta ({\mathbf {y}}) = \mathscr Q^\beta _k u_{\mathbf {x}}({\mathbf {y}}) - |x_i-x_j-(y_i-y_j)|> c/3$
if
$|x_i-x_j-(y_i-y_j)|<c/3$
. Together with Lemma 2.15, this implies (2.11) with
$\delta :=c/3.$
Step 2. In this step we are going to find a useful family of martingales and derive a discrete “Tanaka’s formula,” see (2.13) below. That formula will be instrumental in the later analysis. We will also derive bounds related to these martingales, specifically (2.12), (2.14), and (2.15) below.
If
$I=c {\mathbb {Z}}$
we will henceforth assume without loss of generality that
$\delta =c$
in (2.11), otherwise it may be a smaller value. Let
$\mathbf e_i$
denote the standard basis of
${\mathbb {R}}^k$
. Equation (2.11) easily implies that for all
${\mathbf {x}}=(x_1,...,x_k)\in I^k$
one has
$$ \begin{align}F(|x_i-x_j|) &\leq \inf_{\beta\in[-\eta/(2k),\eta/(2k)]} \delta^{-1}\sum_{m\in {\mathbb{Z}}} F(\delta |m|)\big[ v^\beta_{\delta m (\mathbf e_i-\mathbf e_j)}({\mathbf{x}}) + C|\beta|^{4p}\nonumber\\ &\quad +C|\beta|^{2p-1}\mathrm{F_{decay}}\big(\min_{1\le i'<j'\le k}|x_{i'}-x_{j'}|\big)\big].\end{align} $$
For every
${\mathbf {x}},{\mathbf {y}}\in I^k$
, note that
$u_{\mathbf {x}}(\mathbf R_r)-\sum _{s=0}^{r-1} v_{\mathbf {x}}^\beta (\mathbf R_r)$
is a
${\mathbf {P}}^{(\beta ,k)}_{\mathbf {y}}$
-martingale by Corollary 2.10. Defining
$f_F,g_{\beta ,F}:I^k\to \Bbb R$
by
we see that the process
$$ \begin{align}\mathcal M_{\beta,F} (r):=-(f_F(\mathbf R_r)- f_F ({\mathbf{y}}))+\sum_{s=0}^{r-1}g_{\beta,F} (\mathbf R_s)\end{align} $$
is a
${\mathbf {P}}^{(\beta ,k)}_{\mathbf {y}}$
-martingale (for all
${\mathbf {y}}\in I^k)$
with
$\mathcal M_{\beta ,F} (0)=0$
. Note that f is globally Lipschitz, in fact
$$ \begin{align} |f_F ({\mathbf{x}})-f_F ({\mathbf{y}})| &\leq \bigg(\delta^{-1}\sum_{m\in {\mathbb{Z}}} F(\delta |m|) \bigg)|x_i-x_j-(y_i-y_j)|\leq C|x_i-x_j-(y_i-y_j)|,\end{align} $$
where we absorbed all constants in the last bound. We also claim that
$g_{\beta ,F}$
is a globally bounded function. Indeed, Lemma 2.15 gives that
$\Gamma := \sup _{\beta \in [-\eta /(2k),\eta /(2k)]} \sup _{{\mathbf {x}},{\mathbf {y}}\in I^k} v^\beta _{\mathbf {x}}({\mathbf {y}}) <\infty $
, thus we see that
$$ \begin{align}\sup_{\beta\in [-\eta/(2k),\eta/(2k)]}\sup_{{\mathbf{x}}\in I^k}|g_{\beta,F}({\mathbf{x}})| \leq \bigg(\delta^{-1}\sum_{m\in {\mathbb{Z}}} F(\delta |m|) \bigg) \cdot \Gamma <\infty. \end{align} $$
From here and (2.14) one sees that the increments of the martingales
$\mathcal M_{\beta ,F}$
have
$q^{th}$
moments bounded uniformly over
$\beta \in [-\eta /(2k),\eta /(2k)]$
and all times
$r\in {\mathbb {Z}}_{\ge 0}$
, thus by Lemma 2.17 one obtains that
where C does not depend on q.
Step 3. In this step we will show an a priori bound that the left side of (2.10) can actually be bounded above by some multiple of the left side of (2.9) plus
$C(r-r')^{1/2} +C|\beta |^{4p}r,$
which will effectively reduce the problem to showing the first bound. This bound is (2.18) below.
In (2.12), note that
$\mathrm {F_{decay}}\big (\min _{1\le i'<j'\le k}|x_{i'}-x_{j'}|\big ) \leq \sum _{1\le i'<j'\le k} \mathrm {F_{decay}}(|x_{i'}-x_{j'}|)$
. Furthermore, recall from the beginning of the proof our assumption (without loss of generality) that
$F\ge \mathrm {F_{decay}}$
, thus
$\mathrm {F_{decay}}(|x_{i'}-x_{j'}|) \leq F(|x_{i'}-x_{j'}|)$
. Thus from (2.12) we have the preliminary bound
$$ \begin{align}\notag \mathbf E_{\mathbf{x}}^{(\beta,k)} &[ V^{ij}(F;r)^q ]^{1/q} \\& \leq \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \bigg( \sum_{s=0}^{r-1} g_{\beta,F}(\mathbf R_s) \bigg)^q\bigg]^{1/q}+C|\beta|^{4p}r + C|\beta|^{2p-1} \sum_{1\le i'<j' \le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ V^{i'j'}(F;r)^q]^{1/q}. \end{align} $$
Using (2.13), (2.14), (2.15), and (2.16), we furthermore have
$$ \begin{align*} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \bigg( \sum_{s=0}^{r-1} g_{\beta,F}(\mathbf R_s) \bigg)^q\bigg]^{1/q} &\leq \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |\mathcal M_{\beta,F}(r)|^q]^{1/q}+ \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |f_F (\mathbf R_r) - f_F({\mathbf{x}})|^q ] ^{1/q} \notag \\& \leq Cr^{1/2} + C \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r-(x_i-x_j)|^q]^{1/q}. \end{align*} $$
Combining the last two expressions and summing over all indices
$(i,j)$
with
$1\le i<j\le k$
, then moving all instances of
$ \mathbf E_{\mathbf {x}}^{(\beta ,k)} [ V^{ij}(F;r)^q ]^{1/q}$
to the left side of the inequality, yields
$$ \begin{align*}\big(1- C|\beta|^{2p-1} \tfrac12k(k-1)\big)& \sum_{1\le i<j \le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ V^{ij}(F;r)^q ]^{1/q} \\ &\leq C r^{1/2} + C|\beta|^{4p}r + C \sum_{1\le i<j\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r-(x_i-x_j)|^q]^{1/q}. \end{align*} $$
Now choose
$\beta _0$
small enough so that
$C\beta _0^{2p-1} \tfrac 12k(k-1) \leq \frac 12.$
Then after absorbing all constants into some larger one, the last expression yields for
$|\beta |\leq \beta _0$
that
$$ \begin{align} \sum_{1\le i<j \le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ V^{ij}(F;r)^q ]^{1/q} \leq C r^{1/2} +C |\beta|^{4p}r + C \sum_{1\le i<j\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r-(x_i-x_j)|^q]^{1/q}. \end{align} $$
This bound will be a key step in proving the theorem shortly.
Step 4. In this step we will obtain a bound that is in a sense converse to (2.18). Recall from (2.7) that
$$ \begin{align} \bigg| \int_{I^k} \big( y_i-y_j - (x_i-x_j)\big) \boldsymbol q^{(k)}_\beta({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) \bigg| \leq C|\beta|^{2p-1} \mathrm{F_{decay}}\big(\min_{1\le i'<j'\le k} |x_{i'}-x_{j'}|\big) + |\beta|^{4p}\end{align} $$
uniformly over
$\beta \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}\in I^k.$
Let
$\pi _i:I^k\to I$
be the coordinate function
${\mathbf {x}}\mapsto x_i$
, and define the process
$$ \begin{align*}M^{i,j,\beta}_r:= R^i_r-R^j_r - (x_i-x_j)-\sum_{s=0}^{r-1} \big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s).\end{align*} $$
By Corollary 2.10, this is a
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
-martingale for all
${\mathbf {x}}\in I^k$
, which should not be confused with the Tanaka martingales appearing in (2.13), as they are unrelated. Now we can use Lemma 2.17, because (2.19) and Lemma 2.9 show that the increments of this martingale have uniformly bounded
$q^{th}$
moments. Thus we see that
$ \mathbf E_{\mathbf {x}}^{(\beta ,k)} [ |M^{i,j,\beta }_r|^q]^{1/q} \leq C r^{1/2},$
where C does not depend on integers
$r\ge 0$
. Thus we find that
$$ \begin{align} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q} \leq C r^{1/2} + \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \bigg( \sum_{s=0}^{r-1} \big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s)\bigg)^q\bigg]^{1/q}. \end{align} $$
Equation (2.19) precisely says
$|(\mathscr Q^\beta _k-\mathrm {Id}) (\pi _i-\pi _j)\big )(\mathbf R_s)| \leq C|\beta |^{2p-1} \mathrm {F_{decay}}\big ( \min _{1\le i<j\le k} |R^i_s-R^j_s|\big ) + |\beta |^{4p},$
which can be further bounded above by
$C|\beta |^{2p-1} \sum _{1\le i'<j'\le k}\mathrm {F_{decay}}\big ( |R^{i'}_s-R^{j'}_s|\big ) + |\beta |^{4p}$
because
$\mathrm {F_{decay}}$
is nonincreasing. Using the definition (2.6) of the processes
$V^{ij}$
, this precisely means that
$$ \begin{align*}\sum_{s=0}^{r-1} \big|\big(\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s) \big| \leq Cr|\beta|^{4p}+ C|\beta|^{2p-1} \sum_{1\le i'<j'\le k}V^{i'j'}(\mathrm{F_{decay}};r).\end{align*} $$
Plugging that bound into (2.20), and recalling
$F\ge \mathrm {F_{decay}}$
, we obtain
$$ \begin{align} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q} \leq C r^{1/2} + C|\beta|^{4p} r + C|\beta|^{2p-1} \sum_{1\le i'<j'\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ V^{i'j'}(F,r)^q]^{1/q}. \end{align} $$
uniformly over
$\beta \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}\in I^k.$
Step 5. In this step, we will establish the theorem. Apply (2.21) and then (2.18) in that order, and one finds that
$$ \begin{align*} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q} \leq C r^{1/2} + C|\beta|^{4p} r +C|\beta|^{2p-1} \sum_{1\le i'<j'\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^{i'}_r-R^{j'}_r - (x_i-x_j)|^q]^{1/q}. \end{align*} $$
Now sum the left side over all indices
$(i,j)$
with
$1\le i<j\le k$
, and move all instances of
$ \mathbf E_{\mathbf {x}}^{(\beta ,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q}$
to the left side of the expression, and we obtain
$$ \begin{align*}\big(1-\tfrac12 k(k-1) C |\beta|^{2p-1} \big) \sum_{1\le i<j\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q} \leq Cr^{1/2} + C|\beta|^{4p}r.\end{align*} $$
Now make
$\beta _0>0$
smaller (if needed) so that
$C \cdot \beta _0^{2p-1} \cdot \frac 12 k(k-1) \leq 1/2$
. Then for
$|\beta |\leq \beta _0$
one obtains from the previous expression (after absorbing constants on the right side into some larger constant) that
$$ \begin{align*}\sum_{1\le i<j\le k} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^i_r-R^j_r - (x_i-x_j)|^q]^{1/q} \leq Cr^{1/2} + C|\beta|^{4p}r.\end{align*} $$
This proves (2.9), and then (2.10) follows immediately from (2.9) and (2.18).
Remark 2.19. Note that (2.15) and (2.14) actually imply a stronger bound. Specifically if
$F:[0,\infty )\to [0,\infty )$
is decreasing and satisfies
$\sum _{k=0}^\infty F(k)<\infty $
, then uniformly over all
$0\le f\le F$
with f decreasing, and uniformly over
$r\ge 0$
and
$\beta \in [-\beta _0,\beta _0]$
, one has that
$$ \begin{align} \sup_{{\mathbf{x}}\in I^k} \;\mathbf E_{\mathbf{x}}^{(\beta,k)} \big[ V^{ij}(f;r)^q\big]^{1/q} \leq C \bigg(\sum_{n=0}^\infty f(k) \bigg)\big( r^{1/2}+|\beta|^{4p}r\big).\end{align} $$
The proof of this can be copied almost verbatim from Step 2 above, noting that for decreasing f one has
$\sum _{m\in \Bbb Z} f(\delta |m|) \leq 2\delta ^{-1} \sum _{m=0}^\infty f(m)$
.
Corollary 2.20. Fix
$k\in \mathbb N$
,
$T>0$
, and
$q\ge 1$
, and any decreasing function
$F:[0,\infty )\to [0,\infty )$
with
$\sum _{k=1}^\infty F(k)<\infty $
. Consider any sequence
$\beta _N$
in
$[-\eta /(2k),\eta /(2k)]$
such that
$|\beta _N|\leq MN^{-\frac 1{4p}}$
for some
$M>0$
. There exists some
$C=C(T,k,q,F,M)>0$
such that for all
$N\ge 1$
and
$0\le s\le t \le T$
with
$s,t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
one has
$$ \begin{align*}\sup_{{\mathbf{x}}\in I^k} \mathbf E_{\mathbf{x}}^{(\beta_N,k)} \big[\big|N^{-1/2}\big(R^i_{Nt} - R^j_{Nt} - (R^i_{Ns} -R^j_{Ns})\big)\big|^q\big]^{1/q} &\leq C\sqrt{t-s},\\ \sup_{{\mathbf{x}}\in I^k} \mathbf E_{\mathbf{x}}^{(\beta_N,k)} \big[\big|N^{-\frac12}\big(V^{ij}(F;Nt) - V^{ij}(F;Ns) \big)\big|^q\big]^{1/q} &\leq C\sqrt{t-s}. \end{align*} $$
Proof. This is immediate from (2.9) and (2.10). The extra term of the form
$C|\beta |^{4p}(r-r')$
on the right side of (2.9) and (2.10) will contribute
$C|t-s|$
which can be bounded above by
$C_T |t-s|^{1/2}$
on the bounded time interval
$[0,T].$
Corollary 2.21. Fix arbitrary
$\lambda ,t>0$
and any decreasing function
$F:[0,\infty )\to [0,\infty )$
with
$\sum _{k=1}^\infty F(k)<\infty $
. Consider any sequence
$\beta _N$
in
$[-\eta /(2k),\eta /(2k)]$
such that
$|\beta _N|\leq MN^{-\frac 1{4p}}$
for some
$M>0$
. Then
Proof. Note that (2.10) easily implies the bound that
uniformly over
$N\ge 1$
and
$r\le Nt.$
Since
$V^{ij}$
is an additive functional, one may iteratively apply the Markov property to automatically obtain higher moment bounds just from this first moment bound, specifically we will show that
where C is independent of
$m\in \mathbb N$
and
$r\le Nt$
. If we can prove (2.24), the claim will follow immediately simply by Taylor expanding the exponential, and setting
$r=Nt$
.
The proof of (2.24) follows from a straightforward induction on m. The
$m=1$
case is simply (2.23). For the inductive step, let us define
$\Omega _m$
to be the optimal constant so that the left side of (2.24) is bounded above by
$\Omega _m r^{m/2}$
uniformly over all
$r\ge 0$
. We then have
$$ \begin{align*} h_m(r;{\mathbf{x}})&:= \sum_{s_1,s_2,...,s_m=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \prod_{j=1}^{m} F(|R^i_{s_j}-R^j_{s_j}|) \bigg] \\&= \sum_{\substack{0\le s_1,s_2,...,s_m<r\\ \mathrm{no\;repeated\;index}}} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \prod_{j=1}^{m} F(|R^i_{s_j}-R^j_{s_j}|) \bigg] +\sum_{\substack{0\le s_1,s_2,...,s_m<r\\\mathrm{repeated\;index}}} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \prod_{j=1}^{m} F(|R^i_{s_j}-R^j_{s_j}|) \bigg] \\&\leq m! \cdot \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \sum_{0\leq s_1 < ...<s_m< r} \prod_{j=1}^{m} F(|R^i_{s_j}-R^j_{s_j}|) \bigg] + \|F\|_{\infty} h_{m-1}(r;{\mathbf{x}}) \\&= m! \cdot \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \sum_{s_1=0}^{r-1} F(|R^i_{s_1}-R^j_{s_1}|)\cdot \frac1{(m-1)!} h_{m-1}(r-s_1;\mathbf R_{s_1})] \bigg] + \|F\|_{\infty} h_{m-1}(r;{\mathbf{x}}) \\ &\le m\Omega_{m-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \sum_{s_1=0}^{r-1} F(|R^i_{s_1}-R^j_{s_1}|) \cdot (r-s_1)^{(m-1)/2}\bigg] + \|F\|_{\infty} h_{m-1}(r;{\mathbf{x}}) \end{align*} $$
We used the Markov property in the fourth line. Now for the first term on the right side, use summation by parts and we obtain that
$$ \begin{align*}\sum_{s=0}^{r-1} F(|R^i_{s}-R^j_{s}|)\cdot (r-s)^{(m-1)/2} = \sum_{s=0}^{r-1} \big[ (r-s)^{(m-1)/2}-(r-1-s)^{(m-1)/2}\big]\cdot \bigg(\sum_{t=0}^{s-1} F(|R^i_{t}-R^j_{t}|)\bigg).\end{align*} $$
Now note that
$(r-s)^{(m-1)/2}-(r-1-s)^{(m-1)/2}$
can be bounded above by
$C (\frac {m-1}{2}) (r-s)^{(m-3)/2} $
, then apply the expectation
$ \mathbf E_{\mathbf {x}}^{(\beta ,k)} $
over the last expression, then use the
$m=1$
case once again, and we see that the expectation is bounded above by
$C (\frac {m-1}{2})\cdot \sum _{s=0}^{r-1} (r-s)^{(m-3)/2}s^{1/2} $
which is equal to
$r^{m/2} $
multiplied by a Riemann sum approximation for
$\int _0^1 (1-u)^{(m-3)/2}u^{1/2}du $
. Notice that the latter is
$O(m^{-3/2})$
as
$m\to \infty .$
This whole discussion yields
$$ \begin{align*}\mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \sum_{s_1=0}^{r-1} F(|R^i_{s_1}-R^j_{s_1}|) \cdot (r-s_1)^{(m-1)/2}\bigg] \leq Cm^{-1/2}r^{m/2}\end{align*} $$
for some absolute constant C.
As for the second term
$ \|F\|_{\infty } h_{m-1}(r;{\mathbf {x}}) $
, we can absorb
$\|F\|_{\infty }$
into some absolute constant C since F is just a fixed function. By the inductive hypothesis,
$ \|F\|_{\infty } h_{m-1}(r;{\mathbf {x}}) $
may be bounded above by
$ C\Omega _{m-1} r^{(m-1)/2} \leq C \Omega _{m-1} r^{m/2}. $
The entire above discussion yields the relation
$\Omega _m \leq Cm^{1/2} \Omega _{m-1}$
for some absolute constant C independent of
$m\in \Bbb N,$
which easily implies (2.24).
The next corollary will allow us to control exponential terms of the form appearing in Lemma 2.6, and will be useful in Section 4.
Corollary 2.22. Let
$\mathbf R=(\mathbf R_r)_{r\ge 0}$
denote the canonical process on
$(I^{2k})^{{\mathbb {Z}}_{\ge 0}}$
. For
${\mathbf {x}}\in I^{2k}$
,
$\beta \in [-\eta /(2k),\eta /(2k)]$
, and
$r\in {\mathbb {Z}}_{\ge 0}$
define
$$ \begin{align*}\mathcal H^{2k}(\beta,{\mathbf{x}},\mathbf R,r):=\exp \bigg( \sum_{s=1}^r \bigg\{ \log \mathbf E_{{\mathbf{x}}}^{(0,2k)} \big[ e^{\beta\sum_{j=1}^{2k} (R^j_s -R^j_{s-1})} \big| \mathcal F_{s-1}\big] \; - 2k \log M(\beta)\bigg\}\bigg).\end{align*} $$
Then for all
$q\ge 1$
and
$t\ge 0$
we have the bound
$$ \begin{align*}\sup_{N\ge 1} \sup_{{\mathbf{x}}\in I^{2k}} \mathbf E_{\mathbf{x}}^{(N^{-\frac1{4p}} , 2k)} \big[ \sup_{r\leq Nt} \mathcal H^{2k}(N^{-\frac1{4p}},{\mathbf{x}},\mathbf R,r)^q\big]<\infty.\end{align*} $$
Furthermore, let
$T>0$
. There exists
$C>0$
such that uniformly over
$N\ge 1$
,
${\mathbf {x}}\in I^{2k},$
and
$s,t \in [0,T]\cap (N^{-1}{\mathbb {Z}})$
we have that
Proof. Let us use the bound
$$ \begin{align*}\sup_{r\leq Nt} \mathcal H^{2k}(N^{-\frac1{4p}},{\mathbf{x}},\mathbf R,r)^q &\leq \exp \bigg( q\sum_{s=1}^{Nt} \bigg| \log \mathbf E_{{\mathbf{x}}}^{(0,2k)} \big[ e^{N^{-\frac1{4p}}\sum_{j=1}^{2k} (R^j_s -R^j_{s-1})} \big| \mathcal F_{s-1}\big] \; - 2k \log M(N^{-\frac1{4p}})\bigg|\bigg)\\&=\exp \bigg( q\sum_{s=1}^{Nt} \bigg| \log \int_{I^k} e^{N^{-\frac1{4p}}(y_j - R^j_{s-1})} \boldsymbol p^{(k)} (\mathbf R_{s-1}, {\mathrm{d}} {\mathbf{y}}) \; - 2k \log M(N^{-\frac1{4p}})\bigg|\bigg), \end{align*} $$
where we applied the Markov property in the second line. Using Proposition 2.12 with
$m=0$
and all
$\beta _i = N^{-\frac 1{4p}}$
will show that the expression inside the absolute value can be bounded above by
$$ \begin{align}CN^{-1/2} \sum_{1\le i<j\le k} \mathrm{F_{decay}}\big( |R^i_r-R^j_r| \big) + CN^{-1-\frac1{4p}},\end{align} $$
where C is a large constant independent of N. An application of Corollary 2.21 will then immediately prove the first bound. The second bound can then be proved by first using the elementary bound
$|e^u-1|\leq |u|e^{|u|}$
, then applying Cauchy-Schwarz, and then applying the first bound in conjunction with Corollary 2.20 (noting the termwise bound (2.25) of each summand).
3 Convergence theorems for the tilted processes
In this section, we will prove various convergence theorems for the tilted Markov processes introduced in Definition 2.8. These convergence results (specifically Theorem 3.11 below) will build upon the estimates of the previous section, and they will be crucial to proving Theorem 1.5 in later sections. Due to many additional subtleties not present in the previous works [Reference Das, Drillick and ParekhDDP24a, Reference Das, Drillick and ParekhDDP24b], we will deviate completely from the methods of those papers and derive a new method of studying the tilted measures which leverages the Markov property of the tilted processes proved in Proposition 2.7 (whereas those papers did not use the Markov property).
Proposition 3.1. Fix
$k\in \mathbb N$
and
$T>0$
and
$1\leq i<j\leq k$
. Recall the measure
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
as given in Definition 2.8. Consider any pair of sequences
$\beta _N\in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}_N\in I^k$
, such that
$N^{\frac 1{4p}}\beta _N$
remains bounded and
$N^{-1/2}{\mathbf {x}}_N\to \boldsymbol x\in {\mathbb {R}}^k$
as
$N\to \infty $
. Then the rescaled and linearly interpolated processes
$\big ( N^{-1/2} (R^i_{Nt}-R^j_{Nt})\big )_{t \in (N^{-1} {\mathbb {Z}}_{\ge 0}) \cap [0,T]}$
, viewed as random variables under
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)},$
converge in law as
$N\to \infty $
to a Brownian motion of rate 2 started from
$x_i-x_j$
where
$\boldsymbol x=(x_1,...,x_k)$
. Convergence occurs with respect to the topology of
$C[0,T]$
.
We remark that a Brownian motion of rate 2 is defined as a process distributed as
$\sqrt 2 \cdot B$
for a standard Brownian motion B.
Proof. The tightness of this family of processes has already been established in Corollary 2.20. Thus we need to identify the limit points. For this we will use Levy’s criterion.
Let
$\pi _i:I^k\to I$
be the coordinate function
${\mathbf {x}}\mapsto x_i$
, and define the
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta _N,k)}$
-martingales
$$ \begin{align*}M^{i,j,\beta}_r:= R^i_r-R^j_r - (x_i-x_j)-\sum_{s=0}^{r-1} \big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s).\end{align*} $$
Recall from (2.7) that one has the pathwise bound
$$ \begin{align*}\sup_{r\le Nt} |M^{i,j,\beta}_r-( R^i_r-R^j_r - (x_i-x_j))\big| &\leq \sum_{s=0}^{Nt} \big|\big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s)\big| \\&\leq C|\beta|^{2p-1} \sum_{s=0}^{Nt} \mathrm{F_{decay}}\big( \min_{1\le i<j\le k} |R^i_s-R^j_s|\big) + |\beta|^{4p},\\& \leq C|\beta|^{2p-1} \sum_{1\le i'<j'\le k} \sum_{s=0}^{Nt} \mathrm{F_{decay}}\big( |R^{i'}_s-R^{j'}_s|\big) + |\beta|^{4p}. \end{align*} $$
Specialize to the case where
$\beta =\beta _N$
with
$|\beta _N|\leq CN^{-\frac 1{4p}}$
, and where and
${\mathbf {x}}={\mathbf {x}}_N$
, then multiply both sides by
$N^{-1/2}.$
Then using the result of Corollary 2.20, one finds that the processes
converge in probability to 0 in the topology of
$C[0,T]$
as
$N\to \infty $
, under the measures
${\mathbf {P}}^{(\beta _N,k)}_{{\mathbf {x}}_N}.$
In fact one has for all
$q\ge 1$
that
$$ \begin{align*}\lim_{N\to\infty}\mathbf E^{(\beta_N,k)}_{{\mathbf{x}}_N}\bigg[ \sup_{r\le Nt}\big(N^{-1/2} |M^{i,j,\beta}_r-( R^i_r-R^j_r - (x_i-x_j))\big|\big)^q\bigg]=0 .\end{align*} $$
Consequently, it suffices to show that the family of processes
$\big (N^{-1/2} M^{i,j,\beta _N}_{Nt}\big )_{t\in [0,T]}$
converge in law to a Brownian motion of rate 2 in
${\mathbb {R}}$
started from the origin, as
$N\to \infty $
viewed under the measures
${\mathbf {P}}^{(\beta _N,k)}_{{\mathbf {x}}_N}$
. Since martingality is preserved by limit points as long as one has uniform integrability, it is clear from the
$q^{th}$
moment bounds in the last expression and from Corollary 2.20 that any limit point in
$C[0,T]$
of the family
$\big (M^{i,j,\beta _N}_{Nt}\big )_{t\in [0,T]}$
must be a martingale under the canonical filtration of
$C[0,T]$
. Thus we only need to show that the quadratic variation is
$2t$
.
Define the process
$$ \begin{align*}Y_r^{i,j,\beta}:= (M_r^{i,j,\beta})^2 - \sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ (M_{s+1}^{i,j,\beta}-M_s^{i,j,\beta})^2 |\mathcal F_s]\end{align*} $$
which is also a
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
martingale. Notice just from the definition of these martingales that
$$ \begin{align*} \bigg|2r - \sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ (M_{s+1}^{i,j,\beta}&-M_s^{i,j,\beta})^2 |\mathcal F_s]\bigg| \leq \bigg| 2r - \sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} \big[ \big(R^i_{s+1} - R^j_{s+1} - (R^i_s - R^j_s)\big)^2 \big|\mathcal F_s\big]\bigg| \\& \;\;\; +2\sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \big|R^i_{s+1} - R^j_{s+1} - (R^i_s - R^j_s)\big| \cdot \big|\big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s)\big| \bigg|\mathcal F_s\bigg]\\&\;\;\; + \sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} \bigg[ \big|\big((\mathscr Q^\beta_k-\mathrm{Id}) (\pi_i-\pi_j)\big)(\mathbf R_s)\big| ^2 \bigg|\mathcal F_s\bigg] \end{align*} $$
We claim that all three terms on the right side can be bounded above by
$$ \begin{align*}C|\beta|r + C\sum_{s=0}^{r-1} \mathrm{F_{decay}}\big(\min_{1\le i'<j'\le k} |R_s^{i'}-R_s^{j'}|\big),\end{align*} $$
where the constant C is uniform over
$r\ge 0$
and
$\beta \in [-\eta /(2k),\eta /(2k)].$
Indeed, the first term on the right side is given by the absolute value of
$\sum _{s=0}^{r-1} g_\beta (\mathbf R_s)$
, where the function
$g:I^k\to {\mathbb {R}}$
is given by
We expand the square as
$(a_i-x_i)^2 + (a_j-x_j)^2 -2 (a_i-x_i)(a_j-x_j)$
, and from Corollary 2.13 one obtains that

For the second term on the right hand side, note that
$\big |\big ((\mathscr Q^\beta _k-\mathrm {Id}) (\pi _i-\pi _j)\big )(\mathbf R_s)\big |$
is actually
$\mathcal F_s$
-measurable, thus it can be pulled out of the conditional expectation. By Lemma 2.9, the remaining part
$ \mathbf E_{\mathbf {x}}^{(\beta ,k)} \big [ \big |R^i_{s+1} - R^j_{s+1} - (R^i_s - R^j_s)\big | \big |\mathcal F_s\big ]$
is deterministically bounded uniformly over
$\beta \in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}\in I^k$
, while the term
$\big |\big ((\mathscr Q^\beta _k-\mathrm {Id}) (\pi _i-\pi _j)\big )(\mathbf R_s)\big |$
itself can be bounded from above using (2.7), obtaining an upper bound of
$ C|\beta |^{4p} + C \mathrm {F_{decay}}\big (\min _{1\le i'<j'\le k} |R_s^{i'}-R_s^{j'}|\big ) .$
The third term can also be bounded from above by can be bounded from above using (2.7), obtaining the same upper bound of
$C|\beta |^{4p} + C \mathrm {F_{decay}}\big (\min _{1\le i'<j'\le k} |R_s^{i'}-R_s^{j'}|\big )$
.
Summarizing and specializing to the case where
$r=Nt$
and
$\beta =\beta _N$
, we see that
$$ \begin{align*}\bigg|2t - N^{-1}\sum_{s=0}^{Nt-1} \mathbf E_{\mathbf{x}}^{(\beta_N,k)} [ (M_{s+1}^{i,j,\beta}-M_s^{i,j,\beta})^2 |\mathcal F_s]\bigg| \leq CN^{-1}\cdot(Nt |\beta_N|)+N^{-1} \sum_{r=0}^{Nt} \mathrm{F_{decay}}\big( \min_{1\le i'<j'\le k} |R^{i'}_r - R^{j'}_r|\big) .\end{align*} $$
The latter tends to
$0$
in probability under
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)}$
in the topology of
$C[0,T]$
. Indeed from Corollary 2.20 and the elementary bound
$e^{-\min _\ell u_\ell } \leq \sum _\ell e^{-u_\ell }$
we can immediately deduce that
$$ \begin{align*}N^{-1} \sum_{r=0}^{Nt} \mathrm{F_{decay}}\big( \min_{1\le i'<j'\le k} |R^{i'}_r - R^{j'}_r|\big)\leq N^{-1} \sum_{r=0}^{Nt} \sum_{1\le i'<j'\le k} \mathrm{F_{decay}}\big( |R^{i'}_r - R^{j'}_r|\big) \stackrel{N\to \infty}{\longrightarrow} 0\end{align*} $$
in probability in the topology of
$C[0,T]$
under
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)}$
. This is because the multiplying factor
$N^{-1}$
here is converging to 0 faster than the fluctuation exponent
$N^{-1/2}$
appearing in Corollary 2.20.
We claim that
$(N^{-1} Y^{i,j,\beta _N}_{Nt})_{t\ge 0}$
is a tight family of processes in
$C[0,T]$
as
$N\to \infty $
. Indeed, one easily verifies that
$\big (N^{-1}(M^{j,\beta _N}_{Nt})^2\big )_{t\ge 0}$
is tight and satisfies the same
$L^q$
estimates as the processes
$\big (N^{-1/2}M^{j,\beta _N}_{Nt}\big )_{t\ge 0}$
on any compact time interval
$t\in [0,T]$
, simply because it is the square of a process satisfying such bounds as shown in Steps 2 and 3. On the other hand, by Minkowski’s inequality, we also have that
$$ \begin{align*}\bigg\| N^{-1} \sum_{r=Ns}^{Nt} \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} [ (M^{i,j,\beta_N}_{r+1}-M^{i,j,\beta_N}_r)^2|\mathcal F_s]\bigg\|_{L^q({\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)})}& \leq N^{-1} \sum_{r=Ns}^{Nt} \big\|M^{i,j,\beta_N}_{r+1}-M^{i,j,\beta_N}_r\big\|^{1/2}_{L^{2q}({\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)})}\\ &\leq \bigg(\sup_{N\ge 1} \mathcal S_{2q}(M^{i,j,\beta_N};{\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)})^{1/2} \bigg)|t-s|.\end{align*} $$
Summarizing these bounds, we have that for each
$q\ge 1$
we have
$E_{{\mathbf {x}}_N}^{(\beta _N,k)}\big [ \big (N^{-1}| Y^{j,\beta _N}_{Nt}- Y^{i,j,\beta _N}_{Ns}|\big )^q\big ]^{1/q} \leq C|t-s|^{1/2}$
for all
$s,t$
in a compact interval, thus completing the proof that
$(N^{-1} Y^{j,\beta _N}_{Nt})_{t\ge 0}$
is a tight family of processes in
$C[0,T]$
as
$N\to \infty $
.
Thus if we take a joint limit point
$(\mathcal X,\mathcal Y)$
as
$N\to \infty $
of the processes
$\big ( N^{-1/2}M^{i,j,\beta _N}_{Nt}, N^{-1} Y^{i,j,\beta _N}_{Nt}\big )_{t\ge 0} $
, then it must hold that
$\mathcal Y_t = \mathcal X_t^2 - 2t$
and moreover both processes are continuous martingales, since martingality is preserved by limit points as long as uniform integrability holds. By Levy’s criterion, we conclude that
$\mathcal X$
must be a Brownian motion of rate 2 starting at
$x_i-x_j$
since the starting values of
$N^{-1/2}X_{Nt}$
converge to
$\boldsymbol x$
, that is,
$N^{-1/2}{\mathbf {x}}_N\to \boldsymbol x$
.
While the convergence of the process itself has been established in Proposition 3.1, it is much more subtle to prove convergence of additive functionals of the process to the local time of the Brownian motion. This will be needed to prove the main results, and this will be the main focus going forward.
3.1 Analysis of the annealed difference process
In this subsection we will prove some results for the annealed difference process, that is, the Markov chain on I with transition density
$\boldsymbol {p}_{\mathbf {dif}}$
as defined in Assumption 1.2 Item (6). By Lemma 2.2 this Markov chain equals the averaged law of the difference of two particles in the random environment, that is, the law of the process
$R^i-R^j$
under
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
with
$\beta =0$
(the untilted case). These untilted results will be crucial in deriving the generalizations to the tilted case later.
Definition 3.2. Recall the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}(x,A):= \int _{I^2} \mathbf {1}_{\{a-b\in A\}} \boldsymbol p^{(2)} \big ((x,0),(\mathrm da,\mathrm db)\big )$
from Assumption 1.2 Item (6). We also define the associated Markov operator
for all measurable functions
$f:I\to {\mathbb {R}}$
for which the integral is absolutely convergent.
Definition 3.3. Throughout this subsection, we will let
$(X_r)_{r\ge 0}$
denote the canonical process on the space
$I^{{\mathbb {Z}}_{\ge 0}}$
. We will let
${\mathbf {P}}_x$
denote the probability measure on
$I^{{\mathbb {Z}}_{\ge 0}}$
given by the law of the Markov chain associated to
$\boldsymbol {p}_{\mathbf {dif}}$
started from
$x\in I.$
Thus by Lemma 2.2, the canonical process
$(X_r)_{r\ge 0}$
under
${\mathbf {P}}_x$
is distributed the same as the process
$R^i-R^j$
under
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
with
$\beta =0$
and
$x_i-x_j=x.$
Proposition 3.4 (Existence of a Foster-Lyapunov drift function)
There exists a function
$V:I\to {\mathbb {R}}_+$
with compact sublevel sets such that
$P_{\mathbf {dif}} V(x)\leq V(x)$
for all x outside some compact interval.
We remark that in some references such as [Reference Meyn and TweedieMT93], any function V as above would be said to satisfy the (V1) drift condition (as opposed to (V2) or (V3) which will not be relevant in this paper).
Proof. We set
$V(x):= \sqrt {|x|+1}.$
We will show that
$P_{\mathbf {dif}} V(x)\leq V(x)$
for all positive and sufficiently large x, and the proof for negative x would be completely symmetric.
For
$x>2$
and
$a\in (x/2,3x/2)$
we can write
$$ \begin{align*}V(a)-V(x) &= V'(x) (a-x) + \frac12 V"(x) (a-x)^2 +R(a,x) \\ &= \tfrac12 (x+1)^{-1/2} (a-x) -\frac14 (x+1)^{-3/2} (a-x)^2 + R(a,x), \end{align*} $$
where by Taylor’s remainder theorem one has uniformly over all
$x>2$
and
$a\in (x/2,3x/2)$
the bound
$$ \begin{align*}|R(a,x)| \leq \frac16 |a-x|^3 \sup_{b\in [x/2,3x/2]}|V"'(b)| \le \frac3{48} (x/2)^{-5/2} |a-x|^3 .\end{align*} $$
Now recall that
$\int _I (a-x)\boldsymbol {p}_{\mathbf {dif}}(x,{\mathrm {d}} a) = 0$
for all
$x\in I$
, which is immediate from the definitions, see, for example, Lemma 2.2. Consequently we can disregard the first-order term when calculating
$P_{\mathbf {dif}} V$
and we find that
$$ \begin{align}\notag P_{\mathbf{dif}} V(x)-V(x) &= \int_I (V(a)-V(x)) \boldsymbol{p}_{\mathbf{dif}} (x,{\mathrm{d}} a) \\ \notag &\leq -\frac14 (x+1)^{-3/2}\int_I (a-x)^2 \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a) + \frac3{48}(x/2)^{-5/2} \int_I |a-x|^3 \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a) \\ &\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;+ \int_I |V(a)-V(x)| \mathbf{1}_{\{|x-a|>x/2\}} \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a).\end{align} $$
Let
$\eta $
be as in Assumption 1.2 Item (3), and let us define constants
$$ \begin{align*} \omega_{exp}: &= \sup_{y\in I} \int_I e^{\frac\eta4|a-y|} \boldsymbol{p}_{\mathbf{dif}}(y,{\mathrm{d}} a),\\ \omega_2:&= \sup_{y\in I} \int_I (a-y)^2 \boldsymbol{p}_{\mathbf{dif}}(y,{\mathrm{d}} a),\\ \omega_3 :&= \sup_{y\in I} \int_I |a-y|^3 \boldsymbol{p}_{\mathbf{dif}}(y,{\mathrm{d}} a), \\ \delta_2 :&= \inf_{y\in I} \int_I (a-y)^2 \boldsymbol{p}_{\mathbf{dif}}(y,{\mathrm{d}} a). \end{align*} $$
The first three constants are all finite, for example, by Lemma 2.9, and the last constant is strictly positive as noted in Lemma 2.16. Notice that V is a globally Lipschitz function on I with Lipschitz constant 1/2, that is,
$|V(x)-V(a)|\leq \frac 12 |x-a|,$
thus using Cauchy-Schwarz and Markov’s inequality we see that
$$ \begin{align*} \int_I |V(a)-V(x)| &\mathbf{1}_{\{|x-a|>x/2\}} \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a) \\&\leq \bigg(\int_I (V(a)-V(x))^2\boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a)\bigg)^{1/2} \bigg(\int_I \mathbf{1}_{\{|x-a|>x/2\}} \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a)\bigg)^{1/2} \\ &\leq \bigg(\frac14 \int_I (a-x)^2\boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a)\bigg)^{1/2} \bigg(\frac{\int_I e^{\frac\eta4|x-a|} \boldsymbol{p}_{\mathbf{dif}}(x,{\mathrm{d}} a)}{e^{\frac\eta4(x/2)}}\bigg)^{1/2} \\ &\leq \frac12 \omega_2^{1/2} \omega_{exp}^{1/2}\cdot e^{-\frac{\eta x}{16}}. \end{align*} $$
Plugging this bound back into (3.1), we find that for
$x>2$
one has
The right side is clearly negative for sufficiently large
$x>2$
, thus proving the claim.
Theorem 3.5. Let
$\boldsymbol {p}_{\mathbf {dif}}$
be the Markov kernel from Assumption 1.2 Item (6), and
$P_{\mathbf {dif}}$
its associated Markov operator as in Definition 3.2. Then there exists a unique (up to scalar multiple) invariant measure
$\pi ^{\mathrm {inv}}$
on I, in other words a measure such that
$P_{\mathbf {dif}} f\in L^1(\pi ^{\mathrm {inv}})$
whenever
$f\in L^1(\pi ^{\mathrm {inv}})$
, and moreover
$\int _I (P_{\mathbf {dif}} f-f)d\pi ^{\mathrm {inv}} = 0$
for all
$f\in L^1(\pi ^{\mathrm {inv}})$
. Furthermore,
-
•
$\pi ^{\mathrm {inv}}(K)<\infty $
for all compact subsets
$K\subset I.$
-
•
$\pi ^{\mathrm {inv}}$
has full support on I. -
• For all
$f,g \in L^1(\pi ^{\mathrm {inv}})$
such that
$\int _I g\;d\pi ^{\mathrm {inv}} \neq 0$
we have that
$$ \begin{align*}\lim_{n\to \infty} \frac{\sum_{r=1}^n f(X_k)}{\sum_{r=1}^n g(X_k)} = \frac{\int_I f\;{\mathrm{d}}\pi^{\mathrm{inv}}}{\int_I g\;{\mathrm{d}}\pi^{\mathrm{inv}}},\;\;\;\; {\mathbf{P}}_x\;a.s. \;\;\;\;\; for\;all\; x\in I.\end{align*} $$
Proof. By the first bullet point in Assumption 1.2 Item (6), the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
is Strong Feller, that is, the operator
$P_{\mathbf {dif}}$
sends bounded measurable functions to bounded continuous functions. In particular
$\boldsymbol {p}_{\mathbf {dif}}$
is a “T-chain” in the sense of [Reference Meyn and TweedieMT93, Definition 6.0.0 (iii)].
Furthermore
$\boldsymbol {p}_{\mathbf {dif}}$
is “open-set irreducible” in the sense of [Reference Meyn and TweedieMT93, Section 6.1.2] by the second bullet point of Assumption 1.2 Item (6). In particular, by [Reference Meyn and TweedieMT93, Proposition 6.1.5] the chain is “
$\psi $
-irreducible” where the maximal irreduciblilty measure
$\psi $
has full support due to the open-set irreducibility and the Strong Feller property.
Furthermore, by Proposition 3.4 and [Reference Meyn and TweedieMT93, Theorem 9.4.1], for all
$x\in I$
, there exists a compact set
$K\subset I$
such that the Markov chain
$(X_r)_{r\ge 0}$
visits K infinitely often
${\mathbf {P}}_x$
-a.s.. In other words, the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
is “nonevanescent” in the sense of [Reference Meyn and TweedieMT93, Section 9.2.1]. Now by [Reference Meyn and TweedieMT93, Theorem 9.2.2 (ii)] any nonevanescent
$\psi $
-irreducible T-chain is automatically Harris recurrent.
By [Reference Meyn and TweedieMT93, Theorem 17.3.2], Harris recurrent chains can be characterized as exactly those chains for which a unique invariant measure
$\pi ^{\mathrm {inv}}$
exists (up to scalar multiple) and satisfies the ratio limit theorem as in the third bullet point of the theorem statement. By [Reference Meyn and TweedieMT93, Theorem 10.4.9], the measure
$\pi ^{\mathrm {inv}}$
is equivalent to the maximal irreducibility measure
$\psi $
, which has already been explained to have full support. By [Reference Meyn and TweedieMT93, Theorem 12.3.3], the invariant measure
$\pi ^{\mathrm {inv}}$
is finite on compact sets as long as the Foster-Lyapunov drift criterion is satisfied as we have already shown in Proposition 3.4.
Lemma 3.6. Let
$\eta $
be as in Assumption 1.2 Item (3) and let
$C>0$
. There exists
$C'>0$
such that uniformly over all
$f:I\to {\mathbb {R}}$
and
$b,x\in I$
and
$\theta \in [0,\frac \eta 2 ]$
the following bound holds true.
-
• If
$|f(x)-|x-b||\leq C e^{-\theta |x-b|}$
, then
$|(P_{\mathbf {dif}}f-f)(x)| \leq C' \theta ^{-1} e^{-\theta |x-b|}$
where
$C'$
is a larger constant depending on C but not
$f,x,b,\theta $
.
Proof. For simplicity of notation we will prove the claim with
$b=0$
but the proof for general b is similar, recentering around b rather than
$0$
everywhere in the argument below.
By the triangle inequality, we can write
$$ \begin{align*}|P_{\mathbf{dif}} f(x)-f(x)| &= |\int_I (f(a)-f(x)) \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)| \\ &\leq \int_I |f(a)-|a|| \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da) + \bigg|\int_I (|a|-|x|)\boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)\bigg| + ||x|-|f(x)|| \\ &\leq C \int_I e^{-\theta|a|} \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da) + \bigg|\int_I (|a|-|x|)\boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)\bigg| + C e^{-\theta |x|}. \end{align*} $$
Let us call these terms
$E_1,E_2,E_3$
. Note that
$E_3$
already satisfies the required bound. Since
$-|a| \leq |x-a| - |x|,$
we can bound
$E_1$
by
$$ \begin{align*}C \bigg( \sup_{x'\in I}\sup_{\theta'\in[0,\frac\eta2 ]}\int_I e^{\theta |x'-a|} \boldsymbol{p}_{\mathbf{dif}}(x',\mathrm da)\bigg)\cdot e^{-\theta|x|},\end{align*} $$
and the supremum is finite by the assumption
$\theta <\eta /2.$
This is indeed a bound of the desired form after replacing
$C $
by a larger constant.
Finally we need to bound
$E_2$
. Without loss of generality we assume
$x \ge 0$
, because the case
$x\le 0$
is symmetric. Then we have
$|x| = x = \int _I a \;\boldsymbol {p}_{\mathbf {dif}}(x,\mathrm da),$
since every
$\boldsymbol {p}_{\mathbf {dif}}(x',\cdot )$
is centered at
$x'$
by construction. Thus we have that
$$ \begin{align*} E_2 &= \bigg| \int_I (|a|-a) \boldsymbol{p}_{\mathbf{dif}} (x,\mathrm da)\bigg| = 2 \int_I |a|\mathbf{1}_{\{a\le 0\}} \boldsymbol{p}_{\mathbf{dif}} (x,\mathrm da) \end{align*} $$
Note that for
$a\le 0$
and
$x\ge 0$
one has
We can thus bound
$E_2$
by
$$ \begin{align*}C\bigg( \sup_{x'\in I}\sup_{\theta'\in[0,\frac\eta2 ]}\int_I e^{\theta |x'-a|} \boldsymbol{p}_{\mathbf{dif}}^{r'}(x',\mathrm da)\bigg)\theta^{-1} e^{-\theta |x|},\end{align*} $$
which is again finite by the previous lemma, as explained before. This proves the lemma.
Proposition 3.7. let
$x_N$
be any sequence of values in I such that
$N^{-1/2}x_N\to x\in {\mathbb {R}}$
as
$N\to \infty $
. Consider the rescaled process
$\big (N^{-1/2}X_{Nt}\big )_{t\in (N^{-1}{\mathbb {Z}}_{\ge 0}) \cap [0,T]},$
viewed as random variables in
$C[0,T]$
under the measure
${\mathbf {P}}_{x_N}$
. This is a tight sequence of processes in
$C[0,T]$
and any limit point is given by the law of a Brownian motion of rate 2 starting at x.
Proof. Take
$\beta _N=0$
in Proposition 3.1.
Proposition 3.8. Let
$\pi ^{\mathrm {inv}}$
be the invariant measure constructed in Theorem 3.5, with any fixed scalar multiple. For all decreasing functions
$F:[0,\infty )\to [0,\infty )$
satisfying
$\sum _{k=0}^\infty F(k) <\infty $
, one has
$$ \begin{align*}\int_I F(|x|) \pi^{\mathrm{inv}}({\mathrm{d}} x) \leq C\sum_{k=0}^\infty F(k),\end{align*} $$
where C is some absolute constant that does not depend on
$F.$
Proof. We break the proof into several steps.
Step 1. Let
$u_b(y):=|y-b|$
, and
$v_b:=P_{\mathbf {dif}} u_b-u_b$
. In this step we will prove that for every fixed
$x,b\in I$
, the process
$N^{-1/2}\sum _{r=0}^{Nt} v_b(X_r)$
converges in law under
${\mathbf {P}}_x$
to the process
$L_0^W$
, where the latter denotes the local time at zero of a Brownian motion W of rate 2 started from
$0$
.
Fix
$T>0$
and
$q\ge 1$
and
$b\in I$
. We claim that there exists some
$C>0$
such that for
$N\ge 1$
and
$0\le s\le t \le T$
with
$s,t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
and
$q\ge 1$
one has
$$ \begin{align*}\sup_{x\in I} \mathbf E_x \bigg[\bigg|N^{-1/2}\sum_{r=Ns}^{Nt} v_b(X_r)\bigg|^q\bigg]^{1/q} \leq C(b)\sqrt{t-s}.\end{align*} $$
Indeed, this is immediate from applying Corollary 2.20 with
$\beta =0$
, since we have exponential decay of
$v_b$
by Lemma 3.6 with
$\theta =\eta /2$
. The tightness of the pair of processes is thus immediate from this bound. Therefore we just need to show how to identify the limit points. To do this we will use martingale problems. We already know from Proposition 3.7 that
$(N^{-1/2}X_{Nt})_{t\ge 0}$
is converging to a Brownian motion.
Using Corollary 2.10 with
$\beta =0$
, one sees that the process
$Z_r:=u_b(X_r)- \sum _{s=0}^{r-1} v_b(X_s)$
is a
${\mathbf {P}}_x$
-martingale for every
$x\in I$
. Consider any joint limit point
$(\mathcal X,\mathcal U,\mathcal V)$
of the 3-tuple of processes
$$ \begin{align*}\big(N^{-1/2}X_{Nt},N^{-1/2} u_b(X_{Nt}), N^{-1/2} \sum_{s=0}^{Nt-1} v_b(X_s) \big)_{t\ge 0}.\end{align*} $$
Tightness of the first and third coordinates has already been explained, and the second coordinate is tight by Lemma 3.6 and Corollary 2.20, since Lemma 3.6 with
$\theta =\eta /2$
guarantees that
$|v(x)|\leq C e^{-\frac 2{\eta }|x|}$
. We have already shown that
$\mathcal X$
must be distributed as W for a Brownian motion W. For fixed
$b\in I$
, the functions
$x\mapsto N^{-1/2}u_b(N^{1/2}x)$
are converging uniformly on compacts as
$N\to \infty $
to absolute value, thus we must have
$\mathcal U=|\mathcal X|=|W|$
. Again using the fact that martingality is preserved by limit points, we see that
$\mathcal U- \mathcal V = |W|-\mathcal V$
must be a martingale for the limit point (with respect to the canonical filtration on the space of 3-tuples of continuous paths), which by Tanaka’s formula forces
$\mathcal V = L_0^W,$
where
$L_0^W$
is the local time.
Step 2. In this step we show that
$v_b\in L^1(\pi ^{\mathrm {inv}})$
for every
$b\in I$
, and moreover
$\int _I v_b \; {\mathrm {d}} \pi ^{\mathrm {inv}}$
does not depend on
$b.$
We first show that
$v_0\in L^1(\pi ^{\mathrm {inv}})$
, that is, the case
$b=0$
. Suppose it was not the case, we will now derive a contradiction. By convexity of
$u_0$
and Jensen’s inequality, we know that
$v_0$
is nonnegative, thus
$\int _I v_0\; {\mathrm {d}}\pi ^{\mathrm {inv}} = +\infty .$
On the other hand by the first bullet point of Theorem 3.5 we have that
$\mathbf {1}_{[-J,J]}\in L^1(\pi ^{\mathrm {inv}})$
for all
$J>0$
. Then by the third bullet point of Theorem 3.5 one easily shows that for all
$J>0$
that
$$ \begin{align*}\lim_{N\to \infty} \frac{\sum_{r=0}^N \mathbf{1}_{[-J,J]}(X_r) }{\sum_{r=0}^N v_0(X_r)} = 0\end{align*} $$
${\mathbf {P}}_x$
-a.s. for all
$x\in I$
. By the result of Step 1, this means that for all
$J>0$
we have that
$$ \begin{align*}N^{-1/2} \sum_{r=0}^N \mathbf{1}_{[-J,J]} (X_r) = \frac{\sum_{r=0}^N \mathbf{1}_{[-J,J]}(X_r) }{\sum_{r=0}^N v_0(X_r)} \cdot N^{-1/2} \sum_{r=0}^N v_0(X_r) \;\;\;\stackrel{N\to\infty}{\longrightarrow}\;\;\; 0 \cdot L_0^W(1) = 0,\end{align*} $$
in distribution (hence in probability) under every
${\mathbf {P}}_x$
. On the other hand by Corollary 2.20 with
$\beta =0$
and
$q=2$
we know that for all (fixed)
$J>0$
$$ \begin{align*}\sup_{N\ge 1} \sup_{x\in I} \mathbf E_x \bigg[ \bigg(N^{-1/2} \sum_{r=0}^N \mathbf{1}_{[-J,J]} (X_r)\bigg)^2\bigg]<\infty,\end{align*} $$
thus by uniform integrability and the previous convergence statement, for all
$x\in I$
one has
$$ \begin{align*}\lim_{N\to \infty} \mathbf E_x \bigg[ N^{-1/2} \sum_{r=0}^N \mathbf{1}_{[-J,J]} (X_r)\bigg]=0, \;\;\; for\;all\;\; J>0.\end{align*} $$
Define
$g_N(J):= \mathbf E_0 \big [ N^{-1/2} \sum _{r=0}^N \mathbf {1}_{[-J,J]} (X_r)\big ],$
so we have shown that
$\lim _{N\to \infty }g_N(J)=0$
for all
$J>0.$
For any
$\theta>0 $
note that
$\mathbf {1}_{[-J,J]}(x) \leq e^{\theta J/2} e^{-\theta |x|/2},$
consequently we find that
$$ \begin{align*}\sup_{N\ge 1} g_N(J) \leq e^{\theta J/2} \sup_{N\ge 1} \mathbf E_0 \bigg[ N^{-1/2} \sum_{r=0}^N e^{-\theta |X_r|/2} \bigg] = C(\theta) e^{\theta J/2}\end{align*} $$
where we know that the second supremum is finite by Corollary 2.20. Thus by Fubini’s theorem and the Dominated Convergence Theorem we see that for all fixed
$\theta>0$
$$ \begin{align*}\lim_{N\to \infty} \mathbf E_0 \bigg[ N^{-1/2}\sum_{r=0}^N \bigg( \sum_{J=1}^\infty e^{-\theta J}\mathbf{1}_{[-J,J]} (X_r)\bigg) \bigg] &= \lim_{N\to\infty} \sum_{J=1}^\infty e^{-\theta J} g_N(J) =\sum_{J=1}^\infty e^{-\theta J}\lim_{N\to\infty} g_N(J) = 0. \end{align*} $$
On the other hand, we know that
$v_0$
decays exponentially fast at infinity thanks to Lemma 3.6, which means that there exists some
$C,\theta>0$
such that
$v_0 \leq C \sum _{J=1}^\infty e^{-\theta J}\mathbf {1}_{[-J,J]}$
. Thus the last expression implies that
$N^{-1/2}\sum _{r=0}^N v_0(X_r)$
converges to zero in
$L^1({\mathbf {P}}_0)$
, contradicting the result of Step 1 and thus completing the proof that
$v_0\in L^1(\pi ^{\mathrm {inv}})$
.
It remains to show that
$v_b\in L^1(\pi ^{\mathrm {inv}})$
and
$\int _I v_b \;{\mathrm {d}} \pi ^{\mathrm {inv}} = \int _I v_0 \;{\mathrm {d}} \pi ^{\mathrm {inv}}$
, for all
$b\in I$
. First note that
$\int _I v_0 \;{\mathrm {d}} \pi ^{\mathrm {inv}}\neq 0$
. Indeed
$v_0\ge 0$
and
$\pi ^{\mathrm {inv}}$
has full support by Theorem 3.5. So if the integral vanishes, then
$v_0=0$
which would imply that
$|X_r|$
is a nonnegative martingale, thus convergent, contradicting the result of Proposition 3.7. From the result of Step 1 we know that for every
$b\in I$
$$ \begin{align*}\frac{\sum_{r=0}^{N} v_b(X_r)}{\sum_{r=0}^{N} v_0(X_r)} \;\;\; \stackrel{N\to\infty}{\longrightarrow} \;\;\; \frac{L_0^W(1)}{L_0^W(1)} =1\end{align*} $$
in distribution (hence in probability) under every
${\mathbf {P}}_x$
. Using the third bullet point of Theorem 3.5, it is clear that this would be impossible unless
$\int _I v_b \;{\mathrm {d}} \pi ^{\mathrm {inv}}<\infty $
and
$\int _I v_b \;{\mathrm {d}} \pi ^{\mathrm {inv}} = \int _I v_0 \;{\mathrm {d}} \pi ^{\mathrm {inv}}$
, for all
$b\in I$
.
Step 3. In this step we will finally prove the claim being made in the proposition statement. Let F be as in the theorem statement. The argument will proceed very similarly to the proof of (2.10), but we repeat the details in the present context for additional clarity.
We first show that
$\inf _{b\in I} P_{\mathbf {dif}} u_b(b)>0.$
To prove this, note that
$\boldsymbol {p}_{\mathbf {dif}}(b,\cdot )$
is never a Dirac mass, simply because it has mean x and Assumption 1.2 Item (6) implies that there cannot be absorbing states. Consequently
$P_{\mathbf {dif}} u_b(b)= \int _I |b-a|\boldsymbol {p}_{\mathbf {dif}}(b,\mathrm da)>0$
for all
$b\in I$
. Furthermore
$P_{\mathbf {dif}} u_b(b)$
is a continuous function of b, simply by the assumption of weak continuity of all of the kernels in the paper. If we had that
$u_{b_n}(b_n)\to 0$
for some
$|b_n|\to \infty $
, this would mean that the measures
$\boldsymbol {p}_{\mathbf {dif}}(b_n,b_n+\cdot )$
would be converging weakly to a Dirac mass at
$0$
. But the exponential moments of these measures are uniformly bounded by Lemma 2.9. Thus the variances of
$\boldsymbol {p}_{\mathbf {dif}}(x_n,\cdot )$
would be converging to zero, which is impossible since we have already observed in (2.8) that they converge to
$2>0$
.
We now claim that there exists
$\delta>0$
so that for all
$b\in I$
one has that
$v_b>\delta \mathbf {1}_{(b-\delta ,b+\delta )}$
. To prove this, note that
$$ \begin{align*} |P_{\mathbf{dif}} u_b(y) - P_{\mathbf{dif}} u_y(y)| &= \bigg| \int_I |a-b|\boldsymbol{p}_{\mathbf{dif}} (y, \mathrm da)-\int_I |a-y|\boldsymbol{p}_{\mathbf{dif}} (y, \mathrm da)\bigg| \\ &\le \int_I \big| |a-b|-|a-y| \big| \boldsymbol{p}_{\mathbf{dif}} (y, \mathrm da) \\ &\leq \int_I|b-y| \boldsymbol{p}_{\mathbf{dif}} (y, \mathrm da)=|b-y|. \end{align*} $$
Recall that
$c:= \inf _{y\in I} P_{\mathbf {dif}} u_y(y)$
has already been shown to be strictly positive. Thus for all
$b,y\in I$
we have
$P_{\mathbf {dif}} u_y(y)>c$
, and from the above bound it then follows that
$P_{\mathbf {dif}} u_b(y)> 2c/3$
if
$|b-y|<c/3.$
Thus we see that
$v_x(y) = P_{\mathbf {dif}} u_b(y) - |b-y|> c/3$
if
$|b-y|<c/3$
, thus proving the claim with
$\delta :=c/3.$
If
$I=c {\mathbb {Z}}$
we will henceforth assume that
$\delta =c$
, otherwise it may be a smaller value. Using the fact that
$v_b>\delta \mathbf {1}_{(b-\delta ,b+\delta )}$
, we obtain
Integrating both sides with respect to
$\pi ^{\mathrm {inv}}$
and then using the result of Step 2, we find that
$$ \begin{align*}\int_I h\;{\mathrm{d}}\pi^{\mathrm{inv}} \leq \delta^{-1}\sum_{m\in {\mathbb{Z}}} F(\delta |m|) \int_I v_{\delta m}\;{\mathrm{d}}\pi^{\mathrm{inv}} = \bigg( \delta^{-1}\sum_{m\in {\mathbb{Z}}} F(\delta|m|)\bigg)\int_I v_0\;{\mathrm{d}}\pi^{\mathrm{inv}},\end{align*} $$
and the sum is bounded above by
$2\delta ^{-1} \sum _{k=0}^\infty F(k)$
.
3.2 Analysis and convergence results for nonzero
$\beta $
With all of the necessary prerequisites established for the untilted
$\beta =0$
case, we will now move on to the case where the tilting coefficient
$\beta $
will be nonzero, as will be needed to prove convergence to the KPZ equation. The following lemma will be a crucial step in proving convergence of the additive functional process to Brownian local time under a varying tilt.
Lemma 3.9. Fix
$k\in \mathbb N$
and
$t>0$
and
$1\leq i<j\leq k$
. Recall the tilted path measures
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
as given in Definition 2.8. Consider any sequences
$\beta _N\in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}_N\in I^k$
such that
$|\beta _N|\leq CN^{-\frac 1{4p}}$
and
$N^{-1/2}{\mathbf {x}}_N\to \boldsymbol x\in {\mathbb {R}}^k$
as
$N\to \infty $
. Take any continuous function
$f:I\to {\mathbb {R}}$
such that
$|f(x)| \leq F(|x|)$
for some decreasing
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
. Then we have that
$$ \begin{align*}\lim_{N\to\infty} \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ N^{-1/2} \sum_{r=0}^{Nt-1} f(R_r^i-R_r^j)\bigg]=\gamma(f)\mathbf E_{\mathrm{BM}}^{x_i-x_j}[L_0^W(t)]\end{align*} $$
where the latter denotes expectation with respect to a Brownian motion of rate 2 started from
$x_i-x_j$
where
$\boldsymbol x=(x_1,...,x_k)$
. Furthermore
$L_0^W$
is its local time at zero, and
$$ \begin{align*}\gamma(f):= \frac{\int_I f(x)\pi^{\mathrm{inv}}(dx)}{\int_I \big[\int_I |a|\boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)-|x|\big] \pi^{\mathrm{inv}}(dx)}.\end{align*} $$
Note by Lemma 2.6 that we are ultimately interested in the case
$\beta _N=N^{-\frac 1{4p}}$
, but there is no additional difficulty in considering
$|\beta _N|\leq CN^{-\frac 1{4p}}$
at the moment.
Proof.
Step 1. We first establish the claim in the very special case that
$f=(P_{\mathbf {dif}} - \mathrm {Id})u$
where
$u(x) = |x|.$
In this case it is clear that
$\gamma (f)=1$
. Letting
$\overline u({\mathbf {x}}):= |x_i-x_j|$
and
$\overline f_\beta ({\mathbf {x}}):= (\mathscr Q^\beta _k -\mathrm {Id})\overline u$
, both of which are functions on
$I^k$
, we claim that
for some large enough constant
$C>0$
and some decreasing function
$F:[0,\infty )\to [0,\infty )$
satisfying
$\sum _{k=0}^{\infty } F(k)<\infty $
. To prove (3.2), we first prove an intermediate claim that
$$ \begin{align} \bigg| \int_{I^k} |y_i-y_j| \boldsymbol q^{(k)}_\beta({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) -\mathrm{sign}(x_i-x_j) \int_{I^k} (y_i-y_j) \boldsymbol q^{(k)}_\beta({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) \bigg| \leq Ce^{-\frac\eta{8k} |x_i-x_j| }. \end{align} $$
To prove the latter bound, it suffices by symmetry to consider the case
$x_i-x_j\ge 0$
. In this case, the integral appearing on the left is bounded above by
$\int _{I^k} |y_i-y_j| \mathbf {1}_{\{y_i-y_j\le 0\}} \boldsymbol q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}}), $
which by Cauchy-Schwarz can be bounded from above by
$\big (\int _{I^k} (y_i-y_j)^2 \boldsymbol q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})\big )^{1/2}\big (\int _{I^k} \mathbf {1}_{\{y_i-y_j\le 0\}} \boldsymbol q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}})\big )^{1/2}. $
Using Lemma 2.9, the first term can be bounded above by
$(C+C(x_i-x_j)^2)^{1/2},$
with C uniform over
$\beta $
and
${\mathbf {x}}$
. For the second term, note that the distance of the point
${\mathbf {x}}$
to the set
$\{{\mathbf {y}}:y_i-y_j \le 0\}$
is
$x_i-x_j$
, thus by the uniform exponential moment bounds of Lemma 2.9 and Markov’s inequality we have an upper bound of
$Ce^{-\frac {\eta }{2k} |x_i-x_j|}.$
Finally note that by making the constant larger, one has
$(C+C(x_i-x_j)^2)^{1/2} e^{-\frac {\eta }{2k} |x_i-x_j|} \leq Ce^{-\frac {\eta }{8k} |x_i-x_j|},$
thus establishing (3.3).
Note that (3.3) reduces proving (3.2) to showing that
$$ \begin{align*}\bigg| \int_{I^k} (y_i-y_j) \boldsymbol q^{(k)}_\beta({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) - \int_{I^k} (y_i-y_j) \boldsymbol p^{(k)} ({\mathbf{x}},{\mathrm{d}}{\mathbf{y}}) \bigg| \leq C|\beta|^{2p-1} \mathrm{F_{decay}}\big(\min_{1\le i'<j'\le k} |x_{i'}-x_{j'}| \big) + C|\beta|^{4p}.\end{align*} $$
But this is clear simply by performing a Taylor expansion of
$\boldsymbol q^{(k)}_\beta $
from (2.4) of order
$4p-1$
in the variable
$\beta $
, then applying Items (4) and (5) of Assumption 1.2.
With (3.2) established, use the bound
$F\big (\min _{1\le i'<j'\le k} |x_{i'}-x_{j'}| \big ) \leq \sum _{1\le i'<j'\le k} F( |x_{i'}-x_{j'}| ),$
and we find that
$$ \begin{align*}&\mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ N^{-1/2} \sum_{r=0}^{Nt-1} |f(R_r^i-R_r^j)-\overline f_{\beta_N} (\mathbf R_r)|\bigg]\\ &\quad \leq \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ N^{-1/2} \sum_{1\le i'<j'\le k} \sum_{r=0}^{Nt-1} \big(|\beta_N|^{2p-1} F(|R^i_r-R^j_r|)+|\beta_N|^{4p}\big)\bigg].\end{align*} $$
Using
$|\beta _N|\leq CN^{-\frac 1{4p}}$
, we may use the second bound in Corollary 2.20 to immediately conclude that the above expectation tends to zero. Thus (recalling
$\gamma (f)=1$
) in order to complete this step, it suffices to prove that
$$ \begin{align} \lim_{N\to\infty}\mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ N^{-1/2} \sum_{r=0}^{Nt-1} \overline f_{\beta_N}(\mathbf R_r)\bigg] = \mathbf E_{\mathrm{BM}}^{x_i-x_j}[L_0^W(t)]. \end{align} $$
But by Corollary 2.10 we know that
$|R^i_r-R^j_r| - \sum _{s=0}^{r-1} \overline f_\beta (\mathbf R_s)$
is a
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
-martingale, thus the expectation on the left side is precisely
$$ \begin{align*}\mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ N^{-1/2}\big| R^i_{Nt} -R^j_{Nt}|\bigg] - N^{-1/2}|x_N^i-x_N^j|,\end{align*} $$
where
${\mathbf {x}}_N:= (x^1_N,...,x^k_N)$
in coordinates. By Proposition 3.1 and the uniform
$L^q$
bounds in Corollary 2.20, we immediately have that the last expectation converges to
By Tanaka’s formula, this is precisely
$\mathbf E_{\mathrm {BM}}^{x_i-x_j}[L_0^W(t)],$
thus establishing (3.4) and proving the claim for this special case.
Step 2. Now we consider the case of general f. For this we use a Krylov-Bogoliubov type of trick. Fixing
$t>0$
henceforth, consider the sequence of measures
$\gamma _N$
on I given by
$$ \begin{align*}\int_{I} f \;{\mathrm{d}} \gamma_N:= N^{-1/2} \sum_{s=0}^{Nt-1}\mathbf E^{(\beta_N,k)}_{{\mathbf{x}}_N} [ f(R^i_s-R^j_s)].\end{align*} $$
We will show that any subsequence
$\gamma _{N_k}$
of this sequence of measures has a further subsequence
$\gamma _{N_{k_j}}$
converging as
$N\to \infty $
to an invariant measure
$\gamma $
for the Markov kernel
$\boldsymbol {p}_{\mathbf {dif}}$
. By convergence, we mean that for all continuous
$f:I\to {\mathbb {R}}$
of exponential decay, we have
$\int _I f\;{\mathrm {d}}\gamma _N \to \int _I f\;{\mathrm {d}}\gamma $
along this subsequence
$\gamma _{N_{k_j}}$
.
If we can prove this, then the lemma would be proved for all functions f of exponential decay, because by the uniqueness in Theorem 3.5, this would mean that the subsequential limit
$\gamma $
must be a constant multiple of
$\pi ^{\mathrm {inv}}$
. But the result of Step 1 uniquely identifies the constant as the one appearing in the lemma statement (the expectation of the Brownian local time). If every subsequence of
$\gamma _N$
has a further subsequence converging to some fixed measure, then the sequence
$\gamma _N$
must itself converge to that fixed measure, thus completing the proof.
Thus consider a subsequence
$\gamma _{N_k}$
. There is a further convergent subsequence
$\gamma _{N_{k_j}}$
simply by the second bound in Corollary 2.20 (with
$q=1$
) combined with, for example, Banach-Alaoglu (to extract subsequential limits on every compact set, then applying a diagonal argument). Let us call this subsequential limit
$\gamma $
. We need to show that for all continuous
$f:I\to {\mathbb {R}}$
of exponential decay at infinity, one has
$\int _I P_{\mathbf {dif}} f \;{\mathrm {d}}\gamma = \int _I f\;{\mathrm {d}}\gamma .$
To prove this, consider such f, say
$|f(x)| \leq e^{-\theta |x|}$
where
$\theta \leq \eta /(2k)$
, and define
$\overline f: I^k\to {\mathbb {R}}$
by
$\overline f({\mathbf {x}}) = f(x_i-x_j)$
. To show that
$\int _I P_{\mathbf {dif}} f \;{\mathrm {d}}\gamma = \int _I f\;{\mathrm {d}}\gamma ,$
it suffices to show that
$$ \begin{align}\lim_{N\to\infty}N^{-1/2} \sum_{s=0}^{Nt-1} \mathbf E^{(\beta_N,k)}_{{\mathbf{x}}_N}[ |(\mathscr Q^0_k-\mathscr Q^{\beta_N}_k) \overline f(\mathbf R_s)|] = 0, \end{align} $$
where
$\mathscr Q^0_k$
denotes the operator
$\mathscr Q_k^\beta $
with
$\beta =0$
(these operators were defined in Corollary 2.10). Indeed, it suffices to show this because
$\mathscr Q^0_k\overline f(\mathbf R_s) = P_{\mathbf {dif}} f(R^i_s-R^j_s)$
so that
$N^{-1/2} \sum _{s=0}^{Nt-1}\mathbf E^{(\beta _N,k)}_{{\mathbf {x}}_N} [ \mathscr Q^0_k\overline f(\mathbf R_s)] $
converges to
$\int _I P_{\mathbf {dif}} f \;{\mathrm {d}}\gamma $
along the subsequence
$N_{k_j}$
, and moreover because
$N^{-1/2} \sum _{s=0}^{Nt-1}\mathbf E^{(\beta _N,k)}_{{\mathbf {x}}_N} [ \mathscr Q^{\beta _N}_k \overline f(\mathbf R_s)]$
by the Markov property is equal to
$\int _I f{\mathrm {d}}\gamma _N + N^{-1/2} \big ( \mathbf E^{(\beta _N,k)}_{{\mathbf {x}}_N}[f(R^i_{Nt} -R^j_{Nt})] - f(x^N_i-x^N_j)\big ),$
which converges to
$\int _I f\;{\mathrm {d}}\gamma $
along the subsequence
$N_{k_j}$
.
To prove (3.5), we claim by the exponential decay assumption on f and from the definition (2.4) of the measures
$\boldsymbol q^{(k)}_\beta $
, one has the bound
The second inequality is immediate from the first one using
$\min \{u,v\}\leq u^{1/2}v^{1/2}$
. To prove the first inequality, the upper bound of
$C|\beta |$
is clear simply from differentiability of the kernels
$\boldsymbol q^{(k)}_\beta $
in the variable
$\beta $
near
$\beta =0$
, see their explicit expression (2.4). For the upper bound of
$Ce^{-\frac \theta {2} |x_i-x_j|}$
, we can show that both of the terms
$\mathscr Q^{\beta }_k \overline f({\mathbf {x}})$
and
$\mathscr Q^{0}_k) \overline f({\mathbf {x}})$
are bounded above by such a quantity. To show this, write the definition
$\mathscr Q^\beta _k \overline f({\mathbf {x}}):= \int _{I^k} f(y_i-y_j) \boldsymbol q^{(k)}_\beta ({\mathbf {x}},{\mathrm {d}}{\mathbf {y}}), $
then split the integral into two parts:
$\{{\mathbf {y}}: |{\mathbf {y}}-{\mathbf {x}}| \leq \frac 12|x_i-x_j|\} $
and
$\{{\mathbf {y}}: |{\mathbf {y}}-{\mathbf {x}}|> \frac 12|x_i-x_j|\} $
, where
$|{\mathbf {x}}|=\sum _{j=1}^k |x_j|$
. On the first set, use the exponential decay bound on f to bound
$|f({\mathbf {y}}) |\leq Ce^{-\frac \theta 2|x_i-x_j|}$
. On the second set, use the uniform moment bounds of Lemma 2.9 and Markov’s inequality (and
$|f|\leq 1$
) to obtain an upper bound of
$Ce^{-\frac \eta {4k} |x_i-x_j|} \leq Ce^{-\frac \theta 2 |x_i-x_j|}.$
With (3.6) proved, the claim (3.5) follows immediately from the second bound of Corollary 2.20. While this proves the claim for all f of exponential decay, the general claim for all f as in the theorem statement can be proved using the uniform bound (2.22) (only
$q=1$
is needed there) and an approximation argument of f by some sequence
$f_n$
with
$|f_n|\leq |f|$
with each
$|f_n|$
decaying exponentially.
Proposition 3.10. Fix
$k\in \mathbb N$
and
$T>0$
and
$1\leq i<j\leq k$
. Recall the tilted path measure
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
as given in Definition 2.8. Consider any sequences
$\beta _N\in [-\eta /(2k),\eta /(2k)]$
and
${\mathbf {x}}_N\in I^k$
such that
$|\beta _N|\leq CN^{-\frac 1{4p}}$
and
$N^{-1/2}{\mathbf {x}}_N\to \boldsymbol x\in {\mathbb {R}}^k$
as
$N\to \infty $
. Take any continuous function
$f:I\to {\mathbb {R}}$
such that
$|f(x)| \leq F(|x|)$
for some decreasing
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
. Consider the pair of processes
$$ \begin{align*}\bigg( N^{-1/2} \big(R^i_{Nt} -R^j_{Nt}\big), N^{-1/2} \sum_{r=0}^{Nt-1} f(R_r^i-R_r^j)\bigg)_{t\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,T]},\end{align*} $$
viewed as
$C([0,T],{\mathbb {R}}^2)$
-valued random variables under the measures
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)}.$
This sequence converges in law to the pair
$(W,\gamma (f)L_0^W(t))$
, where W is a Brownian motion of rate 2 starting at
$x_i-x_j$
and
$\boldsymbol x=(x_1,...,x_k)$
. Furthermore
$L_0^W$
is its local time at zero, and
$$ \begin{align*}\gamma(f):= \frac{\int_I f(x)\pi^{\mathrm{inv}}({\mathrm{d}} x)}{\int_I \big[|x| - \int_I |a|\boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)\big] \pi^{\mathrm{inv}}({\mathrm{d}} x)}.\end{align*} $$
Proof. Convergence of the first coordinate was already established in Proposition 3.1. Tightness of the second coordinate is immediate from the second bound in Corollary 2.20. We will thus show how to jointly identify the limit of the second coordinate. Without loss of generality we can assume
$f\ge 0$
, since we can always write
$f=f_+-f_-$
with
$f_+,f_-\ge 0$
, both of which still satisfy the same exponential bounds, and the result respects the linearity in the function variable.
Consider any joint limit point
${\mathbf {P}}_{\mathrm {lim}}$
, which is a probability measure on the canonical space
$C([0,T],{\mathbb {R}}^2)$
. Let
$(W,\mathcal L)$
denote the canonical process on that space, and let
$(\mathcal F_{\mathrm {lim}}(t))_{t\ge 0}$
be the canonical filtration on that space. On one hand
$|W_t|-|x_i-x_j|-L_0^W(t)= \int _0^t \mathrm {sign}(W_s)dW_s $
is a
${\mathbf {P}}_{\mathrm {lim}}$
-martingale, because W is a
${\mathbf {P}}_{\mathrm {lim}}$
-Brownian motion as we already verified in Proposition 3.1. In particular this implies that for
$t\ge s\ge 0$
,
where
$g(t,x):= \mathbf E^x_{\mathrm {BM}}[L_0^B(t)],$
and the expectation is with respect to a Brownian motion B of rate 2 started from
$x \in {\mathbb {R}}$
. So far we have only studied the first part W of the joint limit. Now we study the other part
$\mathcal L$
. Immediately from the Markov property of the prelimiting processes and the result of Lemma 3.9, we find that for
$t\ge s\ge 0$
Combining (3.7) and (3.8), we see that
$\mathcal L-\gamma (f)L_0^W$
is a continuous
${\mathbf {P}}_{\mathrm {lim}}$
-martingale starting from 0. But it is also a difference of two increasing processes, thus of bounded variation
${\mathbf {P}}_{\mathrm {lim}}$
-almost surely. We conclude that
$\mathcal L-\gamma (f)L_0^W=0$
, proving the theorem.
We now come to the result that will be most crucial in proving convergence of the quenched density field to the KPZ equation, namely the invariance principle for the tilted Markov chains
$\boldsymbol q^{(k)}_\beta $
.
Theorem 3.11. Fix any
$k\in \mathbb N$
and constants
$C, T>0$
. Take any sequence
${\mathbf {x}}_N\in I^k$
such that
$N^{-1/2} {\mathbf {x}}_N\to \boldsymbol x\in {\mathbb {R}}^k$
. Recall the drift constant
$d_N$
that was defined in Equation (1.3) of the introduction. Fix any continuous function
$f:I\to {\mathbb {R}}$
such that
$|f(x)| \leq F(|x|)$
for some decreasing
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
. Let
$\mathbf {R}=(\mathbf {R}(r))_{r\ge 0}$
be the canonical process on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
, and define the rescaled processes
$$ \begin{align*} {\mathbf X}_N(t):= \frac{\mathbf R(Nt)-d_N t}{\sqrt{N}},\;\;\;\;\;\;\;\;\mathscr V^{ij}_N(f;t):=\frac{\sum_{r=1}^{Nt} f(R^i_r-R^j_r)}{\sqrt{N}}, \end{align*} $$
where these expressions are valid for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
, and understood to be linearly interpolated for
$t \notin N^{-1}{\mathbb {Z}}_{\ge 0}$
. Now recall the measure
${\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
as given in Definition 2.8. Then the processes
$(\mathbf X_N, \mathscr V_N)$
under
${\mathbf {P}}_{{\mathbf {x}}_N}^{(N^{-\frac 1{4p}},k)} $
converge in law with respect to the topology of
$C([0,T],{\mathbb {R}}^k\times {\mathbb {R}}^{k(k-1)/2})$
to
$(\mathbf U, (\gamma (f) \cdot L^{ij})_{1\le i<j\le k})$
where
-
•
$\mathbf U=(U^1,...,U^k)$
is a standard k-dimensional Brownian motion starting from
$\boldsymbol x$
. -
•
$L^{ij} = L_0^{U^i- U^j}$
denotes the pairwise local time of the
$i^{th}$
and
$j^{th}$
coordinates. -
•
$\gamma (f)$
is a real number given by
$$ \begin{align*}\gamma(f):= \frac{\int_{I} f(x) \pi^{\mathrm{inv}}({\mathrm{d}} x)}{\int_{I} \big[ \int_{I} |a| \; \boldsymbol{p}_{\mathbf{dif}}(x,\mathrm da)-|x|\big]\pi^{\mathrm{inv}}({\mathrm{d}} x)}.\end{align*} $$
Before the proof, we remark that the tilting strength of
$N^{-\frac 1{4p}}$
can be replaced by any sequence
$\beta _N$
satisfying
$|\beta _N|\leq CN^{-\frac 1{4p}}$
but then the constants
$d_N$
need to be replaced by
where
was defined in Corollary 2.13. This will be clear from the proof.
Proof. We break the proof into several steps. Throughout the proof we will set
$\beta _N:= N^{-\frac 1{4p}}$
.
Step 1. In this step we find a useful family of martingales, associated to each of the coordinates of the process. Define the
$j^{th}$
coordinate function
$v_j: I^k \to {\mathbb {R}}$
by
$(x_1,...,x_k)\mapsto x_j.$
Note by Corollary 2.10 that for each
$1\le j \le k$
that the process
$$ \begin{align*}M^{j,\beta}_r := R^j_r - \sum_{s=0}^{r-1} (\mathscr Q^\beta_k-\mathrm{Id})v_j(\mathbf R_s)\end{align*} $$
is a
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
-martingale for all
${\mathbf {x}}\in I^k$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
. Explicitly we have that
with
$\boldsymbol q^{(k)}_\beta $
as defined in (2.4).
Step 2. Fix
$j\in \{1,...,k\}$
, and let
$R^j$
as usual denote the
$j^{th}$
coordinate of the canonical process
$\mathbf R$
on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
. Recall the quantity
$\mathcal S_q(M,P)$
from Lemma 2.17. In this step we will show that for each
$q\ge 1$
,
First note by (3.9) and Lemma 2.9 that
$$ \begin{align}\sup_{\beta\in[-\eta/(2k),\eta/(2k)]}\sup_{{\mathbf{x}}\in I^k} |(\mathscr Q^\beta_k -\mathrm{Id})v_j({\mathbf{x}})| \leq \sup_{\beta\in[-\eta/(2k),\eta/(2k)]} \sup_{{\mathbf{x}}\in I^k} \int_{I^k} |y_j-x_j| \boldsymbol q^{(k)}_\beta({\mathbf{x}},{\mathrm{d}}{\mathbf{y}})<\infty.\end{align} $$
Then note by the Markov property and another application of Lemma 2.9 that
$$ \begin{align}\sup_{r\ge 1} \sup_{{\mathbf{x}}\in I^k} \sup_{\beta\in[-\eta/(2k),\eta/(2k)]} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ |R^j_r-R^j_{r-1}|^q] \leq \sup_{{\mathbf{x}}'\in I^k} \sup_{\beta\in [-\eta/(2k),\eta/(2k)]} \int_{I^k} |y_j-x_j|^q \boldsymbol q^{(k)}_\beta({\mathbf{x}}',{\mathrm{d}}{\mathbf{y}})<\infty .\end{align} $$
Combining (3.11) and (3.12), we immediately obtain (3.10).
Step 3. In this step we study the processes given by
$$ \begin{align*}\mathcal B_N(t):= N^{-1/2}d_Nt - N^{-1/2}\sum_{r=0}^{Nt-1} (\mathscr Q^{\beta_N}_k -\mathrm{Id})v_j(\mathbf R_r).\end{align*} $$
In particular we will show that if
$q\ge 1$
then
$\mathbf E_{{\mathbf {x}}_N}^{(\beta _N,k)} [ |\mathcal B_N(t)-\mathcal B_N(s)|^q]^{1/q} \leq CN^{-(\frac 12-\frac 1{4p})} |t-s|^{1/2},$
so that
$\mathcal B_N$
converges to the zero process in the topology of
$C[0,T]$
. To prove this, notice that in the notation of Corollary 2.13, we have
. Consequently by that corollary we have that for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$

where
$t=Nr$
and we used the fact that
$\beta _N=N^{-\frac 1{4p}}$
in the last equality. Thus using the bound
$\mathrm {F_{decay}}(\min _\ell u_\ell ) \leq \sum _\ell \mathrm {F_{decay}}(u_\ell )$
we find that
$$ \begin{align*} |\mathcal B_N(t)-\mathcal B_N(s)| &\leq \sum_{u \in [s,t]\cap N^{-1}{\mathbb{Z}}_{\ge 0}} |\mathcal B_N(u+N^{-1})-\mathcal B_N(u)| \\ &\leq \bigg(N^{-1+\frac1{4p}}\sum_{r \in [Ns,Nt]\cap{\mathbb{Z}}_{\ge 0}} \sum_{1\le i'<j'\le k} \mathrm{F_{decay}}(|R^{i'}_r-R^{j'}_r| )\bigg) \;\;\;\; + N^{-1/2} (t-s). \end{align*} $$
Thus taking expectation and applying the second estimate of Corollary 2.20 we find that
Step 4. Combining the results of (3.10) and (3.13) and Lemma 2.17, we immediately obtain tightness of the rescaled process
$\mathbf X_N$
in the space
$C[0,T]$
. It remains to identify the limit point as Brownian motion. In this step, we will begin by showing that any limit point
$\mathbf U = (U^1,...,U^k)$
is necessarily a martingale in the joint filtration of all coordinates, and moreover each coordinate
$U^j$
is individually distributed as a standard Brownian motion started from
$x_j$
where
$\boldsymbol x=(x_1,...,x_k)$
is as in the theorem statement.
Write
$\mathbf X_N = (X_N^1,...,X_N^j)$
in coordinates. By (3.13), the difference
$X_N^j(t) - N^{-1/2} M^{j,\beta _N}_{Nt}$
converges in probability to zero in the topology of
$C[0,T]$
, thus it suffices to show that if
$\mathbf U$
is any limit point of the processes
$(N^{-1/2} M^{j,\beta _N}_{Nt})_{1\le j \le k,t\ge 0}$
then
$\mathbf U$
is a martingale and each coordinate is individually a Brownian motion. The martingality of the limit point is clear, since martingality is preserved by limit points as long as one has uniform integrability as guaranteed by the
$L^q$
bounds (3.10).
Now we will show that each coordinate of the limit point
$\mathbf U$
is a standard Brownian motion. To do this, we need to study the quadratic variations of the martingales
$(N^{-1/2} M^{j,\beta _N}_{Nt})_{t\ge 0}$
for each fixed j. More precisely, note that the process
$$ \begin{align*}Y^{j,\beta}_r:= (M^{j,\beta}_r)^2 - \sum_{s=0}^{r-1} \mathbf E_{\mathbf{x}}^{(\beta,k)} [ (M^{j,\beta}_{s+1}-M^{j,\beta}_s)^2|\mathcal F_s]\end{align*} $$
is a
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
-martingale. We claim that
$(N^{-1} Y^{j,\beta _N}_{Nt})_{t\ge 0}$
is a tight family of processes in
$C[0,T]$
as
$N\to \infty $
. Indeed, one easily verifies that
$\big (N^{-1}(M^{j,\beta _N}_{Nt})^2\big )_{t\ge 0}$
is tight and satisfies the same
$L^q$
estimates as the processes
$\big (N^{-1/2}M^{j,\beta _N}_{Nt}\big )_{t\ge 0}$
on any compact time interval
$t\in [0,T]$
, simply because it is the square of a process satisfying such bounds as shown in Steps 2 and 3. On the other hand, by Minkowski’s inequality, we also have that
$$ \begin{align*}\bigg\| N^{-1} \sum_{r=Ns}^{Nt} \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} [ (M^{j,\beta_N}_{r+1}-M^{j,\beta_N}_r)^2|\mathcal F_s]\bigg\|_{L^q({\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)})}& \leq N^{-1} \sum_{r=Ns}^{Nt} \big\|M^{j,\beta_N}_{r+1}-M^{j,\beta_N}_r\big\|_{L^{2q}({\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)})}\\ &\leq \bigg(\sup_{N\ge 1} \mathcal S_{2q}(M^{j,\beta_N};{\mathbf{P}}_{{\mathbf{x}}_N}^{(\beta_N,k)}) \bigg)|t-s|,\end{align*} $$
where the supremum is finite by (3.10). Summarizing these bounds, we have that for each
$q\ge 1$
we have
$\mathbf E_{{\mathbf {x}}_N}^{(\beta _N,k)}\big [ \big (N^{-1}| Y^{j,\beta _N}_{Nt}- Y^{j,\beta _N}_{Ns}|\big )^q\big ]^{1/q} \leq C|t-s|^{1/2}$
for all
$s,t$
in a compact interval, thus completing the proof that
$(N^{-1} Y^{j,\beta _N}_{Nt})_{t\ge 0}$
is a tight family of processes in
$C[0,T]$
as
$N\to \infty $
.
We will now study the joint limit as
$N\to \infty $
under the measures
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)}$
of the pair of processes
$(N^{-1/2} M^{j,\beta _N}_{Nt}, N^{-1} Y^{j,\beta _N}_{Nt})_{t\ge 0}.$
Consider any joint limit point
$(U,V).$
Then the pair are martingales in their joint filtration, since martingality is preserved by limit points as long as one has the
$L^q$
bounds as shown above. Let us study how
$U,V$
must be related. Corollary 2.13 easily implies that

where C is independent of
$\beta \in [-\eta /(2k),\eta /(2k)].$
From this inequality and (3.9), it is immediate that
which implies from the definition of the martingales
$M^{j,\beta }$
that

see proof of Proposition 3.1 for a very similar calculation in more detail. Letting
be as in Corollary 2.13, we also have from that corollary that

where we applied Corollary 2.20 and
$|\beta _N|\to 0$
in the last line. Combining (3.14) and (3.15), and noting from Assumption 1.2 that
as
$\beta _N\to 0,$
we find that
$$ \begin{align}\lim_{N\to \infty} \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \bigg[ \bigg|t\;\;\;-\;\;\;N^{-1}\sum_{s=0}^{Nt-1} \mathbf E_{{\mathbf{x}}_N}^{(\beta_N,k)} \big[ (M_{r+1}^{j,\beta_N} -M_r^{j,\beta_N})^2\big| \mathcal F_r\big]\bigg| \bigg] =0.\end{align} $$
Using (3.16), one may immediately conclude that the joint limit point
$(U,V)$
of the pair of processes
$(N^{-1/2} M^{j,\beta _N}_{Nt}, N^{-1} Y^{j,\beta _N}_{Nt})_{t\ge 0}$
must satisfy
$V_t=U_t^2-t$
. Since both are continuous martingales as explained above, it must be true that U is a standard Brownian motion.
Step 5. In this step we will complete the proof of the theorem. The tightness of
$\mathbf X_N$
has been proved in previous steps. The tightness of the processes
$\mathscr V^{ij}_N(f;\bullet )$
is immediate from Corollary 2.20 and the fact that f is assumed to have exponential decay.
Consider any joint limit point
$(\mathbf U,(\mathcal L^{ij})_{1\le i<j\le k})$
in the space
$C([0,T],{\mathbb {R}}^k\times {\mathbb {R}}^{k(k-1)/2}).$
On one hand, the result of Step 5 shows that
$\mathbf U$
is a martingale in the joint filtration of all processes, with
$\mathbf U_0 = \boldsymbol x$
, and each coordinate has quadratic variation t, that is,
$\langle U^j\rangle _t=t$
. On the other hand, the result of Proposition 3.1 shows that
$\langle U^i-U^j\rangle _t = 2t$
whenever
$i<j$
. This forces
$\langle U^i,U^j\rangle _t = 0$
for
$i\neq j$
. We conclude by Levy’s criterion that
$\mathbf U$
is a standard k-dimensional Brownian motion. Now the result of Proposition 3.10 forces that for each
$i<j$
the marginal law of the pair
$(U^i-U^j, \mathcal L^{ij})$
must satisfy
$\mathcal L^{ij} = \gamma (f)\cdot L_0^{U^i-U^j}.$
This uniquely identifies the joint limit point
$(\mathbf U,(\mathcal L^{ij})_{1\le i<j\le k})$
in the space
$C([0,T],{\mathbb {R}}^k\times {\mathbb {R}}^{k(k-1)/2}),$
completing the proof of the theorem.
As a corollary of the above convergence theorem, we will now prove that moments of the field (1.4) converge to the moments of (1.5). While the moment convergence is not enough to prove the weak convergence of Theorem 1.5, it gives a strong indication of it, and more importantly it implies some useful estimates that will be used later.
Corollary 3.12 (Convergence of moments of the rescaled field to those of the KPZ equation)
Let
$\mathfrak H^N(t,\cdot )$
be as defined in (1.4) in the introduction. Then for each
$k\in \mathbb N, t>0,$
and
$\phi \in \mathcal S({\mathbb {R}}),$
$$ \begin{align*} \lim_{N\to\infty} {\mathbb{E}}[\mathfrak H^N(t,\phi)^k] = \mathbf E_{\mathrm{BM}^{\otimes k}} \bigg[ e^{ \gamma(\boldsymbol{\zeta}) \sum_{1\le i<j\le k} L_0^{U^i-U^j}(t)} \prod_{i=1}^k \phi(U^i_t)\bigg], \end{align*} $$
where the expectation on the right side is with respect to a standard k-dimensional Brownian motion
$(U^1,...,U^k)$
, where
$L_0^X$
denotes local time at zero of the process X, where
$\boldsymbol {\zeta }$
is the function in Definition 2.11, and
$\gamma (f)$
is the coefficient defined in Theorem 3.11.
Proof. This is immediate from Lemma 2.6, Proposition 2.12 (with all
$\beta _i=N^{-\frac 1{4p}}$
and
$m=0$
), and Theorem 3.11, noting that Corollary 2.21 guarantees uniform integrability of all relevant quantities up to exponential scales (hence convergence in law implies convergence of the respective expectations).
The following corollary of Theorem 3.11 will be useful in later sections.
Corollary 3.13. Assume the same notations and assumptions of Theorem 3.11. Suppose
$k=2m$
is even. For
$t>s>0$
define
to be the set of all ordered m-tuples of points in
$[s,t]$
. We define
$\mathcal M(\Delta _m[0,T])$
to be the space of finite and non-negative Borel measures on that simplex, equipped with the topology of weak convergence. Fix some measurable function
$f:I\to {\mathbb {R}}$
such that
$|f(x)|\leq F(|x|)$
for some decreasing
$F:[0,\infty )\to [0,\infty )$
such that
$\sum _{k=0}^\infty F(k)<\infty $
. Consider the following sequence of
$\mathcal M(\Delta _m[0,T])$
-valued random variables
$$ \begin{align} \nu_N^{f,k}:= N^{-m/2}\sum_{u_1\le ...\le u_m \in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,T]}\;\;\;\prod_{j=1}^m f\big(R^{2j-1}_{Nu_j} - R^{2j}_{Nu_j}\big)\delta_{(u_1,\ldots,u_m)} \end{align} $$
where as usual
$(R^1,...,R^{2m})$
is the canonical process on
$(I^k)^{{\mathbb {Z}}_{\ge 0}}$
. The random variables
$\{\nu _N^{f,k}\}_{N\ge 1}$
are tight in the topology of
$\mathcal M(\Delta _m[0,T])$
. Moreover, recalling the notations from Theorem 3.11, as
$N\to \infty $
any limit point under
${\mathbf {P}}_{{\mathbf {x}}_N}^{(\beta _N,k)}$
of the triple of processes
$(\mathbf X_N,\;\big (\mathscr V^{ij}_N(f,\bullet )\big )_{1\leq i<j\leq k},\; \nu _N^{f,k})$
is of the form
$$ \begin{align*} \left(\mathbf{U},\; \big(\gamma(f)\cdot L^{U^i-U^j}\big)_{1\leq i<j\leq k}, \;\gamma(f)^m\prod_{j=1}^m dL^{U^{2j-1}-U^{2j}}(u_j)\right) \end{align*} $$
where
$\mathbf U, L^{ij}$
are as in Theorem 3.11, and
$dL(t)$
denotes the Lebesgue-Stiltjes measure induced by the increasing function
$t\mapsto L(t)$
.
Proof. Note that
$$ \begin{align*} \nu_N^{f,k}(\Delta_m[0,T]) \le \prod_{j=1}^m\left[N^{-\frac12}\sum_{r=0}^{NT} f\big(R^{2j-1}_{Nu} - R^{2j}_{Nu}\big) \right]. \end{align*} $$
From the second bound in Corollary 2.20, we know for all
$q\ge 1$
that the
$q^{th}$
moments of each of the terms in the product are uniformly bounded as
$N\to \infty $
under the measures
$\mathbf E_{{\mathbf {x}}_N}^{(\beta _N,k)}$
, where
$\beta _N=N^{-\frac 1{4p}}$
. Thus for all
$q\ge 1$
, we can deduce using, for example, Hölder’s inequality that
Hence the laws of
$\{\nu _N^{f,k}\}_{N\ge 1}$
are tight as measures on
$\mathcal M(\Delta _m[0,T])$
, because the total mass of
$\nu _N^{f,k}$
is a tight family of random variables by (3.19) and because
$\Delta _m[0,T]$
as defined in (3.17) is a compact space (hence so is the unit ball of
$\mathcal M(\Delta _m[0,T])$
in the weak topology). Let
$\left (\mathbf {U}, \big (K^{ij}\big )_{1\leq i<j\leq k}, \nu \right )$
be any limit point of the sequence
$(\mathbf X_N,\;\big (\mathscr V^{ij}_N(f,\bullet )\big )_{1\leq i<j\leq k},\; \nu _N^{f,k})$
. Note that the joint cumulative distribution function of the measure
$\nu $
is necessarily given by a product of the
$K^{ij}$
, since this is true in the prelimit. Since we know that
$K^{ij}=\gamma (f) L_0^{U^{i}-U^{j}}$
by Theorem 3.11, this immediately implies that
$\nu =\gamma (f)^m \prod _{j=1}^m dL_0^{U^{2j-1}-U^{2j}}(u_j)$
.
This concludes our study of the k-point motion
$\boldsymbol p^{(k)}$
and its Girsanov tilts
$\boldsymbol q^{(k)}_\beta $
, which was initiated in Section 2. Subsequent sections will use the convergence results of this section to prove Theorem 1.5.
4 The discrete stochastic heat equation
In this section, we are going to change topic to studying the rescaled field
$\mathfrak H^N$
from (1.4). In particular, we will derive a discrete stochastic heat equation for this rescaled field
$\mathfrak H^N$
. This discrete equation will be crucial in proving Theorem 1.5, because it will eventually allow us to show that any limit point satisfies the martingale problem characterization of the equation (1.5) [Reference Konno and ShigaKS88, Reference Bertini and GiacominBG97].
First we establish some notations that will be used in this section. For a measure
$\nu $
on I we will write
$\int _I f(x) \nu ({\mathrm {d}} x+y)$
to mean the integral of f with respect to the measure
$\nu _y(A) := \nu (A+y),$
where
$A+y$
is the translate of the Borel set A by the real number y. More intuitively, one has the identity “
$\nu ({\mathrm {d}} x+y) = \nu (x+y) {\mathrm {d}} x$
.”
Let
$\vec d_N$
be as in (1.3), and define the discrete lattice
Let
$\mu $
be the measure from Assumption 1.2 Item (3), and for
$\lambda \in {\mathbb {R}}$
define the measures
$\mu ^\lambda ({\mathrm {d}} x) = e^{\lambda x - \log M(\lambda )} \mu ({\mathrm {d}} x)$
. Note that
$\mu ^\lambda $
makes sense for
$|\lambda |<\eta $
where
$\eta $
is as in Assumption 1.2 Item (3).
Definition 4.1. Let
$\Lambda _N$
be as above, and suppose that f is a signed measure on
$\Lambda _N$
. We then define the discrete heat operator
Since convolution of measures is well-defined, this expression makes sense even if
$f(t,\cdot )$
is a finite measure for all
$t\in \mathbb N$
. The resultant object
$g=\mathcal L_N f$
makes sense as a signed measure on
$\mathbb N\times I$
, such that
$g(t,\cdot )$
is a finitesigned measure for all
$t \in \mathbb N,$
which has exponential moments up to some value.
In particular if
$P^\omega (r,\cdot )$
is given by (1.2), then we can define for
$r\in {\mathbb {Z}}_{\ge 0}$
Then
$\mathcal L_N Z_N$
makes sense as a signed measure on
$\mathbb N\times I.$
Furthermore
$\mathcal L_N Z_N(Nt,\cdot )$
is a finite signed measure on I for all t, and still has exponential moments. Note that the main object of interest in Theorem 1.5 is essentially the diffusively rescaled field
$Z_N(Nt,N^{1/2}{\mathrm {d}} x),$
and roughly speaking the main goal in the remainder of the section will be to find and study a nice family of martingales for this rescaled family.
Lemma 4.2. Let
$N\in \mathbb N$
. For
$\phi \in C_c^\infty ({\mathbb {R}})$
we define
$M_t^N(\phi )$
for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
by the formula
$M_0^N(\phi )=0$
, and
Then for all
$N\in \mathbb N$
and
$\phi \in C_c^\infty ({\mathbb {R}})$
, the process
$M^N(\phi )$
is a martingale indexed by
$N^{-1}{\mathbb {Z}}_{\ge 0}$
, with respect to the filtration
$(\mathcal F^\omega _{Nt})_{t\in N^{-1}{\mathbb {Z}}_{\ge 0}}$
where
$\mathcal F_t^\omega :=\sigma (K_1,...,K_{t})$
for
$t\in \mathbb N$
(and
$\omega =(K_i)_{i=1}^\infty $
are the IID environment kernels as in Assumption 1.2).
Proof. First note by the convolution property of the kernels
$P^\omega $
of (1.2) that if
$r=Nt\in \mathbb N$
then
Now notice that
$$ \begin{align*}D_{N,N^{-1}(r+1),N^{-1/2}(x-N^{-1}d_N)} D_{N,N^{-1}r,N^{-1/2}y}^{-1} &= e^{N^{-\frac1{4p}}(x-N^{-1}d_N-y) + N^{-1-\frac1{4p}}d_N - \log M(N^{-\frac1{4p}})}\\ &= e^{N^{-\frac1{4p}}(x-y) - \log M(N^{-\frac1{4p}})}. \end{align*} $$
On the other hand we also have by definition that
$\mu ^{N^{-\frac 1{4p}}}({\mathrm {d}} x-y) = e^{N^{-\frac 1{4p}}(x-y) - \log M(N^{-\frac 1{4p}})}\mu ({\mathrm {d}} x-y).$
Therefore we can write
From the last expression the martingality is clear because
$Z_N(Nt,{\mathrm {d}} y)$
is
$\mathcal F_{Nt-1}$
-measurable, and
$K_{Nt+1}(y,{\mathrm {d}} x)$
is independent of
$\mathcal F_{Nt-1}$
with
${\mathbb {E}}[ K_{Nt+1}(y,{\mathrm {d}} x)] = \mu ({\mathrm {d}} x-y).$
Next we will calculate the predictable quadratic variations of the above martingales. If
$\phi \in C_c^\infty ({\mathbb {R}})$
define
$$ \begin{align*}J^N\phi(y_1,y_2) &:= \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \phi(N^{-1/2}x_j)\\ &\quad \times\bigg[ \boldsymbol p^{(2)} \big( (y_1,y_2),({\mathrm{d}} x_1,{\mathrm{d}} x_2)\big) - \mu({\mathrm{d}} x_1-y_1)\mu({\mathrm{d}} x_2-y_2)\bigg].\end{align*} $$
By (4.3) we find that
$$ \begin{align}\notag\langle M^N(\phi)\rangle_t &:= \sum_{s=1}^{Nt} {\mathbb{E}} \big[\big( M^N_{sN^{-1}}(\phi) - M^N_{(s-1)N^{-1}}(\phi)\big)^2 \big| \mathcal F_{s-1}^\omega \big] \\ &= \notag \sum_{s=1}^{Nt} {\mathbb{E}} \bigg[\bigg( \int_I \int_I \phi(N^{-1/2}x) e^{N^{-\frac1{4p}}(x-y) - \log M(N^{-\frac1{4p}})}\big[ K_{s+1}(y,{\mathrm{d}} x) -\mu({\mathrm{d}} x-y)\big]Z_N(s,{\mathrm{d}} y)\bigg)^2 \bigg| \mathcal F^\omega_{s-1} \bigg] \\ &= \sum_{s=1}^{Nt} \int_{I^2} (J^N\phi)(y_1,y_2) Z_N(s,{\mathrm{d}} y_1)Z_N(s,{\mathrm{d}} y_2). \end{align} $$
In the last line we again use the fact that
$Z_N(s,\bullet )$
is
$\mathcal F^\omega _{s-1}$
-measurable. Now we let
$E^\omega _{(2)}$
denote a quenched expectation of two independent particles
$(R^1_n,R^2_n)_{n\ge 0}$
sampled from a fixed realization of the environment kernels
$\{K_n\}_{n\ge 1}.$
Then the last expression may be rewritten as
$$ \begin{align*}\sum_{s=1}^{Nt} E^\omega_{(2)} \bigg[\prod_{j=1}^2 D_{N,N^{-1}s, N^{-1/2} (R^j_{Ns} - sN^{-1}d_N)} \big(J^N\phi\big) ( R^1_s,R^2_s) \bigg] .\end{align*} $$
Unfortunately
$J^N\phi $
is not yet in a form that is amenable to asymptotic analysis and tightness arguments, therefore we will need to perform Taylor expansions of some expressions therein, which will yield many “error” terms that we will need to show are inconsequential in the limit.
First let us abbreviate
Define
$$ \begin{align*}\mathcal A^1_N(\phi,y_1,y_2):= \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \cdot \big[ \prod_{j=1}^2\phi(N^{-1/2}x_j)\;\;-\;\phi(N^{-1/2}y_1)^2\big] \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big).\end{align*} $$
We can then write
$$ \begin{align} J^N\phi(y_1,y_2) = \mathcal A^1_N(\phi,y_1,y_2)+\phi(N^{-1/2}y_1)^2\cdot \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big). \end{align} $$
Now we are going to expand the exponential in powers of
$N^{-\frac 1{4p}}$
, truncating after the
$(2p)^{th}$
term. To this end, let us define
$$ \begin{align*}H_{2p}(x) := e^x - \sum_{k=0}^{2p}\frac{x^k}{k!},\end{align*} $$
so that
$|H_{2p}(x)| \leq |x|^{2p+1} e^{|x|}.$
Now write the integral appearing on the right side of (4.6) as
$$ \begin{align*} \int_{I^2}& \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big) \\&= e^{-2 \log M(N^{-\frac1{4p}})} \int_{I^2} \prod_{j=1}^2 \bigg[ H_{2p} \big(N^{-\frac1{4p}}(x_j-y_j)\big) + \sum_{k=0}^{2p} \frac{N^{-\frac{k}{4p}} (x_j-y_j)^k}{k!} \bigg] \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big). \end{align*} $$
From the deterministic moments in Assumption 1.2 Item (4), note that we necessarily have
$$ \begin{align}\int_{I^2} \prod_{j=1}^2(x_j-y_j)^{k_j}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big) = 0,\;\;\;\;\; \text{if} \;\;\;\; k_1<p \;\;\; or \;\;\; k_2<p.\end{align} $$
Therefore we may write
$$ \begin{align*} \int_{I^2} &\prod_{j=1}^2 \bigg[ H_{2p} \big(N^{-\frac1{4p}}(x_j-y_j)\big) + \sum_{k=0}^{2p} \frac{N^{-\frac{k}{4p}} (x_j-y_j)^k}{k!} \bigg] \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big) \\ &= N^{-1/2}\boldsymbol{\zeta}(y_1-y_2) + \sum_{j=2,3,4} \mathcal A^j_N(y_1-y_2), \end{align*} $$
where
$$ \begin{align*}\boldsymbol{\zeta}(y_1-y_2)&:=(p!)^{-2} \int_{I^2} \prod_{j=1}^2 (x_j-y_j)^p \rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\\ \mathcal A^2_N(y_1-y_2)&:= \int_{I^2}\prod_{j=1}^2 H_{2p} \big(N^{-\frac1{4p}}(x_j-y_j)\big)\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\\ \mathcal A^3_N(y_1-y_2)&:= 2\int_{I^2} H_{2p} \big(N^{-\frac1{4p}}(x_1-y_1)\big)\sum_{k=0}^{2p} \frac{N^{-\frac{k}{4p}} (x_2-y_2)^k}{k!}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\\ \mathcal A^4_N(y_1-y_2)&:= \int_{I^2} \sum_{\substack{p\leq k_1,k_2\leq 2p\\k_1+k_2>2p}} \frac{N^{-\frac{(k_1+k_2)}{4p}} (x_1-y_1)^{k_1}(x_2-y_2)^{k_2}}{k_1!k_2!}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big). \end{align*} $$
In the last expression we are using the fact that if
$k_1+k_2\leq 2p$
then either
$(k_1,k_2)=(p,p)$
(whose contribution is already measured by
$\boldsymbol {\zeta }$
) or at least one of
$k_1,k_2$
is less than p (hence the contribution vanishes by (4.7)). Note that the above expressions appear to be functions of
$(y_1,y_2)$
, but we have implicitly used the translation invariance condition in Assumption 1.2 Item (2) to say that they are actually only dependent on
$y_1-y_2$
.
We will ultimately show that
$\boldsymbol {\zeta }$
is important in the limit, while the terms
$\mathcal A^j_N$
have vanishing contribution. Since
$\boldsymbol {\zeta }$
has a prefactor of
$N^{-1/2}$
and since
$\mathcal A^2,\mathcal A^3,\mathcal A^4$
are of similar form but of order
$N^{-\frac 12-\frac {1}{4p}}$
and smaller, this is already somewhat intuitively clear. We therefore define
$$ \begin{align} \mathcal E^j_N(t,\phi)&:= e^{-2\log M(N^{-\frac1{4p}})}\sum_{s=0}^{Nt-1}\int_{I^2} \phi(N^{-1/2}y_1)^2\mathcal A^j_N(y_1-y_2)Z_N(s,{\mathrm{d}} y_1)Z_N(s,{\mathrm{d}} y_2),\;\;\;\;\;\;\;j=2,3,4. \end{align} $$
We also define
$$ \begin{align}\mathcal E^1_N(t,\phi) = \sum_{s=0}^{Nt-1} \int_{I^2}\mathcal A^1_N(\phi,y_1,y_2)Z_N(s,{\mathrm{d}} y_1)Z_N(s,{\mathrm{d}} y_2).\end{align} $$
We then summarize our calculations as follows.
Lemma 4.3. Let
$M^N$
be the martingale from (4.2). For all
$\phi \in C_c^\infty ({\mathbb {R}})$
,
$N\in \mathbb N$
and
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
we have that
$$ \begin{align*}\langle &M^N(\phi)\rangle_t= e^{-2\log M(N^{-\frac1{4p}})} Q^{\boldsymbol{\zeta}}_N(t,\phi^2)+\sum_{j=1}^4 \mathcal E^j_N(t,\phi)\end{align*} $$
where
$\boldsymbol {\zeta }$
is the function from Definition 2.11 and where for any bounded function
$f:I\to {\mathbb {R}}$
, the random field
$Q^f_N$
is defined as
$$ \begin{align}Q^f_N(t,\phi):=N^{-1/2}\sum_{s=1}^{Nt} E^\omega_{(2)} \bigg[\prod_{j=1}^2 D_{N,N^{-1}s, N^{-1/2} (R^j_{s} - N^{-1}d_Ns)} \cdot \phi\big(N^{-1/2} (R^1_{s} - N^{-1}d_Ns)\big)f( R^1_{s}-R^2_{s}) \bigg],\end{align} $$
and furthermore
$\mathcal E^1_N$
is defined in (4.9), where
$\mathcal E^2_N,\mathcal E^3_N,\mathcal E^4_N$
are defined in (4.8), and
$E^\omega _{(2)}$
denotes a quenched expectation of two independent motions
$(R^1_r,R^2_r)_{r\in {\mathbb {Z}}_{\ge 0}}$
in a fixed realization of the environment kernels
$\omega = (K_i)_{i=1}^\infty $
.
The above expression (4.10) for the quadratic variation will be extremely useful when identifying the limit points. In particular, it will help us to obtain the correct noise coefficient because the expression for
$\boldsymbol {\zeta }$
is precisely the expression in the numerator of this coefficient
$\gamma _{\mathrm {ext}}$
in Theorem 1.5, up to a constant.
4.1 Analysis of the quadratic variation field
This subsection will heavily focus on obtaining crucial formulas and estimates for the quadratic variation field (QVF)
$Q_N$
defined in (4.10). Later, these formulas and estimates will allow us to show tightness of (1.4) and also uniquely identify its limit points. The main idea of the proofs in this section is to use the probabilistic interpretation (4.10) of the quadratic variation field together with the strong convergence results of Section 3 to obtain very precise estimates. We first have two lemmas before stating the key estimate of this section. At this point, one should also recall the path measures
${\mathbf {P}}_{\mathrm {RW}^{(k)}}={\mathbf {P}}^{(\beta =0,k)}_{(0,...,0)}$
from Section 2, as well as their tilted generalizations from Definition 2.8.
Lemma 4.4 (Moment formulas)
Fix any bounded functions
$\psi ,\phi $
on
$\mathbb {R}$
and
$N\ge 1$
. Suppose that
$(R^1,\ldots , R^{2k})$
denotes the canonical process on
$(I^{2k})^{{\mathbb {Z}}_{\ge 0}}$
. Recall
$\Delta _k[s,t]$
from (3.17), and define
$\Delta _k^N[s,t]:= (N^{-1}{\mathbb {Z}}_{\ge 0})^k\cap \Delta _k[s,t]$
. We have the following moment formulas.
-
1. For all
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
, all
$\gamma>0$
, and all bounded
$f:I\to {\mathbb {R}}$
, we have that (4.11)where for two processes
$$ \begin{align} \begin{aligned} & \mathbb{E}\left[\bigg(Q_N^f(t,\psi)- \frac{\gamma}N\sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]} \mathfrak H^N(s,\phi)^2 \bigg)^{k}\right] \\ = k! N^{-k}\sum_{(s_1,\ldots,s_k)\in\Delta_k^N[0,t]} &\mathbf E_{\mathrm{RW}^{(2k)}}\bigg[ \prod_{i=1}^k D_{N,s_i,N^{-1/2}(R^{2i-1}_{Ns_i}- d_N s_i)}\cdot D_{N,s_i,N^{-1/2}(R^{2i}_{Ns_i}- d_N s_i)} \Upsilon_N^f({Ns_i};R^{2i-1},R^{2i})\bigg], \end{aligned} \end{align} $$
$X=(X_r)_{R\in {\mathbb {Z}}_{\ge 0}}$
and
$Y=(Y_r)_{R\in {\mathbb {Z}}_{\ge 0}}$
,
$$ \begin{align*} \Upsilon_N^f(r;X,Y) & := N^{1/2}\psi\big(N^{-1/2}(X_r-N^{-1/4}r)\big)f(X_r-Y_r)-\gamma \phi\big(N^{-1/2}(X_r-N^{-1/4}r)\big)\phi\big(N^{-1/2}(Y_r-N^{-1/4}r)\big). \end{align*} $$
-
2. For all
$0\le s<t$
we have the following moment formula for the temporal increment of the QVF (4.12)
$$ \begin{align} \begin{aligned} & \mathbb{E}\left[\bigg(Q_N^f(t,\psi)-Q_N^f(s,\psi)\bigg)^{k}\right] \\ = N^{-k/2}k! \sum_{(s_1,\ldots,s_k)\in\Delta_k^N[s,t]} & \mathbf E_{\mathrm{RW}_{\nu}^{(2k)}}\bigg[ \prod_{i=1}^kD_{N,s_i,N^{-1/2}(R^{2i-1}_{Ns_i}- d_N s_i)} D_{N,s_i,N^{-1/2}(R^{2i}_{Ns_i}- d_N s_i)} \\&\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\cdot \psi(N^{-1/2}(R^{2i}_{Ns_i}-d_N s_i))f(R^{2i-1}_{Ns_i}-R^{2i}_{Ns_i})\bigg]. \end{aligned} \end{align} $$
Proof. Note by (1.4) and (1.2) that
$$ \begin{align*}\notag \mathfrak H^N(t,\phi)^2& = \int_{I^2} D_{N,t,N^{-1/2}(x- d_N t)}D_{N,t,N^{-1/2}(y- d_N t)} \\ \notag &\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\cdot \phi\big( N^{-1/2}(x- d_N t)\big)\phi\big( N^{-1/2}(y- d_N t)\big)P^\omega(Nt,{\mathrm{d}} x)P^\omega(Nt,{\mathrm{d}} y) \\ &=E_{(2)}^\omega\notag \big[D_{N,t,N^{-1/2}(S_{Nt}- d_N t)}D_{N,t,N^{-1/2}(R_{Nt}- d_Nt)}\\ &\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\cdot\phi\big( N^{-1/2}(S_{Nt}- d_Nt)\big)\phi\big( N^{-1/2}(R_{Nt}-d_N t)\big)\big]. \nonumber \end{align*} $$
Here
$E_{(2)}^\omega $
denotes quenched expectation for two independent motions
$(R_r,S_r)_{r\ge 0}$
in the fixed realization of the IID environment kernels
$\omega = (K_i)_{i=1}^\infty .$
Note that if we apply
$\frac 1N\sum _{s\in (N^{-1}{\mathbb {Z}}_{\ge 0})\cap [0,t]}$
to the last expression, then it is of a similar form to the definition (4.10) of
$Q_N$
, hence we have that
$$ \begin{align*}Q_N^f&(t,\psi)- \frac{\gamma}N\sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]} \mathfrak H^N(s,\phi)^2 \\&= \frac1N\sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]}E_{(2)}^\omega[D_{N,s,N^{-1/2}(R_{Ns}- d_N s)}D_{N,s,N^{-1/2}(S_{Ns}- d_N s)} \Upsilon_N^f (Ns;R,S)]. \end{align*} $$
From here one expands out the
$k^{th}$
power of both sides of this equation, then uses the independence of the quenched motions, and finally applies the annealed expectation over the quenched expectation to deduce (4.11). The proof of (4.12) is similar.
Lemma 4.5. Fix any
$k\in \mathbb N$
. Let
$\{\psi _i\}_{i=1}^k$
and
$\{\phi _i\}_{i=1}^{2k}$
be bounded continuous functions on
${\mathbb {R}}$
.
Recall
$\Delta _k[s,t]$
from (3.17), and define
$\Delta _k^N[s,t]:= (N^{-1}{\mathbb {Z}}_{\ge 0})^k\cap \Delta _k[s,t].$
Let A be a subset of
$\{1,2,\ldots , k\}$
. Let
$B=\{1,2,\ldots ,k\}\cap A^c$
. For vectors
$\vec t= (t_1,...,t_k)$
define
$$ \begin{align*} & E_1(\vec{t}):= \prod_{i\in A} D_{N,t_i,N^{-1/2}(R^{2i-1}_{Nt_i}- d_N t_i)}\phi_{2i-1}\big(N^{-1/2}(R^{2i-1}_{Nt_i}- d_N t_i)\big) \\ & \hspace{3cm} \cdot \prod_{i\in A} D_{N,t_i,N^{-1/2}(R^{2i}_{Nt_i}- d_N t_i)}\phi_{2i}\big(N^{-1/2}(R^{2i}_{Nt_i}- d_N t_i)\big), \\ & E_2(\vec{t}):=\prod_{i\in B} D_{N,t_i,N^{-1/2}(R^{2i-1}_{Nt_i}-d_Nt_i)}D_{N,t_i,N^{-1/2}(R^{2i}_{Nt_i}-d_Nt_i)} \psi_i(N^{-1/2}(R_{Nt_i}^{2i} - d_Nt_i))f(R^{2i-1}_{Nt_{i}} -R^{2i}_{Nt_{i}}). \end{align*} $$
For each
$0\le s< t\le T<\infty $
we have
$$ \begin{align} \begin{aligned} & \lim_{N\to\infty} N^{-|A|-\tfrac12|B|}\cdot\mathbf E_{\mathrm{RW}^{(2k)}}\bigg[\sum_{(t_1,\ldots,t_{k})\in \Delta_{k}^N[s,t]} E_1(\vec{t})E_2(\vec{t})\bigg] \\ & = \gamma(f)^{|B|}\mathbf E_{\mathrm{BM}^{\otimes (2k)}} \bigg[\int_{\Delta_{k}[s,t]} e^{\gamma(\boldsymbol{\zeta}) \mathcal{L}_{k}(\vec{t})} \prod_{i\in A} \phi_{2i-1}(U_{t_{i}}^{2i-1}) \phi_{2i}(U_{t_{i}}^{2i})dt_i \cdot \prod_{i\in B} \psi_i(U_{t_{i}}^{2i}) dL_0^{U^{2i-1}-U^{2i}}(t_i) \bigg], \end{aligned} \end{align} $$
where the expectation on the right is with respect to a
$2k$
-dimensional standard Brownian motion
$(U^1,\ldots ,U^{2k})$
starting at the origin, and
$$ \begin{align} \mathcal{L}_k(\vec{t}):=\sum_{i=1}^k\sum_{2i-1\le p<q\le 2k} \left[L_0^{U^p-U^q}(t_{i})-L_0^{U^p-U^q}(t_{i-1})\right]. \end{align} $$
Here
$t_0=0$
, and
$\int _0^t f(s) dL_0^{U^i-U^j}(s)$
denotes the integration of the continuous function
$f:[0,t]\to {\mathbb {R}}$
against the (random) Lebesgue-Stiltjes measure
$dL_0^{U^i-U^j}$
induced from the increasing function
$t\mapsto L_0^{U^i-U^j}(t).$
We remark that A and B are allowed to be empty in the above lemma; these are important cases.
Proof. We are going to abbreviate
$X^i_N(t):=N^{-1/2}(R^i_{Nt}-d_Nt)$
, and we will also abbreviate
Let us illustrate the proof when
$k=2$
and
$A=\{1\}, B=\{2\}$
, as this will already illustrate all of the main ideas of the general argument without becoming extremely notationally cumbersome. We will denote
$(\mathcal F_r)_{r\in {\mathbb {Z}}_{\ge 0}}$
the canonical filtration on
$(I^{2k})^{{\mathbb {Z}}_{\ge 0}}$
. In this case, for each
$(t_1,t_2) \in \Delta _2[s,t]$
we can condition on the smaller time
$Nt_1$
to write
$\mathbf E_{\mathrm {RW}^{(4)}}[E_1(t_1,t_2)E_2(t_1,t_2)] $
as
$$ \begin{align} \notag &\mathbf E_{\mathrm{RW}^{(4)}}\big[ C_N\big(X_N^1(t_1),X_N^2(t_1)\big) \prod_{i=1}^2\phi_1(X_N^i(t_1)) C_N\big(X_N^3(t_2),X_N^4(t_2)\big) \psi_2(X_N^4(t_2)) f(R^3_{Nt_2}-R^4_{Nt_2})\big] \\ \notag &= \mathbf E_{\mathrm{RW}^{(4)}}\bigg[ C_N\big(X_N^1(t_1),X_N^2(t_1)\big) \prod_{i=1}^2\phi_1(X_N^i(t_1)) \\&\hspace{3 cm} \cdot \mathbf E_{\mathrm{RW}^{(4)}}[C_N\big(X_N^3(t_2),X_N^4(t_2)\big) \psi_2(X_N^4(t_2)) f(R^3_{Nt_2}-R^4_{Nt_2})\big|\mathcal F_{Nt_1}\big]\bigg]. \end{align} $$
For
${\mathbf {x}}\in I^k$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
let
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
be the “tilted” path measures from Definition 2.8, then define for
$r\in {\mathbb {Z}}_{\ge 0}$
,
$x,y\in I$
and
$\beta \in [-\eta /(2k),\eta /(2k)]$
$$ \begin{align*}&\mathscr G_N(r,x,y):= \mathbf E_{(x,y)}^{(\beta,2)}\bigg [\mathcal H^{2}\big(N^{-\frac1{4p}},(x,y),\mathbf R,r\big) \cdot \psi_2\big((N^{-1/2} (R^2_r-N^{-1}d_Nr)\big) f(R^1_r-R^2_r)\bigg], \end{align*} $$
where
$$ \begin{align*}\mathcal H^{2k}(\beta,{\mathbf{x}},\mathbf R,r):=\exp \bigg( \sum_{s=1}^r \bigg\{ \log \mathbf E_{{\mathbf{x}}}^{(0,2k)} \big[ e^{\beta\sum_{j=1}^{2k} (R^j_s -R^j_{s-1})} \big| \mathcal F_{s-1}\big] \; - 2k \log M(\beta)\bigg\}\bigg).\end{align*} $$
One sees from the Markov property of the k-point motion that the inner conditional expectation of (4.15) can be rewritten in terms of the tilted measures as
$$ \begin{align*} \mathbf E&_{RW^{(4)}}[C_N\big(X_N^3(t_2),X_N^4(t_2)\big) \psi_2(X_N^4(t_2)) f(R^3_{Nt_2}-R^4_{Nt_2})\big|\mathcal F_{Nt_1}\big] \\&= C_N\big(X_N^3(t_1),X_N^4(t_1)\big) \cdot \mathscr G_N(N(t_2-t_1), R^3_{Nt_1},R^4_{Nt_1}), \end{align*} $$
where the exponential term
$\mathcal H_N$
in the expression for
$\mathscr G_N$
appears completely analogously to Lemma 2.6. This has “reduced” all of the constants
$D_{N,t,x}$
down to time
$t_1$
, now if we apply another tilt we can get rid of those constants as well. More precisely, from the preceding observations and Definition 2.8 we have that
$$ \begin{align*} &\mathbf E_{\mathrm{RW}^{(4)}}[E_1(t_1,t_2)E_2(t_1,t_2)] \\ &= \mathbf E_{\mathrm{RW}^{(4)}}\bigg[ C_N\big(X_N^1(t_1),X_N^2(t_1)\big) C_N\big(X_N^3(t_1),X_N^4(t_1)\big) \prod_{i=1}^2\phi_1(X_N^i(t_1))\mathscr G_N( N(t_2-t_1), R^3_{Nt_1},R^4_{Nt_1})\bigg]\\&= \mathbf E_{(0,0,0,0)}^{(N^{-\frac1{4p}},4)}\bigg [\mathcal H^4(N^{-\frac1{4p}}, (0,0,0,0),\mathbf R, Nt_1)\cdot \prod_{i=1}^2\phi_1(X_N^i(t_1))\mathscr G_N( N(t_2-t_1), R^3_{Nt_1},R^4_{Nt_1})\bigg]. \end{align*} $$
Now to prove the lemma, we want to apply
$N^{-3/2} \sum _{(t_1,t_2)\in \Delta _2^N[s,t]}$
to the last expression, then take the limit and show that it approaches the correct quantity. To do this, define for
$t_1\in [s,t]\cap (N^{-1}{\mathbb {Z}}_{\ge 0})$
$$ \begin{align*}\mathscr U_N(t_1):= N^{-1/2} \sum_{t_2\in [t_1,t]\cap(N^{-1}{\mathbb{Z}}_{\ge 0})} \mathscr G_N( N(t_2-t_1), R^3_{Nt_1},R^4_{Nt_1}).\end{align*} $$
Due to the presence of the term
$f(R^1_r-R^2_r)$
in the expression for
$\mathscr G_N$
, one sees that the expression for
$\mathscr U_N$
can be interpreted as an integral with respect to the measures
$\gamma _N^f$
of Corollary 3.13. Thus using (3.19) and Corollary 2.22 the reader may convince herself that
$\mathbf E_{(0,0)}^{(N^{-\frac 1{4p}},4)}[ |\mathscr U_N(t_1)-\mathscr U_N(t_1')|^q]^{1/q} \leq C|t_1-t_1'|^{1/2}$
with
$C=C(q)$
independent of
$N,t_1,t_1'$
.
By Theorem 3.11, we know that under the measures
$\mathbf E_{(0,0,0,0)}^{(N^{-\frac 1{4p}},4)}$
, the 4-tuple
$(X_N^1,...,X_N^4)$
is converging in law to a standard 4-dimensional Brownian motion
$(W^1,W^2,W^3,W^4)$
. By applying Proposition 2.12 (with
$\beta _i=N^{-\frac 1{4p}}$
and
$m=0$
), one sees that to leading order, the expression for
$\mathcal H^2(N^{-\frac 1{4p}}, (x,y),\mathbf R,r)$
is given by the exponential of
$N^{-1/2}\boldsymbol {\zeta }(R^1_r-R^2_r)$
. Thus using Corollary 3.13 with
$k=1$
and using the fact that the map from
$\mathcal M(\Delta _k[0,T]) \times C(\Delta _k[0,T])\to {\mathbb {R}}$
given by
$(f,\mu )\mapsto \int _{\Delta _k[0,T]} f\;d\mu $
is continuous, one may show that under the measures
$\mathbf E_{(0,0,0,0)}^{(N^{-\frac 1{4p}},4)}$
, the process
$\mathscr U_N$
is converging in law jointly with the
$X_N^i$
(with respect to the topology of
$C[0,T]$
) to the process
$$ \begin{align*}V(t_1):= \gamma(f) \cdot \mathbf E_{\mathrm{BM}^{\otimes 2}}^{(W^3_{t_1},W^4_{t_1})} \bigg[\int_{t_1}^te^{\gamma(\boldsymbol{\zeta}) L_0^B(t_2-t_1)} \psi_2(B_{t_2}) dL_0^B(t_2)\bigg].\end{align*} $$
Here the expectation is with respect to a Brownian motion B of rate 2 that is independent of
$(W^1,...,W^4)$
. Uniform integrability of all quantities involved (hence convergence of the respective expectations) is clear from the moment bounds in Corollary 2.20, Corollary 2.21, and Equation (3.19).
By again applying Proposition 2.12 (with
$\beta _i=N^{-\frac 1{4p}}$
and
$m=0$
), one sees that to leading order the expression for
$\mathcal H^4(N^{-\frac 1{4p}}, (0,0,0,0),\mathbf R,r)$
is given by the exponential of
$N^{-1/2}\sum _{1\le i<j\le 4}\boldsymbol {\zeta }(R^i_r-R^j _r)$
. Summarizing these facts and applying the convergence in law of the 5-tuple
$(X_N^1,...,X_N^4,\mathscr U_N)$
to
$(W^1,...,W^4,V)$
, we see that
$$ \begin{align*} N&^{-3/2} \mathbf E_{\mathrm{RW}^{(4)}} \bigg[\sum_{(t_1,t_2)\in \Delta_2^N[s,t]}E_1(t_1,t_2)E_2(t_1,t_2)\bigg] \\ &= \mathbf E_{(0,0,0,0)}^{(N^{-\frac1{4p}},4)}\bigg [N^{-1}\sum_{t_1\in [s,t]\cap (N^{-1}{\mathbb{Z}}_{\ge 0})}\mathcal H^4(N^{-\frac1{4p}}, (0,0,0,0),\mathbf R, Nt_1)\cdot \prod_{i=1}^2\phi_1(X_N^i(t_1))\cdot \mathscr U_N(t_1)\bigg]\\ &\stackrel{N\to\infty}{\longrightarrow} \mathbf E_{\mathrm{BM}^{\otimes 4}} \bigg[ \int_s^t e^{\gamma(\boldsymbol{\zeta}) \sum_{1\le i<j\le 4} L_0^{W^i-W^j}(t_1) } \prod_{i=1}^2 \phi_1( W^i(t_1)) \cdot V(t_1)dt_1 \bigg] \end{align*} $$
where we used the fact that convergence in the topology of
$C[0,T]$
is strong enough to imply convergence in law of the Riemann sums of the prelimiting process to the integral of the limiting process (and the uniform integrability guaranteed by Corollaries 2.20 and 2.22 guarantees convergence of the respective expectations). It is now an exercise with the definition of the process V and the Markov property of Brownian motion to simplify the last expression to the desired form as in the lemma statement.
While we have proved the claim in a special case when
$k=2$
, the proof of the general case is similar: one applies the Markov property repeatedly and proceeds inductively in the variable k, using the convergence results of Sections 2 and 3 at each step (specifically Corollaries 2.20, 2.21, 2.22, 3.13, and Theorem 3.11 exactly as we have done above).
Proposition 4.6 (Key estimate for the QVF)
Let
$a \in {\mathbb {R}}$
and let
$\xi (x):= \frac 1{\sqrt {\pi }}e^{-x^2}$
, and let
$\xi _\varepsilon ^a (x):= \varepsilon ^{-1}\xi (\varepsilon ^{-1} (x-a))$
. Let
$\boldsymbol {\zeta }:I\to {\mathbb {R}}$
be the specific function from Definition 2.11. Then for all
$t>0$
and
$a\in {\mathbb {R}}\setminus \{0\}$
,
$$ \begin{align} \lim_{\varepsilon \to 0} \;\limsup_{N \to \infty}\; \mathbb{E}\bigg[ \bigg( Q_N^{\boldsymbol{\zeta}}(t,\xi_\varepsilon^a)- \frac{\gamma(\boldsymbol{\zeta})}N\sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]}\mathfrak H^N(s,\xi_{\varepsilon\sqrt{2}}^a)^2\bigg)^2 \bigg] = 0. \end{align} $$
Furthermore, for all
$t>0$
we have the bound
$$ \begin{align}\sup_{\substack{\varepsilon>0\\a\in{\mathbb{R}}\setminus\{0\}}} \limsup_{N \to \infty} \left[1\wedge\frac1{|\log a|^{2}}\right]\!\cdot\!\mathbb{E}\bigg[ \bigg( Q_N^{\boldsymbol{\zeta}}(t,\xi_\varepsilon^a)-\frac{\gamma(\boldsymbol{\zeta})}N\sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]}\mathfrak H^N(s,\xi_{\varepsilon\sqrt{2}}^a)^2\bigg)^2 \bigg] <\infty. \end{align} $$
The above proposition is the key estimate that will allow us to identify limit points. We remark that in the proposition statement
$\boldsymbol {\zeta }$
and
$\gamma (\boldsymbol {\zeta })$
can respectively be replaced by any f and
$\gamma (f)$
for any function
$f:I\to {\mathbb {R}}$
of exponential decay, though we will not need the more general version.
Proof. Applying Lemma 4.4 1 with
$\gamma = \gamma (\boldsymbol {\zeta })$
, and Proposition 4.5 with
$k=2$
, then we obtain
$$ \begin{align} \nonumber & \lim_{N\to \infty} \mathbb{E}\left[\bigg(Q_N(t,\psi)-\frac{\gamma(\boldsymbol{\zeta})}{N} \sum_{s\in (N^{-1}{\mathbb{Z}}_{\ge 0})\cap [0,t]} \mathfrak H^N(s,\phi)^2\bigg)^{2}\right] \\ \nonumber & =2\gamma(\boldsymbol{\zeta})^2 \cdot \mathbf{E}_{\mathrm{BM}^{\otimes 4}}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta}) \mathcal{L}_2(s_1,s_2)}\prod_{i=1}^2 \left(\psi(X_{s_i}^i) \tfrac12 dL_0^{X^i-Y^i}(s_i)-\phi(X^i)\phi(Y^i)ds_i\right)\right] \\ & = 2\gamma(\boldsymbol{\zeta})^2 \hspace{-0.1cm}\!\cdot\! \mathbf{E}_{\mathrm{BM}^{\otimes 4}}\left[\ \int\limits_{\Delta_2(0,t)} \hspace{-0.3cm}e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)}\prod_{i=1}^2 \left(\psi\big(\tfrac12(X_{s_i}^i+Y_{s_i}^i)\big) \tfrac12 dL_0^{X^i-Y^i}(s_i)-\phi(X^i)\phi(Y^i)ds_i\right)\right] \end{align} $$
for all
$\psi ,\phi \in \mathcal {S}(\mathbb {R})$
, where
$\mathcal {L}_2$
is defined in (4.14), and
$(X^1,X^2,Y^1,Y^2)$
is a standard Brownian motion in
$\Bbb R^4$
under
${\mathbf {P}}_{B^{\otimes 4}}$
. The second equality in the above equation follows by observing that
$X_u^i=Y_u^i$
for u in the support of
$L_0^{X^i-Y^i}(du)$
.
We shall now write
$\mathbf {E}$
instead of
$ \mathbf {E}_{B^{\otimes 4}}$
for convenience. Let us now take
in (4.18). Using the identity
$\xi ^a_{\varepsilon \sqrt {2}}(x)\xi ^a_{\varepsilon \sqrt {2}}(y)= \xi ^a_\varepsilon ((x+y)/2)\xi ^0_{2\varepsilon }(x-y)$
, we may now write (4.18) as
$$ \begin{align} 2\gamma(\boldsymbol{\zeta})^2\hspace{-0.1cm} \!\cdot\! \mathbf{E}\left[\ \int\limits_{\Delta_2(0,t)} \hspace{-0.2cm}e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)}\prod_{i=1}^2 \left(\xi_\varepsilon^a\big(\tfrac12(X_{s_i}^i+Y_{s_i}^i)\big)\!\cdot\!(\tfrac12 dL_0^{X^i-Y^i}(s_i)-\xi_{2\varepsilon}^{0}(X_{s_i}^i-Y_{s_i}^i)ds_i)\right)\right]. \end{align} $$
Let us write
$U^{i,-}:=X^i-Y^i$
and
$U^{i,+}:=X^i+Y^i$
. Note that under
${\mathbf {P}}_{B^{\otimes 4}}$
the four processes
$U^{1,-},U^{1,+},U^{2,-},U^{2,+}$
are independent Brownian motions with diffusion coefficient
$2$
. This enables us to view (4.19) as
$$ \begin{align} \nonumber & 2\gamma(\boldsymbol{\zeta})^2 \cdot \mathbf{E}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)}\prod_{i=1}^2 \left(\xi_\varepsilon^a\big(\tfrac12U_{s_i}^{i,+})\big)\cdot(\tfrac12 dL_0^{U^{i,-}}(s_i)-\xi_{2\varepsilon}^{0}(U_{s_i}^{i,-})ds_i)\right)\right] \\ & =: 2\gamma(\boldsymbol{\zeta})^2 [A_1(\varepsilon)-A_2(\varepsilon)-A_3(\varepsilon)+A_4(\varepsilon)], \end{align} $$
where
$$ \begin{align*} A_1(\varepsilon) & := \mathbf{E}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)}\xi_\varepsilon^a\big(\tfrac12U_{s_1}^{1,+}\big)\xi_\varepsilon^a\big(\tfrac12U_{s_2}^{2,+}\big)\,\tfrac12 dL_0^{U^{1,-}}(s_1)\,\tfrac12dL_0^{U^{2,-}}(s_2)\right] \end{align*} $$
$$ \begin{align*} A_2(\varepsilon) & := \mathbf{E}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)} \xi_\varepsilon^a\big(\tfrac12U_{s_1}^{1,+}\big)\xi_\varepsilon^a\big(\tfrac12U_{s_2}^{2,+}\big)\xi_{2\varepsilon}^{0}(U_{s_2}^{2,-})\, \tfrac12 dL_0^{U^{1,-}}(s_1)\, ds_2\right] \end{align*} $$
$$ \begin{align*} A_3(\varepsilon) & := \mathbf{E}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)} \xi_\varepsilon^a\big(\tfrac12U_{s_1}^{1,+}\big)\xi_\varepsilon^a\big(\tfrac12U_{s_2}^{2,+}\big)\xi_{2\varepsilon}^{0}(U_{s_1}^{1,-})\, \tfrac12 dL_0^{U^{2,-}}(s_2)\, ds_1\right] \end{align*} $$
$$ \begin{align*} A_4(\varepsilon) & := \mathbf{E}\left[\int_{\Delta_2(0,t)} e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_2(s_1,s_2)}\xi_\varepsilon^a\big(\tfrac12U_{s_1}^{1,+}\big)\xi_\varepsilon^a\big(\tfrac12U_{s_2}^{2,+}\big)\xi_{2\varepsilon}^{0}(U_{s_1}^{1,-})\xi_{2\varepsilon}^{0}(U_{s_2}^{2,-})\,ds_1\,ds_2\right]. \end{align*} $$
From here, the goal is to show that (4.20) vanishes as
$\varepsilon \to 0$
as long as
$a\neq 0$
, as well as establish a bound given by the right side of (4.17) for each of the four terms
$A_i(\varepsilon )$
.
Formally,
$\tfrac 12 dL_0^{U^{i,\pm }}(s_i)$
is the same as
$\delta _0(U^{i,\pm }) ds_i$
, which suggests that each of the
$A_i$
may be written in terms of Brownian bridge expectations. This is indeed the case, and consequently the proofs of the desired convergence statements and bounds for the terms
$A_i(\varepsilon )$
rely purely on elementary (albeit lengthy) disintegration formulas for Brownian motion at its endpoint, and these proofs can be copied verbatim from [Reference Das, Drillick and ParekhDDP24a, Proof of Proposition 5.3: Steps 2 and 3], replacing the coefficient
$\sigma $
appearing there with our coefficient
$\gamma (\boldsymbol {\zeta })$
throughout the proof. For brevity, we will not reproduce the details here.
With the “key estimate” proved, next we focus on obtaining bounds that will be useful for proving tightness of the rescaled field (1.4).
Proposition 4.7 (Estimates for moments of the increments of QVF)
Fix
$k\in \mathbb N$
,
$T>0$
, and
$F:[0,\infty )\to [0,\infty )$
decreasing such that
$\sum _{k=0}^\infty F(k)<\infty $
. Let
$f(x):=F(|x|)$
. Then there exists a constant
$C=C(k,T,F)>0$
such that for all bounded measurable functions
$\phi $
on
${\mathbb {R}}$
and all
$0\leq s<t\leq T$
with
$s,t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
one has that
Furthermore fix
$q>1$
and
$\varepsilon>0$
. Then there exists
$C=C(q,\varepsilon ,k,T,F)>0$
such that for all functions
$\phi \in L^q(\mathbb {R})$
and all
$\varepsilon \leq s<t\leq T$
one has
Proof. We are going to use the same notation and the same family of martingales from the proof of Lemma 4.5. Using (4.12) together with the trivial bound
$|\phi (N^{-1/2}(R^j(Nt_j)-N^{3/4}t_j))|\leq \|\phi \|_{L^\infty },$
we obtain that
$$ \begin{align*} \notag &\mathbb{E}[\big(Q_N^f(t,\phi)-Q_N^f(s,\phi)\big)^k] \\ & \le N^{-k/2} k!\|\phi\|_{L^\infty}^k\sum_{(t_1,\ldots,t_k)\in \Delta_k^N[s,t]} \mathbf E_{\mathrm{RW}^{(2k)}}\bigg[\prod_{j=1}^k\bigg\{ f(R^{2j-1}_{Nt_j} -R^{2j}_{Nt_j}) \prod_{i=0,1} D_{N,t_j,N^{-1/2}(R^{2j-i}_{{Nt_j}}- d_N t_j)}\bigg\} \bigg]. \end{align*} $$
Now we actually claim, more generally than (4.21), that one has
$$ \begin{align*}\sup_{\substack{{\mathbf{x}}\in I^k\\N\ge 1}} N^{-k/2} \sum_{(t_1,\ldots,t_k)\in \Delta_k^N[s,t]} \mathbf E_{\mathbf{x}}^{(0,2k)}\bigg[\prod_{j=1}^k\bigg\{ f(R^{2j-1}_{Nt_j} -R^{2j}_{Nt_j}) \prod_{i=0,1} D_{N,t_j,N^{-1/2}(R^{2j-i}_{{Nt_j}}- d_N t_j)}\bigg\} \bigg]\leq C(t-s)^{k/2}.\end{align*} $$
The main idea in proving this is to proceed inductively in the variable k, applying similar arguments as in the proof of Lemma 4.5 to get rid of the constants by repeatedly applying the Markov property and writing these expressions in terms of the tilted measures
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
of Definition 2.8.
For clarity, write some of the details of this inductive argument. The claim is vacuously true when
$k=0$
, let us assume (4.21) has been proved up to
$k-1$
where
$k\in \mathbb N.$
By conditioning on time
$t_1$
, using the Markov property, and applying the inductive hypothesis for the conditioned expression on the time interval
$[Nt_1,Nt]$
one will obtain uniformly over
${\mathbf {x}}\in I^k$
and
$N\ge 1$
that
$$ \begin{align*} N&^{-k/2} \sum_{(t_1,\ldots,t_k)\in \Delta_k^N[s,t]} \mathbf E_{\mathbf{x}}^{(0,2k)}\bigg[\prod_{j=1}^k\bigg\{ f(R^{2j-1}_{Nt_j} -R^{2j}_{Nt_j}) \prod_{i=0,1} D_{N,t_j,N^{-1/2}(R^{2j-i}_{{Nt_j}}- d_N t_j)}\bigg\} \bigg] \\ &\leq N^{-1/2} \sum_{t_1\in [s,t]\cap (N^{-1}{\mathbb{Z}}_{\ge 0})} (t-t_1)^{(k-1)/2}\mathbf E_{\mathbf{x}}^{(0,2k)} \bigg[ f(R^1_{Nt_1}-R^2_{Nt_1}) \prod_{j=1}^k \bigg\{ \prod_{i=0,1} D_{N,t_1,N^{-1/2}(R^{2j-i}_{Nt_1}-d_Nt_1)}\bigg\}\bigg]. \end{align*} $$
In other words all of the renormalizing constants
$D_{N,t,x}$
have been “reduced” to time
$t_1$
and the inductive hypothesis has yielded the factor
$(t-t_1)^{(k-1)/2}$
after summing out all other coordinates. Let us define the adapted process
$$ \begin{align*}\mathcal H_N(\mathbf R, r):= \exp \bigg( \sum_{s=1}^r \bigg\{ \log \mathbf E_{\mathrm{RW}^{(2k)}} \big[ e^{N^{\frac1{4p}}\sum_{j=1}^{2k} (R^j_s -R^j_{s-1})} \big| \mathcal F_{s-1}\big] \; - 2k \log M(N^{-\frac1{4p}})\bigg\}\bigg).\end{align*} $$
Then the previous expression can be rewritten in terms of the tilted measures
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
(see, e.g., Lemma 2.6) as
$$ \begin{align*}\mathbf E_{\mathbf{x}}^{(N^{-\frac1{4p}},2k)} \bigg[ N^{-1/2} \sum_{t_1\in [s,t]\cap (N^{-1}{\mathbb{Z}}_{\ge 0})} (t-t_1)^{(k-1)/2}\mathcal H_N(\mathbf R,Nt_1) f(R^1_{Nt_1}-R^2_{Nt_1}) \bigg].\end{align*} $$
If we brutally bound
$\mathcal H_N(\mathbf R,Nt_1) \leq \sup _{0\leq r\leq Nt}| \mathcal H_N(\mathbf R,r)|:= \mathcal H_N^{\mathrm {sup}}(\mathbf R,r) $
, then an application of Corollary 2.22 will show that
$\Gamma (q):=\sup _{N\ge 1} \sup _{{\mathbf {x}}\in I^k} \mathbf E_{\mathbf {x}}^{(N^{-\frac 1{4p}},2k)} [\mathcal H_N^{\mathrm {sup}}(\mathbf R,r)^q]< \infty $
where
$q\ge 1$
is arbitrary. We can also use the bound
$(t-t_1)^{(k-1)/2}$
by
$(t-s)^{(k-1)/2},$
thus by Cauchy-Schwarz the last expression can be bounded above
$$ \begin{align*}(t-s)^{(k-1)/2} \mathbf E_{\mathbf{x}}^{(N^{-\frac1{4p}},2k)} \bigg[ \mathcal H_N^{\mathrm{sup}}(\mathbf R,Nt) \cdot \nu^{f,1}_N([s,t])\bigg] \leq (t-s)^{(k-1)/2} \Gamma(2)^{1/2} \cdot \mathbf E_{\mathbf{x}}^{(N^{-\frac1{4p}},2k)} \big[ \nu^{f,1}_N([s,t])^2\big]^{1/2}\end{align*} $$
where we recall that the measure
$\nu _N^{f,k}$
was introduced in (3.18). By Corollary 2.20 it is immediate that
$\sup _{N\ge 1} \sup _{{\mathbf {x}}\in I^k}\mathbf E_{\mathbf {x}}^{(N^{-\frac 1{4p}},2k)} \big [ \nu ^{f,1}_N([s,t])^2\big ]^{1/2} \leq C\sqrt {t-s},$
which completes the inductive step and thus finishes the proof of (4.21).
For (4.22), the proof is completely analogous to a similar proof given in [Reference Das, Drillick and ParekhDDP24a, Eq. (5.43)] but we rewrite some of the details here for completeness. Appealing to the moment formula for the increment of QVF from (4.12) and the convergence from (4.13) we have
$$ \begin{align} \lim_{N\to \infty}\mathbb{E}\bigg[\big(Q_N^f(t,\phi)-Q_N^f(s,\phi)\big)^k\bigg] = k!\gamma(\boldsymbol{\zeta})^k \cdot \mathbf E_{B^{\otimes 2k}} \bigg[\int_{\Delta_k[s,t]}e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_k(\vec{t})}\prod_{j=1}^k \phi(U_{t_j}^{2j})dL_0^{U^{2j-1}-U^{2j}}(t_j) \bigg], \end{align} $$
where
$\mathcal {L}_k(\vec {t})$
is defined in (4.14). Now [Reference Das, Drillick and ParekhDDP24a, Lemma A.2] formalizes the intuition that
$dL_0^{U^{2j-1}-U^{2j}}(t_j) = \delta _0(U_{t_j}^{2j-1}-U^{2j}_{t_j})$
. Thus by appealing to that lemma, we can write the above expectation in terms of a certain family of concatenated bridge processes whose law we will write as
${\mathbf {P}}_c$
. Specifically we have that
$$ \begin{align} \text{right hand side~of (4.23)} = \int_{\Delta_k[s,t]} \mathbf E_{c} \bigg[e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_k(\vec{t})}\prod_{j=1}^k \phi(U^{2j}_{t_j})\bigg]\prod_{j=1}^k p_{2t_j}(0)dt_1\cdots dt_k, \end{align} $$
where the expectation is taken over a collection of paths
$U^j$
such that
-
•
$(U^{2i-1}-U^{2i}, U^{2i-1}+U^{2i})_{i=1}^k$
are
$2k$
many independent processes. -
•
$U^{2i-1}+U^{2i}$
is a Brownian motion of diffusion rate
$2$
for
$i=1,2,\ldots ,k$
. -
•
$U^{2i-1}-U^{2i}$
is a Brownian bridge (from
$0$
to
$0$
) of diffusion rate
$2$
from
$[0,t_i]$
and an independent Brownian motion of diffusion rate
$2$
from
$[t_i,\infty )$
for
$i=1,2,\ldots ,k$
.
We have used a different notation for the expectation operator
$\mathbf {E}_c$
in (4.24) (c for concatenation) just to stress that the law is different from the standard Brownian motion. We claim that for all
$q>1$
we have
$$ \begin{align} \sup_{\vec{t}\in [0,T]} \mathbf E_{c} \bigg[e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_k(\vec{t})}\prod_{j=1}^k |\phi(U^{2j}_{t_j})|\bigg] \le C\cdot \|\phi\|_{L^q}^k. \end{align} $$
where the
$C>0$
depends on
$q,k,\varepsilon ,T$
. Let us assume (4.25) for the moment. By hypothesis,
$t_j\ge \varepsilon>0$
. Thus,
$p_{2t_j}(0) \le C$
for some constant
$C>0$
depending on
$\varepsilon $
. Thus, in view of (4.25), to get an upper bound for the right-hand side of (4.24), we may take the supremum of the integrand in the right-hand side of (4.24) and pull it outside of the integration. As the Lebesgue measure of
$\Delta _k[s,t]$
is
$\frac {(t-s)^k}{k!}$
, we thus have the desired estimate in (4.22).
Let us now establish (4.25). Fix
$q>1$
and take
$v>1$
so that
$(q')^{-1}+q^{-1}=1$
. Use Hölder’s inequality to write
$$ \begin{align*} \mathbf E_{c} \bigg[e^{\gamma(\boldsymbol{\zeta})\mathcal{L}_k(\vec{t})}\prod_{j=1}^k |\phi(U^{2j}_{t_j})|\bigg] &\leq \mathbf E_{c} \bigg[e^{q'\gamma(\boldsymbol{\zeta})\mathcal{L}_k(\vec{t})}\bigg]^{1/q'}\mathbf E_c \bigg[\prod_{j=1}^k |\phi(U^{2j}_{t_j})|^q\bigg]^{1/q}. \end{align*} $$
For the first expectation above, observe that by [Reference Das, Drillick and ParekhDDP24a, Lemma A.6],
$\mathbf E_{c} \bigg [e^{q'\gamma (\boldsymbol {\zeta })\mathcal {L}_k(\vec {t})}\bigg ]$
is uniformly bounded over
$\vec {t}\in \Delta _k(0,T)$
. For the second expectation above, note that under
$\mathbf {E}_c$
, we have
Thus under
$\mathbf {E}_c$
,
$U_{t_j}^{2j}$
are independent Gaussian random variables with variance
$t_j/2$
. Hence,
$$ \begin{align*} \mathbf E_c \bigg[\prod_{j=1}^k |\phi(U^{2j}_{t_j})|^q\bigg]^{1/q} = \prod_{j=1}^k\mathbf E_c \bigg[ |\phi(U^{2j}_{t_j})|^q\bigg]^{1/q} = \prod_{j=1}^k \bigg(\int_{\mathbb{R}} p_{t_j/2}(y)|\phi(y)|^q\,dy\bigg)^{1/q} \le C\cdot \|\phi\|_{L^q}^k. \end{align*} $$
where the last inequality follows by again using the fact that
$t_j\ge \varepsilon>0$
, together with the uniform bound
$\sup _y p_t(y)\leq (2\pi t)^{-1/2}$
, and noting that the constant C is allowed to depend on
$\varepsilon $
. This establishes (4.25) completing the proof of (4.22).
Lemma 4.8. Fix
$k\in \mathbb N$
. For all
$T>0$
and
$q\ge 0$
we have that
$$ \begin{align*}\sup_{0\leq t_1\leq ... \leq t_k\leq T} \sup_{N\in \mathbb N}\sup_{r\le NT} \mathbf E_{\mathrm{RW}^{(2k)}} \bigg[ \bigg| N^{-1/2}\sum_{1\le i<j\le k} |R^i_{r} -R^j_r|\bigg|^q\cdot \prod_{j=1}^{k} D_{N,t_j,N^{-1/2}(R^j_{Nt_j}-d_Nt_j)}\bigg] <\infty.\end{align*} $$
Proof. The proof is very similar to the proofs of Lemma 4.5 and Proposition 4.7, and we do not give the full argument. The sketch is as follows. One inducts on k to show more generally that
$$ \begin{align*}\sup_{0\leq t_1\leq ... \leq t_k\leq T} \sup_{N\in \mathbb N}\sup_{{\mathbf{x}}\in I^k} \sup_{r\le NT}\mathbf E_{{\mathbf{x}}}^{(0,k)} \bigg[ \bigg| N^{-1/2}\sum_{1\le i<j\le k} |R^i_{r} -R^j_r-(x_i-x_j)|\bigg|^q\prod_{j=1}^{k} D_{N,t_j,N^{-1/2}(R^j_{Nt_j}-x_j-d_Nt_j)}\bigg] <\infty.\end{align*} $$
To perform the induction, one applies the Markov property and the inductive hypothesis to bound the expression on the time interval
$[Nt_1,NT]$
, then one rewrites the remaining expectation in terms of the tilted measures
$ {\mathbf {P}}_{\mathbf {x}}^{(\beta ,k)}$
of Definition 2.8. On the remaining time interval
$[0,Nt_1]$
this will yield an exponential term which we called
$\mathcal H_N(\mathbf R,r)$
in the proof of Proposition 4.7. This exponential term has uniform moment bounds as explained in that proof. The other term
$N^{-1/2}\sum _{1\le i<j\le k} |R^i_{r} -R^j_r|$
also has uniform moment bounds using, for example, Corollaries 2.20 and 2.21. Using Hölder’s inequality will then yield finiteness of the supremum to complete the inductive step.
Recall the martingale field
$M^N$
from (4.2). The next few estimates will obtain a
$L^p$
bound on its predictable and optional quadratic variation, which will be the main tool in obtaining tightness for the field (1.4). First we will need to control the error terms appearing in Lemma 4.3. For
$\phi :{\mathbb {R}}\to {\mathbb {R}}$
denote by
$\|\phi \|_{C^k}=\|\phi \|_{C^k({\mathbb {R}})}:=\sum _{j=0}^k \|\phi ^{(j)}\|_{L^\infty ({\mathbb {R}})},$
where
$\phi ^{(j)}(x) = \frac {d^j}{dx^j}\phi (x).$
Proposition 4.9 (Controlling the error terms)
Let
$\mathcal E^j_N (1\le j \le 4)$
be as in Lemma 4.3. Let
$q\ge 1$
be an even integer, and let
$T>0$
. Then there exists
$C>0$
and a function
$f:I\to {\mathbb {R}}_+$
such that
$|f(x)| \leq F(|x|)$
where
$F:[0,\infty )\to [0,\infty )$
is decreasing and
$\sum _{k=0}^\infty F(k)<\infty $
, such that uniformly over all
$N\ge 1$
and
$s,t\in [0,T]\cap (N^{-1}{\mathbb {Z}})$
we have
$$ \begin{align*}\sum_{j=2,3,4}{\mathbb{E}}[ |\mathcal E^j_N(t,\phi)-\mathcal E^j_N(s,\phi)|^q]^{1/q} \leq C N^{-\frac1{4p}}{\mathbb{E}}[ |Q^f_N(t,\phi^2)-Q^f_N(s,\phi^2)|^q]^{1/q}+N^{-\frac1{4p}}\|\phi\|_{L^\infty}^2|t-s|.\end{align*} $$
Furthermore, if
$\phi \in C_c^\infty ({\mathbb {R}})$
, say
$\phi $
is supported on
$[-J,J]$
then
Proof. As introduced earlier in this section, we let
$H_k(x)= e^x - \sum _{n=0}^k \frac {x^k}{k!}$
so that
$|H_k(x)| \leq |x|^{k+1}e^{|x|}$
for all
$x\in {\mathbb {R}}$
.
First let us deal with
$\mathcal E_N^2$
. In the expression for
$\mathcal A^2_N$
preceding (4.8), use the fact that
$|H_{2p}(N^{-\frac 1{4p}} (x_j-y_j))| \leq CN^{-\frac {(2p+1)}{4p}} |x_j-y_j|^{2p+1} e^{N^{-\frac 1{4p}}|x_j-y_j|} $
and we obtain uniformly over
$N\ge 1$
and
$y_1,y_2\in I$
the brutal bound that
$$ \begin{align*}|\mathcal A_N^2(y_1-y_2)| \leq N^{-\frac{(4p+2)}{4p}} \int_{I^2} \prod_{j=1}^2 |x_j-y_j| e^{N^{-\frac1{4p}}|x_j-y_j|}\big| \rho\big( (y_1,y_2),({\mathrm{d}} x_1,{\mathrm{d}} x_2)\big)\big| \leq CN^{-\frac{(4p+2)}{4p}}\end{align*} $$
where the integral can be bounded independently of
$N,y_1,y_2$
using the definition (4.5) of
$\rho $
and, for example, Lemma 2.9. From the above bound on
$\mathcal A^2_N$
it is immediate from (4.8) that
$$ \begin{align*}|\mathcal E^2_N(t,\phi)-\mathcal E_N^2(s,\phi)| \leq CN^{-\frac{(4p+2)}{4p}}e^{-2\log M(N^{-\frac1{4p}})} \|\phi\|_{L^\infty}^2 \sum_{r=Ns}^{Nt} \int_{I^2} Z_N(s,{\mathrm{d}} y_1)Z_N(s,{\mathrm{d}} y_2).\end{align*} $$
Taking the
$L^{2k}$
-norm and applying Minkowski’s inequality to the sum over r, we obtain that
$$ \begin{align*}{\mathbb{E}}\big[&\big(\mathcal E^2_N(t,\phi)-\mathcal E_N^2(s,\phi)\big)^{2k}\big]^{1/(2k)} \\ &\leq CN^{-\frac{(4p+2)}{4p}}e^{-2\log M(N^{-\frac1{4p}})} \|\phi\|_{L^\infty}^2 \sum_{r=Ns}^{Nt} \mathbf E_{\mathrm{RW}^{(2k)}} \bigg[ \prod_{j=1}^{2k} D_{N,N^{-1}r,N^{-1/2}(R^{j}_{r}-N^{-1}d_Nr)}\bigg]^{1/(2k)}\end{align*} $$
where we applied the definition of
$Z_N$
as defined in (4.1). By Lemma 4.8 (with
$q=0$
) the summands are bounded by some absolute constant independently of
$r,N\in \mathbb N$
, consequently since the number of summands is
$N(t-s)$
the entire expression can be bounded above by
$CN^{-\frac 1{2p}}\|\phi \|_{L^\infty }^2|t-s|.$
This is a bound of the desired form for
$\mathcal E^2_N.$
Let us now deal with
$ \mathcal E^4_N$
. In light of (4.8), it suffices to show that there exists a decreasing function
$F:[0,\infty )\to [0,\infty ) $
with
$\sum _{k=0}^\infty F(k)<\infty $
such that
$\sup _{N\ge 1} N^{\frac 1{4p}} |\mathcal A^4_N(y)| \leq F(|y|)$
. Indeed, this would then yield a pathwise bound of the form
where
$f(x)=F(|x|)$
, which certainly implies the claim. But the fact that
$\sup _{N\ge 1} N^{\frac 1{4p}} |\mathcal A^4_N(y)| \leq \mathrm {F_{decay}}(|y|)$
is immediate from Assumption 1.2 Item (5) since the negative power of N in each of the summands is
$(k_1+k_2)/(4p)>1/2$
(and
$1/2$
is the power in the definition of
$Q_N^f).$
This already gives the desired bound for
$\mathcal E^4_N.$
Next let us deal with
$\mathcal E^3_N.$
By writing
$H_{2p}=(H_{2p}-H_{4p})+H_{4p},$
we may rewrite
$$ \begin{align*} \mathcal A^3_N(y_1-y_2) &= 2 \int_{I^2} H_{4p} \big(N^{-\frac1{4p}}(x_1-y_1)\big)\sum_{k=0}^{2p} \frac{N^{-\frac{k}{4p}} (x_2-y_2)^k}{k!}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big) \\&\;\;+ 2\int_{I^2} \bigg(\sum_{k=2p+1}^{4p} \frac{N^{-\frac{k}{4p}} (x_1-y_1)^k}{k!}\bigg)\bigg(\sum_{k=0}^{2p} \frac{N^{-\frac{k}{4p}} (x_2-y_2)^k}{k!}\bigg)\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big). \end{align*} $$
The first term on the right side has
$|H_{4p}(N^{-\frac 1{4p}} (x_j-y_j))| \leq CN^{-\frac {(4p+1)}{4p}} |x_j-y_j|^{4p+1} e^{N^{-\frac 1{4p}}|x_j-y_j|}$
. Since the power of N is
$-1-\frac 1{4p}$
, and since the number of summands is
$N(t-s)$
we can use brutal absolute value bounds to obtain the desired estimate using arguments very similar to the bound for
$\mathcal E_N^2$
above. For the second term on the right side, one will obtain a sum of polynomials
$(x_1-y_1)^{k_1}(x_2-y_2)^{k_2}$
multiplied by a factor
$N^{-\frac {(k_1+k_2)}{4p}}$
with
$k_1+k_2>2p$
. For those terms where
$k_1+k_2>4p$
we can use brutal absolute value bounds to obtain the desired bound using arguments very similar to the bound for
$\mathcal E_N^2$
above. For the terms where
$2p+1\leq k_1+k_2\le 4p$
note that these are polynomials of a similar form to
$\mathcal E^4_N$
, with all negative powers of N being strictly larger than
$1/2.$
Thus we can use an argument very similar to the one obtained in bounding
$\mathcal E^4_N$
, thus completing the argument for bounding
$\mathcal E^3_N$
by the desired quantity.
Finally let us bound
$\mathcal E^1_N$
. To do this we will need to work with
$\mathcal A^1_N(\phi ,y_1,y_2)$
as defined just before (4.9). Let us perform a second-order Taylor expansion of
$\phi $
centered at
$N^{-1/2}y_1$
. Then the expression for
$\mathcal A^1_N(\phi ,y_1,y_2)$
can be written as a sum
$\mathcal A^{1,a}_N(\phi ,y_1,y_2)+\mathcal A^{1,b}_N(\phi ,y_1,y_2)$
where
$$ \begin{align*}\mathcal A^{1,a}_N(\phi,y_1,y_2)&:= \sum_{\substack{0\leq k_1,k_2\leq 2\\k_1+k_2>0}}\frac1{k_1!k_2!}N^{-(k_1+k_2)/2} \phi^{(k_1)}(N^{-1/2}y_1) \phi^{(k_2)}(N^{-1/2} y_1)\\& \hspace{2 cm} \cdot \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \cdot(x_1-y_1)^{k_1} (x_2-y_1)^{k_2}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big) \end{align*} $$
and by Taylor’s theorem the other term
$\mathcal A^{1,b}_N$
is a “remainder” satisfying
$$ \begin{align*}|\mathcal A^{1,b}_N(\phi,y_1,y_2)| \leq N^{-3/2}\|\phi\|_{C^3}^2 \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}|x_j-y_j| - \log M(N^{-\frac1{4p}})} \cdot|x_1-y_1|^3 |x_2-y_2|^3 \big|\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\big|.\end{align*} $$
As in (4.9) we can write
$\mathcal E^1_N = \mathcal E^{1,a}_N+ \mathcal E^{1,b}_N$
corresponding respectively to the contributions of
$\mathcal A^{1,a}_N$
and
$\mathcal A^{1,b}_N$
respectively.
The bound for
$\mathcal E^{1,b}_N$
is straightforward: since the power of N is
$-3/2$
, and since the number of summands in the expression for
$\mathcal E^{1,b}_N$
is
$N(t-s)$
, we can use brutal absolute value bounds to obtain the desired estimate using arguments very similar to the bound for
$\mathcal E_N^2$
above. By Lemma 2.9, the integral is bounded independently of
$y_1,y_2, N$
.
Let us discuss the bound for
$\mathcal E^{1,a}_N$
. In the expression for
$\mathcal A^{1,a}_N$
, each of the terms
$(x_2-y_1)^{k_2}$
can be further expanded out into a sum of terms of the form
$(x_2-y_2)^{i}(y_1-y_2)^{k_2-i}$
where
$0\leq i \leq k_2.$
Ultimately
$\mathcal A^{1,a}_N$
will be written as a sum of terms of the form
$$ \begin{align*}(y_1-y_2)^{k_2-i}N^{-(k_1+k_2)/2} &\phi^{(k_1)}(N^{-1/2}y_1) \phi^{(k_2)}(N^{-1/2} y_1)\\&\cdot \int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \cdot(x_1-y_1)^{k_1} (x_2-y_2)^{i}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\end{align*} $$
where
$0\leq k_1\le 2$
and
$0\le i\le k_2\leq 2$
. Assuming that
$\phi $
is supported on
$[-J,J]$
we can then take the absolute value to bound the last expression by
$$ \begin{align*}|y_1-y_2|^{k_2-i}N^{-(k_1+k_2)/2} &\|\phi\|_{C^2}^2\mathbf{1}_{[-J,J]}(N^{-1/2} y_1)\\&\cdot \bigg|\int_{I^2} \prod_{j=1}^2 e^{N^{-\frac1{4p}}(x_j-y_j) - \log M(N^{-\frac1{4p}})} \cdot(x_1-y_1)^{k_1} (x_2-y_2)^{i}\rho\big( (y_1,y_2), ({\mathrm{d}} x_1,{\mathrm{d}} x_2) \big)\bigg|.\end{align*} $$
All of this has accomplished the following: modulo the extra factor
$|y_1-y_2|^{k_2-i}N^{-(k_1+k_2)/2},$
we now have an expression that is very similar to the expression for
$\sum _{j=2,3,4} \mathcal A^j_N$
which we already dealt with when bounding
$\mathcal E^2,\mathcal E^3,\mathcal E^4$
. Consequently, the same arguments that were used to obtain the bounds for
$\mathcal E^2,\mathcal E^3,\mathcal E^4$
may now be used to bound
$\mathcal A^{1,a}_N$
. More precisely, one expands
$\prod _{j=1,2} e^{N^{-\frac 1{4p}}(x_j-y_j)}$
as a truncated Taylor expansion, where the degree of the
$(x_1-y_1)$
does not exceed
$4p-k_1$
and the degree of
$(x_2-y_2)$
does not exceed
$4p-i$
. All of the polynomial terms of joint degree less than
$2p$
vanish thanks to Assumption 1.2 Item (4), only leaving the polynomials of joint degree
$\ge 2p$
, and the remainder which will contribute at worst an extra
$N^{-1-\frac 1{4p}}$
. This will yield terms of the same form as those of
$\mathcal E^2_N$
and
$\mathcal E^4_N$
, albeit with the extra factor of
$|y_1-y_2|^{k_2-i}N^{-(k_1+k_2)/2}$
in the front.
By considering all of the different cases of
$(k_1,k_2,i)$
, one verifies that the decay conditions on the function
$\mathrm {F_{decay}}$
in Item (5) of Assumption (1.2) are strong enough so that one may find f as in the proposition statement to bound
$\mathcal E^{1,a}_N$
. For instance if
$k_2=2$
, then the term
$|y_1-y_2|^{k_2}$
seems problematic, but the power of N will be less than
$-1$
in this case (thanks to the Taylor expansion of
$\prod _{j=1,2} e^{N^{-\frac 1{4p}}(x_j-y_j)}$
contributing some factors of
$N^{-\frac {k}{4p}}$
), and since
$x\mapsto x^2 \mathrm {F_{decay}}(x)$
is bounded on
$[0,\infty )$
as can be verified from Item (5) of Assumption (1.2), a brutal absolute value bound will suffice (similar to dealing with the terms of type
$\mathcal E^2_N$
above). The worst terms will actually occur when
$k_2=1$
and
$i=k_1=0$
, as the power of N is exactly
$-1$
in this case, so that extracting an extra factor of
$N^{-\frac 1{4p}}$
is not possible to obtain brutal absolute value bounds. Instead, the condition that
$\int _0^\infty x\mathrm {F_{decay}}(x)<\infty $
ensures that we can take
$f(x):= (1+|x|)\mathrm {F_{decay}}(|x|),$
and all of the polynomials terms in the Taylor expansion of
$\prod _{j=1,2} e^{N^{-\frac 1{4p}}(x_j-y_j)}$
will guarantee that we can obtain a bound of the form
$N^{-1/2}|Q^f_N(t,\mathbf {1}_{[-J,J]})-Q^f_N(s,\mathbf {1}_{[-J,J]})|$
. By Assumption 1.2 Item (5), this f is eventually decreasing and thus bounded above by some decreasing function F that is still integrable, as required in the proposition statement. Meanwhile the remainder terms of the Taylor expansion ca be dealt with using admit brutal absolute value bounds as we saw earlier, because they come with a factor of
$N^{-1-\frac 1{4p}}$
in addition to the factor of
$|y_1-y_2|^{k_2-i}N^{-(k_1+k_2)/2}.$
Using Lemma 4.8 with
$q=0,1$
, or
$2$
(depending on the value of
$k_2-i)$
, one may easily deal with such remainder terms, thus finishing the proof of the required bound for
$\mathcal E^{1,a}_N.$
Corollary 4.10 (Predictable quadratic variation bound)
Let
$q\ge 1$
be an even integer, and let
$T>0$
. Let
$M^N(\phi )$
be the martingale defined in (4.2), and let
$\langle M^N(\phi )\rangle $
denote its predictable quadratic variation as in (4.4). Then there exists
$C>0$
and exists
$C>0$
and a function
$f:I\to {\mathbb {R}}_+$
such that
$|f(x)| \leq F(|x|)$
where
$F:[0,\infty )\to [0,\infty )$
is decreasing and
$\sum _{k=0}^\infty F(k)<\infty $
, such that we have the bound uniformly over all
$N\ge 1$
, all
$s,t\in [0,T]\cap (N^{-1}{\mathbb {Z}}),$
and all
$\phi \in C_c^\infty ({\mathbb {R}})$
supported on
$[-J,J]$
$$ \begin{align*}{\mathbb{E}}\big[ \big|\langle M^N(\phi)\rangle_{t}-\langle M^N(\phi)\rangle_s\big|^q\big]^{1/q}&\leq C \|\phi\|_{C^2}^2 {\mathbb{E}}[ |Q^f_N(t,\mathbf{1}_{[-J,J]})-Q^f_N(s,\mathbf{1}_{[-J,J]})|^q]^{1/q}+ C \|\phi\|_{C^3}^2 N^{-\frac1{4p}}|t-s|\\ &\leq C \|\phi\|_{C^3}^2|t-s|^{1/2}. \end{align*} $$
Proof. The first bound is immediate from Lemma 4.3 and Proposition 4.9. The second inequality follows from (4.21) and the fact that
$|t-s|\leq C|t-s|^{1/2}$
for
$s,t\in [0,T]$
.
Proposition 4.11 (Martingale tightness bound)
Fix
$q\ge 1$
and
$T>0$
. For all
$s,t\in (N^{-1}{\mathbb {Z}}_{\ge 0})\cap [0,T]$
and all
$\phi \in C_c^\infty ({\mathbb {R}})$
we have that
Here C is independent of
$N,\phi ,s,t$
.
Proof. It suffices to prove the claim when
$p=2k$
for some positive integer k. By [Reference Hall and HeydeHH14, Theorem 2.11] and Corollary 4.10, we have the Burkholder-type bound
$$ \begin{align*} {\mathbb{E}}\big[ \big|M_t^N(\phi)-M_s^N(\phi)\big|^q\big]^{1/q} & \leq C_p {\mathbb{E}} \big[ \big( \langle M^N(\phi)\rangle_t - \langle M^N(\phi)\rangle_s \big)^{q/2} \big]^{1/q} + {\mathbb{E}} \big[ \sup_{u \in N^{-1} {\mathbb{Z}} \cap [s,t]} |M_{u+N^{-1}}(\phi) - M_u(\phi)|^q \big]^{1/q}\\ &\leq C\|\phi\|_{C^{3}} \cdot |t-s|^{1/4} + {\mathbb{E}} \big[ \sup_{u \in N^{-1} {\mathbb{Z}} \cap [s,t]} |M_{u+N^{-1}}(\phi) - M_u(\phi)|^q \big]^{1/q} \\ &\le C\|\phi\|_{C^{3}} \cdot |t-s|^{1/4} + {\mathbb{E}} \bigg[ \sum_{u \in N^{-1} {\mathbb{Z}} \cap [s,t]} |M_{u+N^{-1}}(\phi) - M_u(\phi)|^q \bigg]^{1/q}. \end{align*} $$
We now claim that
$$ \begin{align} {\mathbb{E}} \bigg[ \sum_{u \in N^{-1} {\mathbb{Z}} \cap [s,t]} |M_{u+N^{-1}}(\phi) - M_u(\phi)|^q \bigg]^{1/q} \leq C|t-s|, \end{align} $$
which is stronger than is necessary to prove the lemma. To prove this, recall from (4.3) that when
$r=Nt$
we have
$$ \begin{align*} \notag (M^N_{t+N^{-1}}(\phi)- M^N_t(\phi))^q = \bigg( \int_I \int_I \phi(N^{-1/2}x) e^{N^{-\frac1{4p}}(x-y) - \log M(N^{-\frac1{4p}})}\big[ K_{r+1}(y,{\mathrm{d}} x) -\mu({\mathrm{d}} x-y)\big]Z_N(r,{\mathrm{d}} y)\bigg)^q. \end{align*} $$
For
$r\in {\mathbb {Z}}_{\ge 0}$
, define the signed (random) measure
$H_{r+1}(y,{\mathrm {d}} x):= K_{r+1}(y,{\mathrm {d}} x) -\mu ({\mathrm {d}} x-y). $
Then define

Note by Assumption 1.2 Item (3) that
$\int _I (y-x)^j H_{r+1}(y,{\mathrm {d}} x) =0$
for all
$0\le j \le p-1$
a.s.. Since
$q=2k$
, we take expectation of this quantity and we obtain

We claim that there exists
$C>0$
such that uniformly over
$y_1,...,y_{2k}\in I$
, and
$\phi \in C_c^\infty ({\mathbb {R}})$

To prove this, we claim a decomposition of
as the sum of three parts

where
$c_{m_1,m_2,...,m_{4k-1}}$
are deterministic constants not depending on N, where

and where the “remainder” term
satisfies for any
$q\ge 1$

Indeed, this three-term decomposition follows using exactly the same arguments as in deriving the error decomposition in Proposition 4.9. More precisely, in (4.27) one writes
$e^{N^{-\frac 1{4p}}(x_j-y_j)}$
as the sum
$H_{2p}(N^{-\frac 1{4p}}(x_j-y_j))+ \sum _{k=0}^{2p} \frac 1{k!}N^{-\frac {k}{4p}}(x_j-y_j)^k,$
then Taylor expanding each copy of
$\phi $
around
$N^{-1/2}y_1$
exactly as we did in the proof of Proposition 4.9 while bounding
$\mathcal E^1_N$
, and then noting that the constants
$c_{m_1,...,m_{4k-1}}$
can be obtained from combining all of the relevant Taylor coefficients of a given collection of exponents
$m_1,...,m_{4k-1}$
. If either of
$k_1$
or
$k_2$
is less than p the contribution in (4.27) would vanish a.s. by Assumption 1.2 Item (3), hence why we have the summation over
$p\le m_1,...,m_{2k}\le 2p$
rather than
$0\leq m_1,..., m_{2k}\leq 2p$
in the decomposition above. The remainder terms
will come from the Taylor expansion’s remainder beyond the order-two terms. Then (4.28) follows immediately from the three-term decomposition.
Henceforth assume
$k \ge 2$
, which we can assume without loss of generality since it suffices to prove the lemma for large k. Given (4.28), the required bound (4.26) is then immediate from Lemma 4.8. This is because for
$k\ge 2$
the first term of (4.28) is
$O(N^{-1})$
and thus easily dealt with using Lemma 2.9 and a brutal absolute value bound. The remaining terms are of even smaller order.
5 Tightness of the rescaled field and identification of the limit points
In this section we will finally prove Theorem 1.5 using the estimates of the previous section.
5.1 Weighted Hölder spaces and Schauder estimates
We now introduce various natural topologies for our field
$\mathscr {U}_N$
and its limit points. We then discuss how the heat flow affects these topologies and record Kolmogorov-type lemmas that will be key in showing tightness under these topologies. We begin by recalling many familiar and useful spaces of continuous and differentiable functions that have natural metric structures. For
$d\ge 1$
, we denote by
$C_c^{\infty }(\mathbb {R}^d)$
the space of all compactly supported smooth functions on
$\mathbb {R}^d$
. For a smooth function on
$\mathbb {R}^d$
, we define its
$C^r$
norm as
$$ \begin{align*} \|f\|_{C^r}:=\sum_{\#\vec{k} \le r} \sup_{{\mathbf{x}}\in \mathbb{R}^d} |D^{\vec k} f({\mathbf{x}})| \end{align*} $$
where the sum is over all
$\vec {k} = (k_1,...,k_d)\in \mathbb {Z}_{\ge 0}^d$
with
$\#\vec k:=\sum k_i\le r$
and
$D^{\vec {k}}:=\partial _{x_1}^{k_1}\cdots \partial _{x_d}^{k_d}$
denotes the mixed partial derivative.
We now recall the definition of weighted Hölder spaces from [HL15, Definitions 2.2 and 2.3]. For the remainder of this paper, we shall work with elliptic and parabolic weighted Hölder spaces with polynomial weight function
for some fixed
$\tau>1$
. We introduce these weights because weighted spaces will be more convenient to obtain tightness estimates. Since the solution of (1.5) started from Dirac initial condition is known to be globally bounded away from
$(0,0)$
, we expect that it is possible to remove the weights throughout this section, but this would require more precise moment estimates than the ones we derived in previous sections, which take into account spatial decay of the fields.
Definition 5.1 (Elliptic Hölder spaces)
For
$\alpha \in (0,1)$
we define the space
$C^{\alpha ,\tau }({\mathbb {R}})$
to be the completion of
$C_c^\infty ({\mathbb {R}})$
with respect to the norm given by
$$ \begin{align*}\|f\|_{C^{\alpha,\tau}({\mathbb{R}})}:= \sup_{x\in{\mathbb{R}}} \frac{|f(x)|}{w(x)} + \sup_{|x-y|\leq 1} \frac{|f(x)-f(y)|}{w(x)|x-y|^{\alpha}}.\end{align*} $$
For
$\alpha <0$
we let
$r=-\lfloor \alpha \rfloor $
and we define
$C^{\alpha ,\tau }({\mathbb {R}})$
to be the completion of
$C_c^\infty ({\mathbb {R}})$
with respect to the norm
$$ \begin{align*}\|f\|_{C^{\alpha,\tau}({\mathbb{R}})}:= \sup_{x\in{\mathbb{R}}} \sup_{\lambda\in (0,1]} \sup_{\phi \in B_r} \frac{(f,S^\lambda_{x}\phi)_{L^2({\mathbb{R}})}}{w(x)\lambda^\alpha}\end{align*} $$
where the scaling operators
$S^\lambda _{x}$
are defined by
and where
$B_r$
is the set of all smooth functions of
$C^r$
norm less than 1 with support contained in the unit ball of
${\mathbb {R}}$
.
One may verify that these spaces embed continuously into
$\mathcal S'({\mathbb {R}})$
.
Definition 5.2 (Function spaces)
Let
$C^{\alpha ,\tau }({\mathbb {R}})$
be as in Definition 5.1. We define
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
to be the space of continuous maps
$g:[0,T]\to C^{\alpha ,\tau }({\mathbb {R}}),$
equipped with the norm
Here and henceforth we will define
$\Psi _{[a,b]}:=[a,b]\times {\mathbb {R}}$
and we will define
$\Psi _T:=\Psi _{[0,T]}.$
So far we have used
$\phi ,\psi $
for test functions on
$\mathbb {R}$
. To make the distinction clear between test functions on
$\mathbb {R}$
and
$\mathbb {R}^2$
, we shall use variant Greek letters such as
$\varphi , \vartheta , \varrho $
for test functions on
$\mathbb {R}^2$
. In many instances below, we will explicitly write
$(f,\varphi )_{{\mathbb {R}}^2}$
or
$(g,\phi )_{{\mathbb {R}}}$
when we want to be clear about the space in which we are applying the
$L^2$
-pairing.
Definition 5.3 (Parabolic Hölder spaces)
We define
$C_c^\infty (\Psi _T)$
to be the set of functions on
$\Psi _T$
that are restrictions to
$\Psi _T$
of some function in
$C_c^\infty ({\mathbb {R}}^2)$
, in particular we do not impose that elements of
$C_c^\infty (\Psi _T)$
vanish at the boundaries of
$\Psi _T.$
For
$\alpha \in (0,1)$
we define the space
$C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)$
to be the completion of
$C_c^\infty (\Psi _T)$
with respect to the norm
$$ \begin{align*}\|f\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T)}:= \sup_{(t,x)\in \Psi_T} \frac{|f(t,x)|}{w(x)} + \sup_{|s-t|^{1/2}+|x-y|\leq 1} \frac{|f(t,x)-f(s,y)|}{w(x)(|t-s|^{1/2}+|x-y|)^{\alpha}}.\end{align*} $$
For
$\alpha <0$
we let
$r=-\lfloor \alpha \rfloor $
and we define
$C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)$
to be the completion of
$C_c^\infty (\Psi _T)$
with respect to the norm
$$ \begin{align*}\|f\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T)}:= \sup_{(t,x)\in\Psi_T} \sup_{\lambda\in (0,1]} \sup_{\varphi \in B_r} \frac{(f,S^\lambda_{(t,x)}\varphi)_{L^2(\Psi_T)}}{w(x)\lambda^\alpha}\end{align*} $$
where the scaling operators are defined by
and where
$B_r$
is the set of all smooth functions of
$C^r$
norm less than 1 with support contained in the unit ball of
${\mathbb {R}}^2$
.
An important property of both the parabolic and elliptic spaces is that one has a continuous embedding
$C^{\alpha ,\tau } \hookrightarrow C^{\beta ,\tau }$
whenever
$\beta <\alpha .$
In fact this embedding is compact, though we will not use this. We also have the following embedding of function spaces inside parabolic spaces.
Lemma 5.4. For
$\alpha <0,\tau>0$
one has a continuous embedding
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))\hookrightarrow C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)$
given by identifying
$v=(v(t))_{t\in [0,T]}$
with the tempered space-time distribution given by
$$ \begin{align*}(v,\varphi)_{{\mathbb{R}}^2} = \int_0^T (v(t) ,\varphi(t,\cdot))_{\mathbb{R}}dt.\end{align*} $$
The proof is straightforward from the definitions and is omitted.
Remark 5.5 (Derivatives of distributions)
Let
$\alpha <0$
. By definition, any element
$f\in C_{\mathfrak {s}}^{\alpha ,\tau }(\Psi _T)$
admits an
$L^2$
-pairing with any smooth function
$\varphi : \Psi _T\to {\mathbb {R}}$
of rapid decay (i.e.,
$\varphi $
is a Schwartz function restricted to
$\Psi _T$
). We will write this pairing as
$(f,\varphi )_{\Psi _T}.$
Consequently there is a natural embedding
$C_{\mathfrak {s}}^{\alpha ,\tau }(\Psi _T)\hookrightarrow \mathcal S'({\mathbb {R}}^2)$
which is defined by formally setting f to be zero outside of
$[0,T]\times {\mathbb {R}}$
. More rigorously, this means that the
$L^2({\mathbb {R}}^2)$
-pairing of f with any
$\varphi \in \mathcal S({\mathbb {R}}^2)$
is defined to be equal to
$(f,\varphi |_{\Psi _T})_{\Psi _T}$
.
The image of this embedding consists of some specific collection of tempered distributions that are necessarily supported on
$[0,T]\times {\mathbb {R}}.$
Thus we can sensibly define
$\partial _tf$
and
$\partial _xf$
as elements of
$\mathcal S'({\mathbb {R}}^2)$
whenever
$f\in C_{\mathfrak {s}}^{\alpha ,\tau }(\Psi _T)$
, by the formulas
for all smooth
$\varphi :\Psi _T\to {\mathbb {R}}$
of rapid decay. This convention on derivatives will be useful for certain computations later. From the definitions, when
$\alpha <0$
one can check that for such f one necessarily has
$\partial _tf \in C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T)$
and
$\partial _x f \in C_{\mathfrak {s}}^{\alpha -1,\tau }(\Psi _T).$
The latter statement fails for
$\alpha>0$
. Indeed by our convention of derivatives,
$\partial _tf$
may no longer be a smooth function (or even a function) even if
$f\in C_c^\infty (\Psi _T)$
. This is because such an f gets extended to all of
${\mathbb {R}}^2$
by setting it to be zero outside
$\Psi _T$
. In particular, if f does not vanish on the boundary of
$\Psi _T$
, then it may become a discontinuous function under our convention of extension to
${\mathbb {R}}^2$
. Due to these discontinuities, the distributional derivative
$\partial _tf$
can be a tempered distribution with singular parts along the boundaries. In our later computations, we never take derivatives of functions in
$C_{\mathfrak {s}}^{\alpha ,\tau }(\Psi _T)$
with
$\alpha>0$
. Sometimes we will write
$\partial _s$
to mean the same thing as
$\partial _t$
.
Proposition 5.6 (Smoothing effect of the heat flow on elliptic spaces)
For
$f\in C_c^\infty ({\mathbb {R}})$
and
$t>0$
define
where
$p_t(x) = (2\pi t)^{-1/2} e^{-x^2/2t}$
is the standard heat kernel. Then for all
$\alpha \le \beta <1$
(possibly negative), there exists
$C=C(\alpha ,\beta ,T)>0$
such that
uniformly over
$f\in C_c^\infty $
and
$t\in [0,T]$
. In particular,
$P_t$
extends to a globally defined linear operator on
$C^{\alpha ,\tau }({\mathbb {R}})$
which maps boundedly into
$C^{\beta ,\tau }({\mathbb {R}}).$
A proof may be found in [HL15, Lemma 2.8] in the case of an exponential weight. The proof for polynomial weights is very similar.
For the parabolic Hölder spaces, the following lemma states that the heat flow improves the regularity by a factor of
$2$
and provides a Schauder-type estimate.
Proposition 5.7 (Schauder estimate)
For
$f\in C_c^\infty (\Psi _T)$
let us define
where
$p_t$
is the standard heat kernel for
$t>0,x\in {\mathbb {R}}$
and
$p_t(x):=0$
for
$t<0.$
Then for all
$\alpha <-1$
with
$\alpha \notin {\mathbb {Z}}$
there exists
$C=C(\alpha )>0$
independent of f such that
In particular K extends to a globally defined linear operator on
$C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)$
that maps boundedly into
$C^{\alpha +2,\tau }_{\mathfrak {s}}(\Psi _T).$
Furthermore, if
$f\in C^{\alpha ,\tau }_{\mathfrak {s}}(\Lambda _{T})$
, then
$K(\partial _t-\frac 12\partial _x^2)f= f$
.
It is important to note that the last statement is only true because of our convention on distributional derivatives that we have explained in Remark 5.5. Without that convention, that statement would be false even for a smooth function f that does not vanish on the line
$\{t=0\}$
.
Proof. See [Reference Das, Drillick and ParekhDDP24a, Proposition 6.6] for the proof.
Corollary 5.8. Define
$J: C^{\alpha ,\tau }({\mathbb {R}}) \to C_{\mathfrak {s}}^{\alpha ,\tau }(\Psi _T)$
by
J is a bounded linear operator for any
$\alpha <1$
and any
$\tau>0.$
Consider the operator
$\widehat P_t:= JP_t$
with
$P_t$
defined in Proposition 5.6. By the semigroup property of the heat kernel, we have that
Furthermore, we have the operator norm bound
where
$C=C(\alpha ,\beta ,T)>0$
does not depend on t and f.
Proof. The first part follows by using the Schauder estimate in Proposition 5.7, noting that
$Jf = K(\delta _0\otimes f)$
where for
$f \in C^{\alpha ,\tau }({\mathbb {R}})$
the latter distribution is defined by
$(\delta _0\otimes f, \varphi )_{{\mathbb {R}}^2}:= (f,\varphi (0,\cdot ))_{{\mathbb {R}}}.$
Directly from the definitions one can check that
$f\mapsto \delta _0\otimes f$
is bounded from
$C^{\alpha ,\tau }({\mathbb {R}})\to C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T).$
Finally, (5.5) follows from Proposition 5.6.
We next define a space-time distribution that is supported on a single temporal cross-section:
Definition 5.9. Let
$\alpha <0$
. Given some
$f\in C^{\alpha ,\tau }({\mathbb {R}})$
and
$b \in [0,T]$
we define
$\delta _b \otimes f\in C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T)$
by the formula
$(\delta _b\otimes f, \varphi )_{{\mathbb {R}}^2}:= (f,\varphi (b,\cdot ))_{{\mathbb {R}}}.$
Directly from the definitions of the scaling operators in (5.1) and (5.2), it is clear that for fixed
$b\in [0,T]$
, the linear map
$f\mapsto \delta _b\otimes f$
is bounded from
$C^{\alpha ,\tau }({\mathbb {R}})\to C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T)$
as long as
$\alpha <0.$
It is also clear that
$\delta _b \otimes f$
is necessarily supported on the line
$\{b\}\times {\mathbb {R}}$
.
We end this subsection by recording a Kolmogorov-type lemma for the three spaces introduced at the beginning of this subsection. It will be crucial in proving tightness in those respective spaces.
Lemma 5.10 (Kolmogorov lemma)
Let
$L^2(\Omega ,\mathcal {F},\mathbb {P})$
be the space of all random variables defined on a probability space
$(\Omega ,\mathcal {F},\mathbb {P})$
with finite second moment. We have the following:
-
1. (Function space) Let
$(t,\phi ) \mapsto V(t,\phi )$
be a map from
$[0,T]\times \mathcal S({\mathbb {R}})$
into
$L^2(\Omega ,\mathcal F,\mathbb P)$
which is linear and continuous in
$\phi $
. Fix a non-negative integer r. Assume there exists some
$\kappa>0, q>1/\kappa $
and
$\alpha <-r$
and
$C=C(\kappa ,\alpha ,q,{T})>0$
such that one has uniformly over all smooth functions
$$ \begin{align*}\mathbf{E}[|V(t,S^\lambda_{x}\phi)|^q]^{1/q}&\leq C\lambda^{\alpha },\\ \mathbf{E}[|V(t,S^\lambda_{x}\phi)-V(s,S^\lambda_{x}\phi)|^q]^{1/q}&\leq C\lambda^{\alpha -\kappa} |t-s|^{\kappa }, \end{align*} $$
$\phi $
on
${\mathbb {R}}$
supported on the unit ball of
${\mathbb {R}}$
with
$\|\phi \|_{C^r}\leq 1$
, and uniformly over
$\lambda \in (0,1]$
and
$0\leq s,t\leq T$
. Then for any
$\tau>1$
and any
$\beta <\alpha -\kappa $
there exists a random variable
$\big (\mathscr V(t)\big )_{t\in [0,T]}$
taking values in
$C([0,T],C^{\beta ,\tau }({\mathbb {R}}))$
such that
$(\mathscr V(t),\phi )=V(t,\phi )$
almost surely for all
$\phi $
and t. Furthermore, one has that where
$$ \begin{align*}\mathbf{E}[\|\mathscr V\|^q_{C([0,T],C^{\beta,\tau}({\mathbb{R}}))}]\leq C',\end{align*} $$
$C'$
depends on the choice of
$\alpha ,\beta ,q,\kappa ,$
and the constant C appearing in the moment bound above but not on
$V,\Omega ,\mathcal F,\mathbb P$
.
-
2. (Parabolic Hölder Space) Let
$\varphi \mapsto V(\varphi )$
be a bounded linear map from
$\mathcal S({\mathbb {R}}^2)$
to
$L^2(\Omega ,\mathcal F,\mathbb P)$
. Assume
$V(\varphi )=0$
for all
$\varphi $
with support contained in the complement of
$\Psi _T$
. Recall
$S^\lambda _{(t,x)}$
from (5.2). Fix a non-negative integer r. Assume there exists some
$q>1$
and
$\alpha <0$
and
$C=C(\alpha ,q)>0$
such that one has uniformly over all smooth functions
$$ \begin{align*}\mathbf{E}[|V(S^\lambda_{(t,x)}\varphi)|^q]^{1/q}\leq C\lambda^{\alpha },\end{align*} $$
$\varphi $
on
${\mathbb {R}}^2$
supported on the unit ball of
${\mathbb {R}}^2$
with
$\|\varphi \|_{C^r}\leq 1$
, and uniformly over
$\lambda \in (0,1]$
and
$(t,x)\in \Psi _T$
. Then for any
$\tau>1$
and any
$\beta <\alpha -3/q$
there exists a random variable
$\mathscr V$
taking values in
$C^{\beta ,\tau }_{\mathfrak {s}}(\Psi _T)$
such that
$(\mathscr V,\varphi )=V(\varphi )$
almost surely for all
$\varphi .$
Furthermore one has that where
$$ \begin{align*}\mathbf{E}[\|\mathscr V\|^q_{C^{\beta,\tau}_{\mathfrak{s}}(\Psi_T)}]\leq C',\end{align*} $$
$C'$
depends on the choice of
$\alpha ,q,$
and the constant C appearing in the moment bound above but not on
$V,\Omega ,\mathcal F,\mathbb P$
.
A proof of the above results may be adapted from the proof of Lemma 9 in Section 5 of [Reference Mourrat and WeberMW17]. We remark that we do not actually need uniformity over a large class of test functions as we have written above, just a single well-chosen test function would suffice (e.g., the Littlewood-Paley blocks as used in [Reference Mourrat and WeberMW17] or the Daubechies wavelets in [HL15]).
5.2 Tightness
Throughout this section, we are going to fix a terminal time
$T>0$
. Let
$\mu $
and
$\eta $
be as in Assumption 1.2 Item (3). As in the beginning of Section 4, we denote
$\mu ^\lambda ({\mathrm {d}} x):= e^{\lambda x - \log M(\lambda )} \mu ({\mathrm {d}} x).$
We claim that for any
$f\in \mathcal S'({\mathbb {R}})$
and
$\lambda \in [-\eta ,\eta ]$
the spatial convolution
$g:= f*\mu ^\lambda $
given by
is well-defined as another element of
$\mathcal S'({\mathbb {R}})$
. To prove this, we must show that
$\phi \mapsto \phi * \mu ^\lambda $
is a well-defined and continuous map from
$\mathcal S({\mathbb {R}})$
into itself. By taking the Fourier transform, it suffices to show that the multiplication operator
$T_\lambda \psi (\xi ) = \widehat {\mu ^\lambda }(\xi ) \psi (\xi )$
is a continuous linear operator from
$\mathcal S({\mathbb {R}})\to \mathcal S({\mathbb {R}})$
. But this is clear from the fact that the Fourier transform
$\widehat {\mu ^\lambda }$
(i.e., the “characteristic function” of the measure
$\mu ^\lambda $
) is smooth with all derivatives bounded, since the measure
$\mu ^\lambda $
has exponentially decaying tails.
Likewise if
$f\in \mathcal S'({\mathbb {R}}^2)$
and one would like to define a tempered distribution
$g\in \mathcal S'({\mathbb {R}}^2)$
such that formally one has
$g(t,x) = \int _{{\mathbb {R}}} f(t,x-y) \mu ^\lambda ({\mathrm {d}} y),$
the procedure is completely analogous, defining it via the Fourier transforms
$\widehat g(\xi _1,\xi _2):= \widehat {\mu ^\lambda }(\xi _2) \widehat f(\xi _1,\xi _2).$
Definition 5.11. Define
Definition 5.12. For
$(s,y) \in {\mathbb {Z}}\times I$
define
$p_N(s,dy)$
to be the transition density at time s and position y of a random walker on
$\Lambda _N$
with increment distribution
$\mu ^{N^{-\frac 1{4p}}}.$
Define linear operators
$D_N,L_N,K_N$
on
$\mathcal S'({\mathbb {R}}^2)$
by
$$ \begin{align*} D_Nf(t,x) &= N \big[ f(t+N^{-1},x)-f(t,x)\big],\\ L_N f(t,x) &= N\bigg[f(t+N^{-1},x-N^{-1}d_N) -\int_I f(t,x-N^{-1/2}y) \mu^{N^{-\frac1{4p}}} ({\mathrm{d}} y)\bigg],\\ K_N f(t,x) &= N^{-1}\sum_{s\in [0,T]\cap (N^{-1}{\mathbb{Z}})} \int_{I} f(t-s,x-N^{-1/2}y)p_N(Ns,{\mathrm{d}} y). \end{align*} $$
These equalities should be understood by integration against smooth functions
$\varphi \in \mathcal S({\mathbb {R}}^2)$
, and the convolutions need to be understood using the explanation after (5.6).
Intuitively it is clear that
$L_N$
is a diffusively rescaled version of the discrete heat operator
$\mathcal L_N$
from Section 4, but which acts on tempered distributions rather than measures on
${\mathbb {Z}}\times I$
. Indeed if
$\varphi \in \mathcal S({\mathbb {R}}^2)$
then a second-order Taylor expansion shows that
$L_N\varphi $
converges in
$\mathcal S({\mathbb {R}}^2)$
as
$N\to \infty $
to
$(\partial _t-\frac 12\partial _x^2)f$
, that is,
$L_N$
approaches the continuum heat operator. Then
$K_N$
is the inverse operator to
$L_N$
, in the following sense.
Lemma 5.13.
$L_NK_Nf=K_NL_N f=f$
whenever f is a tempered distribution supported on
$[a,b]\times {\mathbb {R}}$
with
$b-a<T+1$
.
Proof. Note that for each N the operators
$K_N,L_N$
are continuous on
$\mathcal S'({\mathbb {R}}^2)$
, as may be verified using the Fourier transform as explained after (5.6). Therefore it suffices to prove the claim for all smooth functions f that have compact support contained in
$(a,b)\times {\mathbb {R}}$
, since these are dense in the subset of distributions supported on
$[a,b]\times {\mathbb {R}}$
, with respect to the topology of
$\mathcal S'({\mathbb {R}}^2)$
. The smooth analogue is true by direct calculations, since
$p_N(s,{\mathrm {d}} y)$
is the kernel for the inverse operator to the discrete heat operator
$\mathcal L_N$
introduced in Section 4.
Definition 5.14. With the fields
$M_N$
and
$Q_N$
as defined in (4.2) and (4.10) respectively, we will associate random elements of
$C([0,T+1],\mathcal S'({\mathbb {R}}))$
by the formulas
$\widehat M_N(t,\phi )=0$
for
$t\in N^{-1}$
and
$$ \begin{align*}\widehat{M}_N(t,\phi) &:= M^N_t(\phi),\;\;\;\;\; t\geq 0. \\ \widehat{Q}_N^f(t,\phi)&:= Q_N^f(t,\phi) \end{align*} $$
for
$\phi \in \mathcal S({\mathbb {R}})$
and
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
. We define these fields by linear interpolation for
$t\notin N^{-1}{\mathbb {Z}}_{\ge 0}.$
For any T, note that
$C([0,T],\mathcal S'({\mathbb {R}}))$
embeds naturally into the linear subspace of
$\mathcal S'({\mathbb {R}}^2)$
consisting of distributions supported on
$[0,T]\times {\mathbb {R}}$
, thus we can make sense of
$D_Nf,L_Nf,K_Nf$
for all
$f\in C([0,T],\mathcal S'({\mathbb {R}})),$
and these will be elements of
$\mathcal S'({\mathbb {R}}^2)$
in general.
Definition 5.15. We will say that two tempered distributions
$f,g \in \mathcal S'({\mathbb {R}}^2)$
agree on
$[0,T]$
if there exists
$\varepsilon>0$
such that
$(f,\varphi ) = (g,\varphi )$
for all
$\varphi $
supported on
$[-\varepsilon ,T+\varepsilon ]\times {\mathbb {R}}.$
Definition 5.16. Then define the distribution
$\mathfrak p_N\in C([0,T+1],\mathcal S'({\mathbb {R}}))$
for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
by
and linearly interpolated for
$t\notin N^{-1}{\mathbb {Z}}_{\ge 0}$
.
Since
$\mathfrak p_N$
is deterministic, its scaling limit will simply be the continuum heat kernel.
Lemma 5.17. Let
$\widehat M_N$
be as in Definition 5.14. Furthermore, let
$\mathfrak H^N$
be as defined in (1.4). Restrict
$\mathfrak H^N$
to
$[0,T]$
thus viewed as an element of
$C([0,T],\mathcal S'({\mathbb {R}}))$
. Then
$L_N(\mathfrak H^N-\mathfrak p_N)$
agrees with
$D_N\widehat {M}_N$
on
$[0,T]$
. Consequently
$\mathfrak H^N$
agrees with
$\mathfrak p_N+K_N D_N\widehat {M}_N$
on
$[0,T]$
.
Proof. Firstly, notice that
$L_N \mathfrak p^N (t,\cdot ) = 0$
for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
. Let
$V^N := D_NM^N$
. Then it is clear from (4.2) that
for all
$0\le t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
and
$\phi \in \mathcal S({\mathbb {R}}^d)$
. Now, with
$Z_N$
defined in (4.1), for
$t\in N^{-1}\mathbb N$
we have
With
$\mathcal L_N$
as defined in Section 4, we than apply the convolution defining
$L_N$
to both sides and it is then clear that for
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
the expression for
$L_N\mathfrak H^N(t,\cdot )$
can be written as
which is the same as the expression for
$V^N(t,\phi ).$
Now we need to show that equality holds even if
$t\notin N^{-1}{\mathbb {Z}}_{\ge 0}$
. Since linear operators respect linear interpolation, and since all of the fields
$\mathfrak H^N, V_N, M^N$
are defined via linear interpolation, this is actually immediate.
Finally, if we view the restriction of
$V^N$
to
$[0,T]$
as an element of
$\mathcal S'({\mathbb {R}}^{d+1})$
supported on
$[0,T]\times {\mathbb {R}}$
, then we can apply
$K_N$
to both sides and we obtain that
$(\mathfrak H^N-\mathfrak p^N)(t,\cdot ) = K_N V^N(t,\cdot ) = K_ND_NM^N(t,\cdot )$
for all
$t\in [0,T]$
by Lemma 5.13.
Lemma 5.1. Fix
$\alpha <0,\tau>0$
. Let
$K_N,D_N$
be the operators from Definition 5.12. Recall the function spaces from Definition 5.2. Then we have the operator norm bound
Proof. We will adapt the argument from the continuum version of this lemma proved in [Reference Das, Drillick and ParekhDDP24a, Proof of Theorem 1.4(b)].
Note that
$D_Nv$
is indeed a continuous path taking values in
$C([0,T+1-N^{-1}], C^{\alpha ,\tau }(\mathbb {R}))$
, consequently so is
$K_ND_Nv$
. Let us now define four families of linear operators on
$\mathcal S'({\mathbb {R}})$
, indexed by N. For
$f\in \mathcal S'({\mathbb {R}})$
and
$s\in N^{-1}{\mathbb {Z}}_{\ge 0}$
let
$$ \begin{align*}P_N(s)f(x) &:= \int_I f(x-N^{-1/2}y) p_N(Ns,{\mathrm{d}} y), \\ \delta_Nf(x) &:= N(P_N(N^{-1})-\mathrm{Id})f (x) =N \int_I \big(f(x-N^{-1/2}y) -f(x)\big)p_N(Ns,{\mathrm{d}} y) \\ \delta_N^* f(x) &:= N \int_I \big(f(x+N^{-1/2}y) -f(x)\big)p_N(Ns,{\mathrm{d}} y) , \\ P^*_N(s) f(x)& := \big(\mathrm{Id}+N^{-1}\delta_N^*\big)^{Ns}f(x)=\int_I f(x+N^{-1/2}y) p_N(Ns,{\mathrm{d}} y). \end{align*} $$
Then for
$f\in \mathcal S'({\mathbb {R}})$
each of the four expressions
$P_N(s)f, \delta _N f, \delta _N^*f, P_N^*(s)f$
also make sense as elements of
$\mathcal S'({\mathbb {R}})$
, see, for example, the discussion following (5.6). Note that
$\delta _N$
and
$\delta ^*_N$
, both of which approximate the spatial Laplacian
$\partial _x^2$
, are adjoint to each other on
$L^2({\mathbb {R}})$
. Therefore
$P_N$
and
$P^*_N$
are also adjoint to each other. Now let
$v = (v(t))_{t\in [0,T+1]} \in \mathcal C([0,T+1],C^{\alpha ,\tau }({\mathbb {R}}))$
. For
$\phi \in C_c^\infty ({\mathbb {R}})$
and
$t\in [0,T]$
, if we apply summation by parts and use the fact that v vanishes on
$[0,N^{-1}]$
we can write
$$ \begin{align*}\big( K_ND_N v(t),\phi \big) &= \bigg( N^{-1} \sum_{s\in [0,t]\cap( N^{-1}{\mathbb{Z}}_{\ge 0})} NP_N(t-s)\big[ v(s+N^{-1})-v(s)\big]\;\;, \;\; \phi \bigg) \\ &= \bigg( v(t+N^{-1}) - 0 + \sum_{s\in [0,t]\cap( N^{-1}{\mathbb{Z}}_{\ge 0})} \big[ P_N(t-s)-P_N(t-s-N^{-1}) \big] (v(s))\;\;, \;\; \phi \bigg) \\ &= (v(t+N^{-1}),\phi) + \frac1N \sum_{s\in [0,t]\cap( N^{-1}{\mathbb{Z}}_{\ge 0})} (\delta_N P_N(t-s-N^{-1}) v(s) , \phi) \\ &= (v(t+N^{-1}),\phi) + \frac1N \sum_{s\in [0,t]\cap( N^{-1}{\mathbb{Z}}_{\ge 0})} (v(s), P_N^*(t-s-N^{-1})\delta^*_N \phi). \end{align*} $$
Here we are using the
$L^2({\mathbb {R}})$
-pairing between distributions and smooth functions. By the definition of the function spaces (Definition 5.2), we must replace
$\phi $
by
$S^\lambda _x\phi $
(where the scaling operators are given in Definition 5.1) in the last expression and then study the growth as
$\lambda $
becomes close to 0.
The first term on the right side is completely straightforward to deal with: the growth is at worst
$\lambda ^\alpha (1+x^2)^\tau $
uniformly over
$\lambda , \phi , x, T$
, since we assumed
$v\in C([0,T],C^{\alpha ,\tau }({\mathbb {R}})).$
To deal with the second term, we claim that one has
uniformly over
$s<t\in [0,T]$
, as well as
$\lambda ,\phi ,x, N$
and v with
$\|v\|_{C([0,T+1],C^{\alpha ,\tau }({\mathbb {R}}))}\leq 1$
. Indeed the bound of the form
$\lambda ^{\alpha -2}$
follows by noting that when
$t-s$
is much smaller than
$\lambda $
,
$P_N^*(t-s)$
is essentially the identity operator, so we can effectively disregard the heat kernel and note that
$\delta _N^* S^\lambda _x\phi $
paired with
$v(s)$
satisfies a bound of order
$\lambda ^{\alpha -2}$
, uniformly over N, for example, by second-order Taylor expansion. Likewise the bound of the form
$(t-s)^{\frac {\alpha }2-1}$
is obtained by noting that when
$\lambda $
is very small compared to
$t-s$
,
$P_N^*(t-s-N^{-1})\delta ^*_N S^\lambda _x\phi $
behaves like
$S_x^{\sqrt {t-s}}\phi $
, giving a bound of order
$(\sqrt {t-s})^{\alpha -2}$
after applying
$\delta ^*_N$
and pairing with
$v(s)$
. This proves the bound (5.7).
Now the fact that
$\|K_ND_Nv\|$
can be controlled by
$\|v\|$
follows simply by noting that uniformly over
$0<\lambda ^2 \leq t\leq T$
one has
$$ \begin{align*}\int_0^t \big(\lambda^{\alpha-2} \wedge (t-s)^{\frac{\alpha}2-1}\big) ds = \int_0^{t-\lambda^2}(t-s)^{\frac{\alpha}2-1}ds + \int_{t-\lambda^2}^t \lambda^{\alpha-2}ds \leq C(\alpha)\cdot \lambda^\alpha.\\[-40pt] \end{align*} $$
Proposition 5.18 (Tightness of all relevant processes)
The following are true.
-
1. The fields
$\widehat {M}_N$
from Definition 5.14 may be realized as an element of
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
for any
$\alpha <-5$
and
$\tau>1$
. Moreover, they are tight with respect to that topology. -
2. The fields
$\mathfrak H^N$
from (1.4) may be realized as an element of as an element of
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
for any
$\alpha <-5$
and
$\tau>1$
. Moreover, they are tight with respect to that topology. -
3. For any
$f:I\to {\mathbb {R}}$
of exponential decay, the fields
$\widehat Q_N^f$
from Definition 5.14 may be realized as an element of
$C([0,T],C^{\gamma ,\tau }({\mathbb {R}}))$
for any
$\gamma <-1$
and
$\tau>1$
. Moreover, they are tight with respect to that topology. -
4. Let
$\alpha ,\gamma <0$
be as in the previous items. Let
$(M^\infty ,Q^\infty , H^\infty )$
be a joint limit point of
$(\widehat M_N,\widehat Q_N^{\boldsymbol {\zeta }},\mathfrak H^N)$
in the space
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}})\times C^{\gamma ,\tau }({\mathbb {R}})\times C^{\alpha ,\tau }({\mathbb {R}})),$
where
$\boldsymbol {\zeta }$
is the specific function from Definition 2.11. For all
$\phi \in C_c^\infty ({\mathbb {R}})$
the process
$(M_t^\infty (\phi ))_{t\in [0,T]}$
is a continuous martingale with respect to the canonical filtration on that space, and moreover its quadratic variation is given by (5.8)
$$ \begin{align} \langle M^\infty(\phi)\rangle_t = Q_t^\infty(\phi^2). \end{align} $$
Proof. Take any
$\phi \in C_c^\infty ({\mathbb {R}})$
with
$\|\phi \|_{L^\infty }\leq 1.$
Recall
$S_{(t,x)}^{\lambda }$
from (5.2). Using the first bound in Proposition 4.7, we have
uniformly over
$x\in {\mathbb {R}}, \lambda \in (0,1], 0\leq s,t\leq T$
with
$s,t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
. Since
$Q_N(0,\phi )=0$
by definition, the assumptions of Lemma 5.10 1 are therefore satisfied for any
$\kappa \leq 1/4$
, any
$p>1/\kappa $
, and any
$\alpha \leq -1$
, and we conclude the desired tightness for
$\widehat Q_N=Q_N$
. This proves Item (3).
Now we address the tightness of the
$\widehat M_N$
. Using Proposition 4.11, we have that
where
$\phi \in C_c^\infty ({\mathbb {R}})$
,
$C=C(k)>0$
is free of
$\phi ,s,t,N.$
This gives
uniformly over
$x\in {\mathbb {R}}, \lambda \in (0,1], 0\leq s,t\leq T$
, and
$\phi \in C_c^\infty ({\mathbb {R}})$
with
$\|\phi \|_{C^3}\leq 1$
with support contained in the unit interval. Moreover
$\widehat M_N(0,\phi )=0$
by definition, therefore the assumptions of Lemma 5.10 1 are satisfied with
$\kappa =1/4$
,
$q=8>1/\kappa $
, and any
$\alpha \leq -4$
. Hence, we conclude the desired tightness for
$\widehat M_N$
. This proves Item (1).
Now tightness for the fields
$\mathfrak H^N$
is immediate from Lemma 1, since we know from Lemma 5.17 that
$\mathfrak H^N = \mathfrak p_N + K_ND_N\widehat M_N$
where we view
$K_ND_N $
as a bounded operator from
$\mathcal C([0,T+1],C^{\alpha ,\tau }({\mathbb {R}}))\to C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
and the convergence of
$\mathfrak p_N$
in this topology is straightforward to deal with. This proves Item (2).
We next show that the limit point
$M^{\infty }(\phi )$
is a martingale indexed by
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
. Since
$M^N_0(\phi )=0$
, from (5.9), we see that for any
$q>1$
one has
$\sup _N \mathbb {E}[M^N_t(\phi )^q]<\infty .$
Thus
$M^{\infty }(\phi )$
is a martingale since martingality is preserved by limit points under the uniform integrability assumption. Continuity is guaranteed by the definition of the spaces in which we proved tightness. In the prelimit we know from Lemma 4.3 that
$$ \begin{align*}M^N_t(\phi)^2-e^{-2\log M(N^{-\frac1{4p}})} Q^{\boldsymbol{\zeta}}_N(t,\phi^2)-\sum_{j=1}^4 \mathcal E^j_N(t,\phi)\end{align*} $$
is a martingale indexed by
$t\in N^{-1}{\mathbb {Z}}_{\ge 0}$
, where the error terms
$\mathcal E^j_N$
are defined in (4.9) and (4.8), and satisfy the bounds in Proposition (4.9). By the latter proposition and the tightness estimates (4.21) of
$Q_N^f$
, it follows that
$\sum _{j=1}^4 \mathcal E_N^j(t,\phi )$
vanishes in probability (in the topology of
$C[0,T]$
for each
$\phi \in C_c^\infty ({\mathbb {R}})$
), so we conclude (again by uniform
$L^p$
boundedness guaranteed by Proposition 4.7) that
$M_t^\infty (\phi )^2-Q_t^\infty (\phi ^2)$
is a martingale. This verifies (5.8) completing the proof of Item (4).
5.3 Identification of limit points
Lemma 5.19 (Controlling the difference between the discrete and continuum time-derivatives)
Fix
$\alpha <0, \tau>1$
. The derivative operator
$\partial _s : C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T) \to C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T)$
which was defined in Remark 5.5, is a bounded linear map. Furthermore, let
$D_N$
be as in Definition 5.12. Then we have the operator norm bounds
$$ \begin{align*}\sup_{N\ge 1}\|D_N\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T)\to C_{\mathfrak{s}}^{\alpha-2,\tau}(\Psi_T)}&<\infty. \\ \|D_N-\partial_s\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T) \to C^{\alpha-4,\tau}_{\mathfrak{s}}(\Psi_T)} &\leq \frac12 N^{-1}. \end{align*} $$
Proof. The proof is fairly immediate from the definitions, see [Reference Das, Drillick and ParekhDDP24b, Lemma 6.21] for the complete argument.
Lemma 5.20 (Controlling the difference between the discrete and continuum heat operators)
Fix
$\alpha <0, \tau>1$
. Let
$K_N$
be as in Definition 5.12, and let K be as in (5.3). Then we have the operator norm bound
Here C is independent of N.
The above bound is crude, we do not claim optimality of the Hölder exponents here.
Proof. Define
$f_\varphi ^\lambda (t,x):= (f,S^\lambda _{(t,x)}\varphi )_{L^2(\Psi _T)}$
. It suffices to show that
$$ \begin{align*}\sup_{\|f\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T)}\leq 1} \sup_{(t,x)\in\Psi_T}\sup_{\lambda\in [(0,1]} \sup_{\varphi \in B_r} \frac{|(K_N-K)f_\varphi^\lambda(t,x)|}{(1+x^2)^{\tau}\lambda^{\alpha-1}}\leq CN^{-1/4},\end{align*} $$
where the scaling operators are defined by
$S^\lambda _{(t,x)}\varphi (s,y) = \lambda ^{-3}\varphi (\lambda ^{-2}(t-s),\lambda ^{-1}(x-y)) ,$
and where
$B_r$
is the set of all smooth functions of
$C^r$
norm less than 1 with support contained in the unit ball of
${\mathbb {R}}^2$
. We shall use a probabilistic interpretation of the kernels to prove this. Note that
$$ \begin{align*}(K_N-K)f_\varphi^\lambda (t,x) = \frac1N \sum_{s\in [0,T]\cap (N^{-1}{\mathbb{Z}})} \mathbf E[f_\varphi^\lambda(t-s,x- W_N(s))] \;\;\;-\;\;\;\; \int_0^t \mathbf E[f_\varphi^\lambda (t-s,x-W(s))]ds,\end{align*} $$
where
$W_N(s) = N^{-1/2}R_N(Ns)$
for the discrete-time random walk
$(R_N(r))_{r\ge 0}$
with increment distribution described in Definition 5.12, and W is a standard Brownian motion. Define
$W_N(s)$
by linear interpolation for
$s\notin N^{-1}{\mathbb {Z}}_{\ge 0}$
.
By the KMT coupling [Reference Komlós, Major and TusnádyKMT76], we may assume that
$W_N$
and W are all coupled onto the same probability space so that
$\sup _{s\in [0,T]} |W_N(s) - W(s)| \leq G_N N^{-1/4}$
where
$G_N$
are random variable on that same probability space such that
$\sup _N\mathbf E|G_N|^p<\infty $
for all p (note that
$-1/4$
is the optimal exponent here because, despite KMT giving a better exponent in principle, the increments of the original random walk
$R_N$
are not centered but rather they have a nonzero mean of order
$N^{-3/4}$
). Consequently, by the definition of the parabolic spaces, we have uniformly over
$s,t \in [0,T]$
the bound
$$ \begin{align*} & |f_\varphi^\lambda(t-s,x- W_N(s))-f_\varphi^\lambda (t-s,x-W(s))| \\ & \leq \bigg( \sup_{u \in \big[-|x| - |W(s)|-|W_N(s)|, \;|x| + |W(s)|+|W_N(s)|\big]} | \partial_x f_\varphi^{\lambda}(t-s,u)|\bigg)\cdot | W_N(s) - W(s)| \\ &\leq \bigg( \|f\|_{C^{\alpha,\tau}_{\mathfrak{s}}(\Psi_T)} \lambda^{\alpha-1} \big(1+4x^2 + 4W(s)^2 + 4W_N(s)^2\big)^\tau \bigg)\cdot G_NN^{-1/4}. \end{align*} $$
Here the factor
$\|f\|_{C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)} \lambda ^{\alpha -1}$
can be deduced using, for example, Remark 5.5 which says that
$\partial _x$
boundedly reduces the parabolic regularity by 1 exponent. We also used the fact that
$(|x|+|y|+|z|)^2 \leq 4x^2+4y^2+4z^2.$
We claim that
for a constant
$C=C(\alpha ,T,\tau )$
independent of
$N,x,s.$
To prove this, one uses
$(a+b+c+d)^\tau \leq 4^\tau (a^\tau +b^\tau +c^\tau +d^\tau )$
, then one notes that
$\sup _{N} \mathbf E|G_N \cdot \sup _{s\leq T} (|W(s)|^{2\tau } +|W_N(s)|^{2\tau })|<\infty $
, for example, by Cauchy-Schwarz.
Consequently the above expression for
$(K_N-K)f_\varphi ^\lambda (t,x)$
can be bounded in absolute value by a universal constant times
$\|f\|_{C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)} \lambda ^{\alpha -1} N^{-1/4}$
.
Since we expect to obtain a Dirac initial data in the limiting SPDE, there is an additional singularity at the origin that we have not yet taken into account. To fix this issue, we now formulate a tightness result taking into account this singularity, by starting the field (1.4) from some positive time
$\varepsilon $
(which should be thought of as being close to
$0$
).
Proposition 5.21 (Regularity of limit points)
Fix
$\varepsilon>0,$
and consider the fields
$\mathfrak H^{N,\varepsilon }(t,\cdot ):=\mathfrak H^N(t+\varepsilon ,\cdot ),$
for
$t\ge 0$
. Set
$\mathfrak H^{N,\varepsilon }(t,\cdot ):=0$
for
$t<0$
. The fields
$\mathfrak H^{N,\varepsilon }$
may be realized as random variables taking values in
$C([0,T], C^{\alpha ,\tau }({\mathbb {R}}))$
for any
$\tau>1$
and
$\alpha <-5$
. Furthermore, they are tight with respect to that topology, and any limit point is necessarily supported on
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _T)$
for any
$\kappa>0$
.
In particular, since
$-\kappa +1/2>0$
, the last statement implies that the limit point must be supported on a space of continuous functions.
Proof. We already know from Item (2) of Proposition 5.18 that the
$\mathfrak H^{N}$
are tight in
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
. We also know from Item (1) of Proposition 5.18 that
$\widehat M_N$
are tight in
$C([0,T+1],C^{\alpha ,\tau }({\mathbb {R}}))$
, which is embeds continuously into
$C^{\alpha ,\tau }_{\mathfrak {s}}(\Psi _T)$
by Lemma 5.4. Thus using the first bound in Lemma 5.19, we have that
$D_N\widehat M_N$
are tight in
$C_{\mathfrak {s}}^{\alpha -2}(\Psi _T)$
.
Consider any joint limit point
$(H^\infty ,\mathscr M^\infty )$
as
$N\to \infty $
of the pair
$(\mathfrak H^N,D_N\widehat M_N)$
in the product space
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))\times C^{\alpha -2,\tau }_{\mathfrak {s}}(\Psi _T)$
. By Lemmas 5.17 and 5.20 we may conclude that
where
$p(t,x) = (2\pi t)^{-1/2}e^{-x^2/2t} \mathbf {1}_{\{t\ge 0\}}$
is the (deterministic) standard heat kernel, and K is the operator defined in (5.3).
Fix
$\varepsilon>0$
. Now we consider the
$\varepsilon $
-translates of these statements. It is automatic from Item (2) of Proposition 5.18 that
$\mathfrak H^{N,\varepsilon }$
are tight in
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
. Likewise it is automatic from Item (1) that the family
$\widehat M_{N,\varepsilon } := \widehat M_N(\varepsilon +\cdot )$
is tight in
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
so that from Lemma 5.19 we see that
$D_N\widehat M_{N,\varepsilon }$
are tight in
$C^{\alpha -2,\tau }_{\mathfrak {s}}(\Psi _T).$
Let us now take a joint limit point
$(H^\infty , \mathscr M^\infty , H^{\infty ,\varepsilon }, \mathscr M^{\infty ,\varepsilon })$
of the family
$(\mathfrak H^N, D_N\widehat M_N, \mathfrak H^{N,\varepsilon } , D_N\widehat M_{N,\varepsilon }).$
On one hand we necessarily have
$H^{\infty ,\varepsilon }(t) = H^{\infty }(t+\varepsilon )$
and on the other hand we necessarily have that
$H^\infty = p+K\mathscr M^\infty .$
From the last statement in Proposition 5.7, this then forces the relation
where the tensor product was defined in Definition 5.9. Now we notice
$K(\delta _0\otimes H^\infty (\varepsilon ,\cdot ))= J \big (H^\infty (\varepsilon ,\cdot )\big ) $
, where J is defined in Remark 5.8. We thus have the Duhamel equation
For technical reasons that will be made clear below, we now replace
$\varepsilon $
by
$\varepsilon /2$
and T by
$T+1$
.
We now claim that for
$q>1$
Let us now complete the proof of regularity assuming (5.12).
-
• Using (5.12) and Proposition 5.7, it follows that
This implies that the restriction of
$$ \begin{align*}{\mathbb{E}}\bigg[\big\|K\mathscr M^{\infty,\varepsilon/2}\big\|_{C_{\mathfrak{s}}^{-\kappa+1/2,\tau}(\Psi_{T+1})}^q\bigg]<\infty.\end{align*} $$
$K\mathscr M^{\infty ,\varepsilon /2}$
to
$[\varepsilon /2,T+\varepsilon /2]\times {\mathbb {R}}$
lies in
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _{[\varepsilon /2,T+\varepsilon /2]})$
.
-
• Let
$h:= \mathfrak H^N(\varepsilon /2,\cdot )$
. By Proposition 5.18 Item (2) we have
$h\in C^{-5-\kappa ,\tau }$
almost surely. So from (5.5) with
$\alpha :=-5-\kappa $
and
$\beta := \frac 12-\kappa ,$
it follows that
$\widehat P_{\varepsilon /2}h \in C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _{T+1})$
almost surely. Since from (5.4), we know
$Jh(t+\varepsilon /2,x) = \widehat P_{\varepsilon /2} h(t,x)$
, we thus have that the restriction of
$Jh$
to
$[\varepsilon /2,T+\varepsilon /2]\times {\mathbb {R}}$
lies in
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _{[\varepsilon /2,T+\varepsilon /2]})$
.
Thanks to the above two bullet points and the relation (5.11), we have showed that the restriction of
$H^{\infty ,\varepsilon /2}$
to
$[\varepsilon /2,T+\varepsilon /2]\times {\mathbb {R}}$
lies in
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _{[\varepsilon /2,T+\varepsilon /2]})$
. This is equivalent to the fact that
$H^{\infty ,\varepsilon }$
lies in
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }(\Psi _T)$
. This completes the proof modulo (5.12).
Proof of (5.12)
We shall show (5.12) with
$\varepsilon /2$
and
$T+1$
replaced by
$\varepsilon $
and T respectively. For
$\alpha <0$
, the derivative operator
$\overline \partial _s : C([0,T],C^{\alpha ,\tau }({\mathbb {R}})) \to C_{\mathfrak {s}}^{\alpha -2,\tau }(\Psi _T)$
which is defined by sending
$(v_t)_{t\in [0,T]}$
to the distribution
$$ \begin{align}(\partial_s v,\varphi)_{L^2(\Psi_T)}:=v_T(\varphi(T,\cdot)) -\int_0^T v_t(\partial_t\varphi(t,\cdot))dt -v_0(\varphi(0,\cdot)),\end{align} $$
whenever
$\varphi \in C_c^\infty (\Psi _T),$
is a bounded linear map. Indeed this operator is just the composition of the embedding map of Lemma 5.4 with the operator
$\partial _s$
that was proved to be bounded in Lemma 5.19.
Let
$\widehat M_{N,\varepsilon }:= \widehat M_N(\varepsilon +\cdot )-\widehat M_N(\varepsilon )$
and
$\widehat Q_{N,\varepsilon }:= \widehat Q^{\boldsymbol {\zeta }}_N(\varepsilon +\cdot )-\widehat Q^{\boldsymbol {\zeta }}_N(\varepsilon )$
. Let us consider any joint limit point
$(Q^{\infty ,\varepsilon }, M^{\infty ,\varepsilon },\mathscr M^{\infty ,\varepsilon })$
of
$(\widehat Q_{N,\varepsilon }, \widehat M_{N,\varepsilon },D_N\widehat M_{N,\varepsilon }$
) in the space
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))\times C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))\times C^{\alpha -2,\tau }(\Psi _T),$
where
$\gamma <-1$
. Then one may verify from the second bound in Lemma 5.19 that one necessarily has
$\mathscr M^{\infty ,\varepsilon } = \overline \partial _s M^{\infty ,\varepsilon }$
. Furthermore by Proposition 5.18 one has martingality of
$M_t^{\infty ,\varepsilon }(\phi )$
, with
$\langle M^{\infty ,\varepsilon }(\phi )\rangle = Q^{\infty ,\varepsilon }(\phi ^2)$
.
Let us consider any smooth function
$\varphi $
supported on the unit ball of
${\mathbb {R}}^2$
such that
$\|\varphi \|_{C^1}\leq 1.$
Recall
$S_{(t,x)}^{\lambda }$
from (5.2), and notice that
$\big (\partial _t S_{(t,x)}^\lambda \varphi \big )^2 \leq \lambda ^{-10} \mathbf {1}_{[t-\lambda ^2,t+\lambda ^2]\times [x-\lambda ,x+\lambda ]}$
. We will now estimate the
$q^{th}$
moments of
$\mathscr M^{\infty ,\varepsilon }(S_{(t,x)}^{\lambda }\varphi )=\overline \partial _s M^{\infty ,\varepsilon }(S_{(t,x)}^{\lambda }\varphi )$
. By making
$\varepsilon $
smaller and T larger we may simply ignore the boundary terms in (5.13). Note that for fixed
$s,t,x,\lambda $
the martingale
$u \mapsto M_u^{\infty ,\varepsilon } (\partial _tS_{(t,x)}^\lambda \varphi (s,\cdot ))$
is constant outside of the interval
$[t-\lambda ^2,t+\lambda ^2]$
. Thus we may apply the Minkowski’s inequality and then Burkholder-Davis-Gundy in (5.13) to obtain
$$ \begin{align*} \mathbb{E}[ |\mathscr M^{\infty,\varepsilon}(S_{(t,x)}^{\lambda}\varphi)|^q]^{1/q} & = \mathbb{E}[ |\overline \partial_s M^{\infty,\varepsilon}(S_{(t,x)}^{\lambda}\varphi)|^q]^{1/q}\\&\leq \int_{t-\lambda^2}^{t+\lambda^2} {\mathbb{E}}\big[\big| M_s^{\infty,\varepsilon} (\partial_tS_{(t,x)}^\lambda \varphi(s,\cdot)) \big|^q\big]^{1/q}ds \\ &\leq C_q \int_{t-\lambda^2}^{t+\lambda^2} {\mathbb{E}} \bigg[ \bigg( Q_{t+\lambda^2}^{\infty,\varepsilon} \big((\partial_tS_{(t,x)}^\lambda\varphi)^2 (s,\cdot)\big) - Q_{t-\lambda^2}^{\infty,\varepsilon} \big((\partial_tS_{(t,x)}^\lambda\varphi)^2 (s,\cdot)\big)\bigg)^{q/2}\bigg]^{1/q}\\ &\leq 2C_q\lambda^{-3} {\mathbb{E}} \bigg[ \bigg( Q_{t+\lambda^2}^{\infty,\varepsilon} \big(\mathbf{1}_{[x-\lambda,x+\lambda]} \big) - Q_{t-\lambda^2}^{\infty,\varepsilon} \big(\mathbf{1}_{[x-\lambda,x+\lambda]}\big)\bigg)^{q/2}\bigg]^{1/q} \end{align*} $$
where
$C=C(q)>0$
. For any
$\delta>0$
, by the second bound in Proposition 4.7 we find that the last expression may be bounded above by
Given any
$\kappa>0$
, take
$\delta =\delta (\kappa )$
close to
$0$
and
$q=q(\kappa )$
large enough, then we may then apply Lemma 5.102 to conclude (5.12).
Now we may finally identify the limit points as the solution of the stochastic heat equation (1.5) and thereby prove our weak convergence theorem: Theorem 1.5.
Lemma 5.22. Let
$\boldsymbol w$
be a random variable in
$\mathcal S'({\mathbb {R}})$
such that for some smooth even (deterministic)
$\phi \in \mathcal S({\mathbb {R}})$
and some
$\delta>0$
one has
where
$\phi _\varepsilon ^a (x):= \varepsilon ^{-1}\phi (\varepsilon ^{-1} (x-a)).$
Then
$(\boldsymbol w,\psi )=0$
almost surely for all
$\psi \in \mathcal S({\mathbb {R}})$
.
Proof. Define
$\boldsymbol w^\varepsilon (x):= \boldsymbol w * \phi _\varepsilon (x)$
so that
$(\boldsymbol w,\phi ^a_\varepsilon ) = \boldsymbol w^\varepsilon (a).$
Given some smooth
$\psi :{\mathbb {R}}\to {\mathbb {R}}$
of compact support note that
$(\boldsymbol w^\varepsilon , \psi ) \to (\boldsymbol w,\psi )$
a.s. as
$\varepsilon \to 0$
. This is a purely deterministic statement. Thus it suffices to show that
$(\boldsymbol w^\varepsilon ,\psi ) \to 0$
in probability. To prove that, suppose that the support of
$\psi $
is contained in
$[-S,S]$
and note by Cauchy-Schwarz that
$$ \begin{align*}|(\boldsymbol w^\varepsilon, \psi)| =\bigg|\int_{{\mathbb{R}}} \boldsymbol w^\varepsilon(a) \psi(a)da\bigg|\leq \bigg[ \int_{[-S,S]} \boldsymbol w^\varepsilon(a)^2 da\bigg]^{1/2} \|\psi\|_{L^2({\mathbb{R}})},\end{align*} $$
so that by taking expectation and applying Jensen we find
$$ \begin{align*}{\mathbb{E}}[|(\boldsymbol w^\varepsilon, \psi)|] \leq \|\psi\|_{L^2} \bigg[\int_{[-S,S]} {\mathbb{E}}[\boldsymbol w^\varepsilon(a)^2]da \bigg]^{1/2}.\end{align*} $$
By assumption,
${\mathbb {E}}[\boldsymbol w^\varepsilon (a)^2]\leq Ca^{\delta -1}$
and
${\mathbb {E}}[\boldsymbol w^\varepsilon (a)^2]\to 0$
as
$\varepsilon \to 0$
, therefore the dominated convergence theorem now gives the result by letting
$\varepsilon \to 0$
on the right side. This proves the claim for
$\psi $
of compact support. For general
$\psi \in \mathcal S({\mathbb {R}})$
we may find a sequence
$\psi _n \to \psi $
in the topology of
$\mathcal S({\mathbb {R}})$
, with each
$\psi _n$
compactly supported. Then
$0=(\boldsymbol w,\psi _n)\to (\boldsymbol w,\psi )$
a.s. as
$n\to \infty $
.
Theorem 5.23 (Solving the martingale problem)
Consider the triple of processes
$(\mathfrak H^N,\widehat M_N,Q_N^{\boldsymbol {\zeta }})_{t\ge 0}$
, where
$\mathfrak H^N$
,
$\widehat M_N$
,
$Q_N^f,$
and
$\boldsymbol {\zeta }$
are defined in (1.4), Definition 5.14, (4.10), and Definition 2.11 respectively. Fix
$\alpha <-5, \gamma <-1,$
and
$\tau>1.$
These triples are jointly tight in the space
Consider any joint limit point
$(H^\infty , M^\infty , Q^\infty )$
. Then for any
$s>0$
, the process
$(t,x)\mapsto H_{s+t}^\infty (x)$
is necessarily supported on the space
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }( \Psi _T)$
. Furthermore,
$(M_t^\infty (\phi ))_{t\ge 0}$
is a continuous martingale for all
$\phi \in C_c^\infty ({\mathbb {R}})$
, and moreover for all
$0< s\leq t< T$
one has the almost sure identities
$$ \begin{align} M_t^\infty(\phi)-M_s^\infty(\phi) &= \int_{\mathbb{R}} \big(H_t^\infty(x)-H_s^\infty(x)\big)\phi(x){\mathrm{d}} x -\frac12 \int_s^t \int_{\mathbb{R}} H_u^\infty(x)\phi"(x){\mathrm{d}} x {\mathrm{d}} u \end{align} $$
$$ \begin{align} Q_t^\infty(\phi)-Q_s^\infty(\phi) &= \gamma(\boldsymbol{\zeta}) \int_{\mathbb{R}}\int_s^t H_u^\infty(x)^2 \phi(x)\,{\mathrm{d}} u\,{\mathrm{d}} x. \end{align} $$
Before going into the proof, we remark that with some inspection it may be verified that all quantities make sense given the spaces they lie in, as long as we choose
$\kappa <1/2$
to ensure that
$H^\infty $
is a continuous function in space-time away from
$t=0$
.
Proof. Most parts of the following theorem are already established in previous propositions and lemmas. The tightness of the triple was shown in Proposition 5.18. Consider any limit point
$(H^\infty ,M^\infty ,Q^\infty ).$
The fact that for any
$s>0$
, the process
$(t,x)\mapsto H^\infty _{s+t}(x)$
is necessarily supported on the space
$C_{\mathfrak {s}}^{-\kappa +1/2,\tau }( \Psi _T)$
was proved in Proposition 5.21. From Proposition 5.18 we have that for all
$\phi \in C_c^\infty ({\mathbb {R}})$
the process
$(M_t^\infty (\phi ))_{t\ge 0}$
is a martingale. (5.15) is already proven in Proposition 5.18 as (5.8). Now we just need to show (5.14) and (5.16).
Proof of (5.14)
In (5.10) we obtained that
$H^\infty = p +K\mathscr M^\infty $
where (just as we observed after (5.13)) one necessarily has
$\mathscr M^\infty =\overline \partial _s M^\infty $
. Disregarding the boundary terms implies that
for all
$\varphi $
of compact support contained in
$[\varepsilon ,T-\varepsilon ]\times {\mathbb {R}}$
for some
$\varepsilon>0$
. Since both
$M^\infty $
and
$H^\infty $
lie in spaces with strong enough topologies (see Proposition 5.21), taking
$\varphi (u,x)$
to approach the function
$(u,x) \mapsto \mathbf {1}_{[s,t]}(u) \phi (x)$
in the above equation leads to (5.14).
Proof of (5.16)
Fix any
$0\le t\le T$
and let
$\boldsymbol w$
be a
$\mathcal {S}'(\mathbb {R})$
-valued random variable defined as
$$ \begin{align} (\boldsymbol w,\phi):= Q_t^{\infty}(\phi)-\gamma(\boldsymbol{\zeta})\int_{\mathbb{R}}\int_0^t H_s^{\infty}(x)^2\phi(x)\,{\mathrm{d}} s\,{\mathrm{d}} x. \end{align} $$
We claim that
$ (\boldsymbol w,\phi )=0$
a.s. for all
$\phi \in \mathcal {S}(\mathbb {R}).$
This will validate (5.16). To verify this, let us define
$\xi _{\varepsilon }^a(x):=\varepsilon ^{-1}\xi (\varepsilon ^{-1}(x-a))$
with
$\xi (x):=\frac 1{\sqrt \pi }e^{-x^2}$
. Using the “key estimate” (4.16) of Proposition 4.6, one verifies that the assumptions of Lemma 5.22 hold true for
$\boldsymbol w$
with this family of mollifiers.
To see why Lemma 5.22 is applicable, first note that for fixed
$\phi \in C_c^\infty ({\mathbb {R}})$
, the difference between the sum appearing in (4.16) and integral over
$[0,t]$
of the same quantity tends to zero in probability with respect to the topology of
$C[0,T]$
, since we proved tightness of
$\mathfrak H^N$
in a topology given by the norm of
$C([0,T],C^{\alpha ,\tau }({\mathbb {R}}))$
. Therefore that sum in (4.16) converges in law, jointly with
$\mathfrak H^N$
, to the integral appearing in (5.17). Then an application of (4.16) and (4.17) shows that
$(\boldsymbol w,\phi )$
as defined by (5.17) satisfies the conditions of Lemma 5.22 for any
$\delta \in (0,1)$
, since (4.17) gives a polylogarithmic bound which is less than
$C a^{-\delta }$
for arbitrary choice of
$\delta>0$
. This is enough to complete the proof.
We now complete the proof of our main theorem, Theorem 1.5.
Proof of Theorem 1.5
We continue with the notation and setup of Theorem 5.23. We have already established the tightness of
$\mathfrak H^N$
in Proposition 5.18. Consider any limit point
$ H^\infty $
of
$\mathfrak H^N$
. From the previous theorem, we already know that
$(t,x) \mapsto H_{t+\varepsilon }^\infty (x)$
is a continuous function in space and time. From the three equations (5.14), (5.15), and (5.16) in Theorem 5.23 it follows that the martingale problem for (1.5) is satisfied by any limit point
$H^\infty $
. We refer the reader to [Reference Bertini and GiacominBG97, Proposition 4.11] for the characterization of the law of (1.5) as the solution to this martingale problem.
The result there is only stated for continuous initial conditions, so what this really shows is that for any
$\varepsilon>0$
the law of the continuous field
$(t,x) \mapsto H^\infty _{t+\varepsilon }(x)$
is that of the solution of (1.5) with initial condition
$H^\infty _\varepsilon (\cdot ).$
Thus we still need to pin down the initial data as
$\delta _0$
, by showing that we can let
$\varepsilon \to 0$
and see that the limit of
$H^\infty _\varepsilon (\cdot )$
is equal to
$\delta _0$
in some sense.
In [Reference ParekhPar18, Section 6] there is a general approach to doing this. Specifically, it suffices to show, as in Lemma 6.6 of that paper, the two bounds
where
$q>1$
is arbitrary,
$p_t(x)=(2\pi t)^{-1/2}e^{-\frac {x^2}{2t}}$
is the standard heat kernel, and C is independent of
$t>0$
and
$x\in {\mathbb {R}}$
. Clearly, it suffices to show this when q is an even integer. The solution of (1.5) with
$\delta _0$
initial condition certainly satisfies this bound, and by the moment convergence result in Section 3, we know any limit point
$H^\infty $
must satisfy
$\mathbb {E}[(\int _{\mathbb {R}} H_t^\infty (x)\phi (x){\mathrm {d}} x)^k] = \mathbf {E}[(\int _{\mathbb {R}} \mathcal {U}_t(x)\phi (x){\mathrm {d}} x)^k]$
for all
$k\in \mathbb N$
and
$\phi \in C_c^\infty ({\mathbb {R}})$
, where
$(t,x) \mapsto \mathcal {U}_t(x)$
solves (1.5) with
$\delta _0$
initial data. From here we can conclude by letting
$\phi \to \delta _x$
that
$\mathbb {E}[H_t^\infty (x)^k] = \mathbf {E}[ \mathcal {U}_t(x)^k]$
for all
$k\in \mathbb N$
and all
$x\in {\mathbb {R}}$
. Then we may immediately deduce (5.18) and (5.19) by the corresponding bounds for
$\mathcal {U}_t$
. This completes the proof.
Remark 5.24 (Extension to other initial data)
Throughout this paper we have taken the convention of Dirac initial data, where one starts
$\mathfrak H^N$
from a Dirac mass at zero, corresponding to all particles starting from 0 in the quenched random walks. In fact, this may be generalized, where one assumes instead that
$\omega _N$
is a sequence of probability measures on I such that the measures
$e^{xN^{\frac {2p-1}{4p}}} \omega _N(N^{1/2} {\mathrm {d}} x)$
converge in the topology of
$C^{\alpha ,\tau }({\mathbb {R}})$
to some limiting bounded and continuous function
$\mathcal U_0.$
Then define
$P^\omega (n,\cdot ):= \omega _NK_1\cdots K_n$
and let
$\mathfrak H^N$
be as in (1.4) with
$P^\omega $
defined in this more general fashion. The above methods can be generalized to show that
$\mathfrak H^N$
as defined this way converge in law (in the same sense as Theorem 1.5) to the solution of (1.5) started from
$\mathcal U_0$
. Indeed, this is because all of the bounds and convergence theorems we derived in Sections 2 and 3 worked with varying initial starting point, and this generality can be propagated into the remainder of the paper. Some aspects of the paper need an appropriate modification, for instance Proposition 4.6 will no longer require the polylogarithmic factor at the origin.
Of course, this leads to the question of
$e^{xN^{\frac {2p-1}{4p}}} \omega _N(N^{1/2}{\mathrm {d}} x)$
converging to a more singular limit in the topology of
$C^{\alpha ,\tau }({\mathbb {R}})$
, as opposed to a continuous function (e.g., a measure or tempered distribution). For a reasonable limiting measure, we still expect to see (1.5) in the limit started from this singular initial data, but it is beyond the scope of the present paper to determine how to generalize Proposition 4.6 and [Reference Bertini and GiacominBG97, Proposition 4.11] to this setting.
6 Examples of models satisfying the hypotheses
In this section, we will introduce a number of different models, prove that Assumption 1.2 is satisfied for all these models, and, if possible, calculate the coefficient
$\gamma _{\mathrm {ext}}$
from Theorem 1.5 for these models. As explained in Remark 1.7 of the introduction, if all other conditions of Assumption 1.2 hold true but
$m_2-m_1^2 \neq 1$
, then the condition
$m_2-m_1^2=1$
can always be forced to be true by replacing I by
$c I$
for some
$c>0$
, and modifying the kernels
$K_i$
to reflect this rescaling. Therefore we will not verify this condition for the models below, and it should be implicitly understood that this would affect the diffusion coefficient in (1.5).
6.1 A generalized model of random walk in random environment on
${\mathbb {Z}}$
Here we consider a generalized non-nearest neighbor model of random walks in random environments whose KPZ fluctuations were recently conjectured in the physics paper [Reference Hass, Drillick, Corwin and CorwinHDCC24].
Define the set
$\Omega :=\{ v \in [0,1]^{\mathbb {Z}}: \sum _{i\in {\mathbb {Z}}} v(i) = 1\}$
, and let P be any probability measure on
$\Omega $
(defined on the
$\sigma $
algebra generated by the coordinate maps
$v \mapsto v(i)$
with
$i\in {\mathbb {Z}}$
). Assume the following four conditions on the probability measure P:
-
• The increment measure
$q(i) := E[v(i)]$
(
$i\in \Bbb Z$
) defines an aperiodic and irreducible random walk on
$\Bbb Z$
. -
•
$E[v(i)] \leq Ce^{-d|i|}$
for some constants
$C,d>0.$
-
•
$\sum _{i\in {\mathbb {Z}}} E[v(i)^2]<1$
, that is, v is not P-a.s. concentrated at a single point of
${\mathbb {Z}}$
. -
• The first
$p-1$
moments of v sampled from P are deterministic but the
$p^{th}$
one is not, in other words
$\sum _{i\in \Bbb Z} i^\ell v(i)$
is deterministic for integers
$0\le \ell < p$
, but is genuinely random for
$\ell =p$
.
It is clear that such P exists for every choice of
$p\in \mathbb N$
, for instance choose a probability distribution on
$\{-1,...,p-1\}$
so that the first
$p-1$
moments are deterministic but the
$p^{\text {th}}$
one may take, for example, either of two possible distinct values with equal probability. Using linear algebra one may show that this is possible.
With P as above, let us now sample
$\{v_{r,x}\}_{x\in {\mathbb {Z}}, r\in {\mathbb {Z}}_{\ge 0}} $
according to the measure
$P^{\otimes ({\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}})}$
on the product space
$\Omega ^{{\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}}}.$
In other words all of the
$v_{r,x}(\bullet )$
are IID
$\Omega $
-valued random variables sampled from P for every lattice site
$(r,x)\in {\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}}.$
Now sample the random kernels
Proposition 6.1.
$K_r$
as defined above satisfies all of the conditions of Assumption 1.2.
Proof. With the exception of Item (5), it is immediate from the above four bullet points that all of the items of Assumption 1.2 are satisfied. To show Item (5), one may actually show that the stronger condition (1.1) is satisfied. But this is fairly obvious since the total variation distance is zero as long as no two
$x_i$
are the same, simply by the assumption of independence (i.e., sampling the kernels from the product measure
$P^{\otimes ({\mathbb {Z}}_{\ge 0}\times {\mathbb {Z}})}$
).
Theorem 1.5 thus gives convergence of the rescaled and recentered quenched density field (1.4) to the solution of (1.5) for such a model of random walks in a random environment with such IID kernels at every lattice site. This generalizes our result from [Reference Das, Drillick and ParekhDDP24b, Theorem 1.1] and confirms a conjecture from the physics work of [Reference Hass, Drillick, Corwin and CorwinHDCC24].
Suppose the measure P has the following property: there exists some
$c\in [-1,1]$
such that for all
$i,j\in \Bbb Z$
with
$i\ne j$
, one has
$E[v(i)v(j)] = cE[v(i)]E[v(j)]$
. Then one may show that the invariant measure
$\pi ^{\mathrm {inv}}$
of the annealed difference process is just counting measure on
${\mathbb {Z}}$
except at the origin where it carries extra mass, specifically
$c^{-1}$
. We see that the annealed difference process
$\boldsymbol {p}_{\mathbf {dif}}$
is actually reversible in this case, that is, the detailed balance equations can be solved, and the amount of extra mass
$\pi ^{\mathrm {inv}}$
has at the origin can be found fairly explicitly in terms of the parameters of the model. This was first noted in the physics paper [Reference Hass, Drillick, Corwin and CorwinHDCC24] which led to several interesting alternate forms of the coefficient
$\gamma _{\mathrm {ext}}$
. In the present work we do not explore questions such as massaging the coefficient into other forms, determining when the 2-point motion is reversible, etc.
Remark 6.2. The aperiodicity condition in the first bullet point may seem overly restrictive, since it omits many nearest-neighbor models. The point is that periodic kernels may not satisfy the irreducibility condition in Assumption 1.2 (6), as the communicating class of the origin under the difference chain
$\boldsymbol {p}_{\mathbf {dif}}$
can be a proper subgroup
$I\subset {\mathbb {Z}}^d$
. For instance even in the simplest nearest-neighbor case the difference chain would be supported on
$2\Bbb Z$
. However, this periodicity for nearest-neighbor models is not a fatal flaw and can be remedied by considering the two-step chain in which the kernel
$K_n$
is replaced with the product
$ K_{2n-1} K_{2n}$
, and
$I=\Bbb Z$
is replaced with
$I=2\Bbb Z.$
In general, such periodicity issues can be fixed by modifying the kernels similarly, taking into account multiple steps of the random walk at once.
6.2 A “random landscape model” with strongly mixing weight fields
Here we will study a particular type of stochastic flow model introduced to us by I. Corwin. We will see that KPZ fluctuations arise in systems where there may be infinite range of particle interaction or non-IID weights satisfying strong enough mixing conditions. This seems to be somewhat novel among KPZ-type convergence results.
Let
$(\omega _{r,x})_{r\in \mathbb N, x\in {\mathbb {Z}}}$
be a collection of random variables, and for each
$r\in \mathbb N$
denote
$\boldsymbol \omega _r:= (\omega _{r,x})_{x\in {\mathbb {Z}}}.$
Let
$b: {\mathbb {Z}}\to [0,1]$
be any deterministic function, with
$b(0)>0$
and assume that the additive subgroup of
${\mathbb {Z}}$
generated by
$\{x: b(x)>0\}$
is all of
${\mathbb {Z}}.$
Define the kernels
$$ \begin{align}K_r(x,y):= \frac{b(y-x) e^{\omega_{r,y}}}{\sum_{y'\in {\mathbb{Z}}} b(y'-x) e^{\omega_{r,y'}}}. \end{align} $$
In order to prove KPZ fluctuations for this type of model, we will need to verify Assumption 1.2, for which we will need to assume that either one of the following two conditions holds:
-
(A) b is finitely supported, furthermore the
$\boldsymbol \omega _1, \boldsymbol \omega _2, \boldsymbol \omega _3,...$
are independent families with each
$\boldsymbol \omega _r$
having the same finite-dimensional marginals as
$\boldsymbol \omega _1,$
and moreover
$\boldsymbol \omega _1$
is strictly stationary and
$\alpha $
-mixing with
$\alpha (n)\leq C e^{-\theta n}$
for some
$C,\theta>0$
. -
(B)
$|b(x)| \leq Ce^{-\theta |x|}$
for some
$C,\theta>0$
, and furthermore
$(\omega _{r,x})_{r\in \mathbb N, x\in {\mathbb {Z}}}$
are IID with
${\mathbb {E}}[e^{16 |\omega _{1,0}|}]<\infty $
.
Except for Items (3) and (5), the remaining items of Assumption 1.2 are trivially satisfied from the assumptions on the weights and the probabilities (taking
$p=1$
in Item (4)). Note that for
${\mathbf {x}},{\mathbf {y}}\in I^k$
we have
$$ \begin{align*}\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}}) &= {\mathbb{E}}[ \prod_{j=1}^k K_1(x_j,y_j)] = {\mathbb{E}} \bigg[ \prod_{j=1}^k \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{y'\in {\mathbb{Z}}} b(y'-x_j) e^{\omega_{1,y'}}}\bigg]\\ \mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}}) &= \prod_{j=1}^k {\mathbb{E}}[ K_1(x_j,y_j)]=\prod_{j=1}^k{\mathbb{E}} \bigg[ \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{y'\in {\mathbb{Z}}} b(y'-x_j) e^{\omega_{1,y'}}}\bigg].\end{align*} $$
Using these expressions, let us focus on verifying Items (3) and (5) in the conditions of Assumption 1.2. This will be done separately for each of the two bulleted assumptions above.
Proposition 6.3. Under either Condition (A) or Condition (B), Assumption 1.2 is satisfied by the kernels
$K_r$
from (6.1), with
$p=1$
.
Proof. Verification of Assumption 1.2 under Condition (A). Since b is finitely supported, Item (3) is trivially satisfied, so we just need to verify Item (5). To do this, we will verify the stronger assumption (1.1).
Let us first recall the definition of exponential
$\alpha $
-mixing for a strictly stationary sequence of random variables. If
$\boldsymbol \omega := (\omega _x)_{x\in {\mathbb {Z}}}$
is such a sequence, then we have a definition
where
$\mathcal F_{m,n} := \sigma (\omega _x: m\le x\le n)$
for
$m\le n$
and
$\alpha (\mathcal F,\mathcal G):= \sup \{ |\mathbb P(A\cap B) -\mathbb P(A)\mathbb P(B)|: A\in \mathcal F,B\in \mathcal G\}$
for two sub
$\sigma $
-algebras
$\mathcal F,\mathcal G$
. The sequence
$\boldsymbol \omega $
is called
$\alpha $
-mixing if
$\alpha (n)\to 0$
as
$n\to \infty $
, and it is called exponentially
$\alpha $
-mixing if
$\alpha (n)\to 0$
exponentially fast (e.g. IID sequences are
$\alpha $
-mixing with
$\alpha (n) = \mathbf {1}_{\{n=0\}}.)$
Suppose
$\boldsymbol \omega $
is any
$\alpha $
-mixing sequence with
$\alpha (n)\leq Ce^{-\theta n}$
where
$C,\theta>0$
. Let
$[m_i,n_i] (1\le i \le k)$
be disjoint intervals of
${\mathbb {Z}}$
such that
$n_i+J \leq m_{i+1}$
where
$J>0$
(large). By iteratively applying the
$\alpha $
-mixing property and using the triangle inequality, one may show that if
$f_i$
are
$\mathcal F_{m_i,n_i}$
-measurable with
$|f_i|\leq 1$
, then
$$ \begin{align}\bigg|{\mathbb{E}}\big[ \prod_{j=1}^k f_i \big] - \prod_{j=1}^k {\mathbb{E}}[f_i] \bigg| \leq C k \cdot e^{-\theta J}.\end{align} $$
Let us apply this bound in our context. Suppose that our weight function b is supported on
$[-B,B]$
where
$B>0.$
Now if
${\mathbf {x}}= (x_1,...,x_k)\in I^k$
, letting
$2J:= \min _{1\le i<j \le k} |x_i-x_j|$
we see from (6.2) and the above expressions for
$\boldsymbol p^{(k)}({\mathbf {x}},{\mathbf {y}})$
and
$\mu ^{\otimes k}({\mathbf {y}}-{\mathbf {x}})$
that
where
$C'=Cke^{\theta B}.$
Since b is finitely supported on
$[-B,B]$
, it follows that the measures
$\boldsymbol p^{(k)}({\mathbf {x}},\bullet )$
and
$\mu ^{\otimes k}$
have support contained on at most
$(2B)^{k}$
sites. Thus the last expression is already enough to imply the total variation bound (1.1), thus completing the proof.
Verification of Assumption 1.2 under Conditition (B). Using Jensen’s inequality on the denominator, we have
Taking expectation, we see that
where the (possibly infinite) product is convergent since we know that the weights are IID and
$\sum _{y'} b(y'-x)<\infty $
. Thus Item (3) of Assumption 1.2 is satisfied. More generally, the same arguments together with Hölder’s inequality can be used to show that as long as
$e^{|\omega _{1,0}|}$
has at least m finite moments, we have for all
$k\le m/4$
that
$$ \begin{align} {\mathbb{E}}[\prod_{j=1}^k K_r(x_j,y_j)] \le C\prod_{j=1}^k b(x_j-y_j), \end{align} $$
Let us verify Item (5) by verifying the stronger assumption (1.1). Notice that if
$X,Y,Z\ge 0$
are three random variables defined on any probability space
$(\Omega , \mathcal F,P)$
then we have that
$$ \begin{align}\notag E\bigg| \frac{X}{Y+Z}-\frac XY\bigg| &\leq E[X^2]^{1/2} E\bigg[ \bigg(\frac{1}{Y+Z}-\frac 1Y\bigg)^2\bigg]^{1/2} = E[X^2]^{1/2} E \bigg[ \bigg( \frac{Z}{Y(Y+Z)}\bigg)^2\bigg]^{1/2} \\&\leq E[X^2]^{1/2} E \bigg[ \frac{Z^2}{(Y+Z)^2}\bigg]^{1/2} \leq E[X^2]^{1/2} E [ Z^4]^{1/4} E[(Y+Z)^{-4}]^{1/4}. \end{align} $$
Fix
$J>0$
(large) and
${\mathbf {x}},{\mathbf {y}}\in {\mathbb {Z}}^k$
where
$1\le k \le 4$
. We will use this inequality with
$$ \begin{align*}X:= \prod_{j=1}^k e^{\omega_{1,y_j}} ,\;\;\;\;\;\;\; Y = \prod_{j=1}^k \bigg(\sum_{ y':|y'-x_j| \leq J }b(y'-x_j) e^{\omega_{1,y'}}\bigg),\;\;\;\;\;\;\;\;\; Z:= \sum_{{\mathbf{y}}': \max |y_j'-x_j|>J} \prod_{j=1}^k b(y_j'-x_j) e^{\omega_{1,y_j'}},\end{align*} $$
where the last sum is over vectors
${\mathbf {y}}'=(y_1',...,y_k').$
By Minkowski’s inequality we have
$$ \begin{align*}{\mathbb{E}}[Z^4]^{1/4} \leq \sum_{{\mathbf{y}}': \max |y_j'-x_j|>J} \prod_{j=1}^k b(y_j'-x_j) {\mathbb{E}}[e^{4\omega_{1,y_j'}}]^{1/4} \leq Ce^{-\theta J},\end{align*} $$
where we are using the exponential decay of the kernel b and the fact that the weights are IID. Now by Jensen we also have
$$ \begin{align*}(Y+Z)^{-1} = \prod_{j=1}^k \big(\sum_{y'\in {\mathbb{Z}}} b(y'-x_j) e^{\omega_{1,y'}}\big)^{-1} \leq \prod_{j=1}^k e^{-\sum_{y'\in {\mathbb{Z}}} b(y'-x_j) \omega_{1,y'}},\end{align*} $$
therefore by applying Hölder’s inequality we have
$$ \begin{align*}{\mathbb{E}}[(Y+Z)^{-4}] \leq \prod_{j=1}^k {\mathbb{E}} [ e^{-4k \sum_{y'\in {\mathbb{Z}}} b(y'-x_j) \omega_{1,y'}}]^{1/k} = \prod_{j=1}^k\prod_{y'\in {\mathbb{Z}}} {\mathbb{E}} [ e^{-4k b(y'-x_j) \omega_{1,y'}}]^{1/k} ,\end{align*} $$
and the latter is finite and bounded by an absolute constant since
$b(y'-x)$
is summable over
$y'$
and the weights are IID satisfying
${\mathbb {E}}[e^{4k|\omega _{1,0}|}]<\infty $
. Summarizing the last three expressions and applying (6.4), we find uniformly over
${\mathbf {x}},{\mathbf {y}}\in I^k$
that
$$ \begin{align}\bigg|\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}})-{\mathbb{E}} \bigg[ \prod_{j=1}^k \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{ y':|y'-x_j| \leq J } b(y'-x_j) e^{\omega_{1,y'}}}\bigg]\bigg| \leq Ce^{-\theta J}. \end{align} $$
The same argument with
$k=1$
also yields that
$$ \begin{align*}\bigg|{\mathbb{E}} \bigg[ \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{y'\in {\mathbb{Z}}} b(y'-x_j) e^{\omega_{1,y'}}}\bigg] - {\mathbb{E}} \bigg[ \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{ y':|y'-x_j| \leq J } b(y'-x_j) e^{\omega_{1,y'}}}\bigg]\bigg| \leq Ce^{-\theta J},\;\;\;\;\;\;\; 1\le j \le k.\end{align*} $$
By taking products over
$j\le k$
this will yield
$$ \begin{align}\bigg|\mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}})-\prod_{j=1}^k{\mathbb{E}} \bigg[ \frac{b(y_j-x_j) e^{\omega_{1,y_j}}}{\sum_{ y':|y'-x_j| \leq J } b(y'-x_j) e^{\omega_{1,y'}}}\bigg]\bigg| \leq Ce^{-\theta J}, \end{align} $$
where the constant C may have grown. Using the independence of the weights we see that if
${\mathbf {x}}=(x_1,...,x_k)\in I^k$
such that
$\min _{1\le i<j\le k}|x_i-x_j|=:2J$
then the expectation of the product in (6.5) is equal to the product of expectations in (6.6), thus
$\sup _{{\mathbf {y}}\in I^k}\big |\boldsymbol p^{(k)}({\mathbf {x}},{\mathbf {y}}) - \mu ^{\otimes k}({\mathbf {y}}-{\mathbf {x}})\big | \leq Ce^{-\theta J}.$
In other words, we have shown that
Using (6.3) we also have
$$ \begin{align*}\big|\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}}) - \mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}})\big| \leq \big|\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}})|+|\mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}})\big| \leq C\prod_{j=1}^k b(x_j-y_j).\end{align*} $$
Combining the last two expressions and using
$\min \{u,v\}\leq u^{1/2}v^{1/2}$
we thus find that
$$ \begin{align*}\big|\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}}) - \mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}})\big| \leq C e^{-\frac\theta4 \min_{1\le i<j\le k}|x_i-x_j|} \prod_{j=1}^k b(x_j-y_j)^{1/2} \leq C e^{-\frac\theta4 \min_{1\le i<j\le k}|x_i-x_j|} e^{-\frac\theta2\sum_{j=1}^k |x_j-y_j|},\end{align*} $$
where we applied the assumption of exponential decay of b. Consequently we find that
$$ \begin{align*} \big\| \mu^{\otimes k}(\bullet-{\mathbf{x}}) \;\;-\;\; \boldsymbol p^{(k)}\big( {\mathbf{x}},\bullet \big)\big\|_{TV} &= \sum_{{\mathbf{y}} \in I^k} \big|\boldsymbol p^{(k)}({\mathbf{x}},{\mathbf{y}}) - \mu^{\otimes k}({\mathbf{y}}-{\mathbf{x}})\big| \\ &\leq C e^{-\frac\theta4 \min_{1\le i<j\le k}|x_i-x_j|} \sum_{{\mathbf{y}} \in I^k} e^{-\frac\theta2 \sum_{j=1}^k|x_j-y_j|} \\ &\leq C e^{-\frac\theta4 \min_{1\le i<j\le k}|x_i-x_j|}, \end{align*} $$
which proves (1.1) as desired. The conditions needed to make sure all expectations above are finite is that
${\mathbb {E}}[ e^{4k |\omega _{1,0}|}]<\infty $
. Since
$p=1$
in Assumption 1.2 Item (4), we need (1.1) to hold only up to
$k=4$
, and it follows that
$16$
moments is all that is necessary.
Remark 6.4. Note that Condition (A) does not require any moment conditions on the weights
$\omega _{r,x}$
. We also remark that the mixing rate could be weakened to
$\alpha (n)\leq Cn^{-2-\epsilon }$
, and thanks to the assumptions on
$\mathrm {F_{decay}}$
in Item (5) of Assumption 1.2, this rate would still suffice, despite the exponential bound (1.1) not holding.
Additionally, while we formulated the model on discrete space
${\mathbb {Z}}$
, there is also a continuous-space analogue of the above
$\alpha $
-mixing landscape model. Let
$(\omega (x))_{x\in {\mathbb {R}}}$
be any strictly stationary
$\alpha $
-mixing field with continuous sample paths. Let
$b:{\mathbb {R}}\to {\mathbb {R}}$
be a bounded nonnegative measurable function of compact support that is not Lebesgue-a.e. zero. Then we may define the kernel
$$ \begin{align*}K_1(x,dy):= \frac{b(y-x) e^{\omega(y)}}{\int_{\mathbb{R}} b(y'-x) e^{\omega(y')}dy'}.\end{align*} $$
Essentially the same arguments as those given above will verify the conditions of Assumption 1.2 for IID samplings
$K_1,K_2,K_3,... $
of this kernel. An example of such a field
$(\omega (x))_{x\in {\mathbb {R}}}$
is an Ornstein-Uhlenbeck process. In fact, it is known that any strictly stationary Markov process that is geometrically ergodic is exponentially
$\beta $
-mixing and thus exponentially
$\alpha $
-mixing, see [Reference BradleyBra05, Theorem 3.7]. Consequently we can take
$\omega $
to be the stationary solution to the SDE
$d\omega (t) = -V'(\omega (t))dt +dW(t)$
where, for example,
$\lim _{|x|\to \infty } \frac {V(x)}{|x|} = \infty .$
Finally we remark that for the “landscape-type” models considered above, we do not know if the invariant measure
$\pi ^{\mathrm {inv}}$
can be explicitly described in terms of the parameters of the model, as was the case for the Dirichlet-type models of the previous subsection. We are not sure if the annealed difference process
$\boldsymbol {p}_{\mathbf {dif}}$
is reversible with respect to
$\pi ^{\mathrm {inv}}$
, though we see no reason for this to be the case given the complex correlation structure between the particles.
6.3 Transport-type SPDE model: diffusion in random environment
While Theorem 1.5 perhaps gives the impression that one can only consider discrete-time systems, this is not at all the case. We now discuss a continuous-time stochastic PDE model that motivated much of the discussion of KPZ fluctuations arising in stochastic flow models [Reference Barraquand and Le DoussalBLD20, Reference Le Doussal and ThieryLDT17, Reference Brockington and WarrenBW22]. This model was originally inspired by physics works of Kraichnan [Reference KraichnanKra68, Reference KraichnanKra94], see also [Reference Gawedzki and KupiainenGK95, Reference Gawedzki and KupiainenGK06, Reference Gawedzki and HorvaiGH04].
Let
$\xi $
be a standard Gaussian space-time white noise on
${\mathbb {R}}_+\times {\mathbb {R}},$
and consider any smooth even compactly supported function
$\phi :{\mathbb {R}}\to {\mathbb {R}}.$
Consider the spatially mollified noise
$V:= \xi * \phi $
which is smooth in space and white in time. Assume that
$\|\phi \|_{L^2(\Bbb R)}<1$
, that is,
$\phi * \phi (0)<1$
which implies that
$\sup _{x\in \Bbb R} \phi *\phi (x)<1$
.
Formally the covariance kernel of V is given by
${\mathbb {E}}[ V(t,x)V(s,y) ] = (\phi *\phi )(x-y) \cdot \delta (t-s)$
. As explained in [Reference Brockington and WarrenBW22], one may sensibly construct the Itô solution to the stochastic PDE
where
$t\ge 0$
and
$x\in {\mathbb {R}}$
. Note that, unlike (SHE), there is a derivative around the multiplicative term. A pathwise solution u that is continuous in both variables (away from the origin) and adapted to the filtration of the noise V is only possible due to the spatial smoothness of V, and it turns out that the solution preserves mass of the initial data, that is,
$\int _{\mathbb {R}} u(t,x) {\mathrm {d}} x =1$
for all
$t>0$
a.s., see [Reference KunitaKun94, Reference KunitaKun97, Reference Le Jan and RaimondLJR02].
For
$t>s\ge 0$
and
$x,y\in {\mathbb {R}}$
we define the “propagators”
$u_{s,t}(x,y)$
associated to the equation (6.7) as follows: let
$(t,y)\mapsto u_{s,t}(x,y)$
be the solution of (6.7) started from
$\delta _x(y)$
at time s. All of these solutions are coupled to the same realization of the driving noise V. Then we have the semigroup property
and furthermore if
$\{(s_i,t_i]\}_{i=1}^m$
are disjoint intervals then
$\{u_{s_i,t_i}\}_{i=1}^m$
are independent (because
$u_{s,t}$
is always measurable with respect to the restriction of V to
$[s,t]\times {\mathbb {R}}$
) [Reference KunitaKun94, Reference KunitaKun97, Reference Le Jan and RaimondLJR02].
We may thus define the kernels
and the preservation of mass implies that
$\int _{\mathbb {R}} K_r(x,y)dy=1$
for all
$x\in {\mathbb {R}}$
and
$r\in {\mathbb {Z}}_{\ge 0}$
.
Proposition 6.5. With
$K_r$
as defined above, all of the items of Assumption 1.2 are satisfied for these kernels, with
$p=1$
.
Proof. Since the driving noise V of (6.7) is spatially and temporally stationary, it follows that
$K_1$
is translation invariant and that the kernels
$K_i$
all have the same distribution. The fact that the kernels
$K_1,K_2,...$
are independent was already explained above. Let us explicitly calculate the annealed one-step law
$\mu $
as in Item (3) of Assumption 1.2. Letting
$\theta := 1- (\phi *\phi )(0)$
, [Reference Dunlap and GuDG22, Proposition 2.1] proves that (6.7) is the Kolmogorov forward equation associated to the SDE formally given by
where W is a Brownian motion independent of the noise V.Footnote
2
It turns out that this formal SDE admits a solution theory by construction of a stochastic integral
$\int _0^t V (s,X(s))ds,$
see, e.g., [Reference Bertini and CancriniBC95]. Moreover, for each realization of V, one may prove that
$\int _B u_{s,t}(x,y)dy = P^{V}(X(t)\in B|X(s)=x)$
where
$P^{V}$
is a quenched expectation given V [Reference KunitaKun94, Reference KunitaKun97, Reference Le Jan and RaimondLJR02]. In the annealed sense the stochastic integral
$\int _0^t V (s,X(s))ds,$
is just a Brownian motion of rate
$(\phi *\phi )(0)= \|\phi \|_{L^2({\mathbb {R}})}^2,$
consequently the annealed path X is itself just a Brownian motion of rate
$\theta +(\phi *\phi )(0)=1.$
This discussion shows that the annealed one-step law
$\mu $
is a normal distribution of mean
$0$
and variance
$1,$
which is enough to verify Item (3) of Assumption 1.2. Let us now move onto the proof of Item (6). If
$\gamma _1,\gamma _2: [0,t]\to {\mathbb {R}}$
are two deterministic paths then we have
$$ \begin{align*}{\mathbb{E}} \bigg[\bigg(\int_0^t V(s,\gamma_1(s))ds - \int_0^t V(s,\gamma_2(s))ds\bigg)^2\bigg] & = \int_0^t \psi (\gamma_1(s) - \gamma_2(s))ds,\;\;\; \text{ where }\\ \psi(x) &:= 2\big( \phi*\phi(0) - \phi*\phi(x)\big).\end{align*} $$
Using this fact and the SDE interpretation (6.9) of (6.7), consider two independent particles
$X^1,X^2$
satisfying that SDE, with the same realization of V but independent driving Brownian motions
$W^1,W^2$
. One may verify that the annealed difference
$Y(t):= X^1(t)-X^2(t)$
of two particles is itself an Itô diffusion (with deterministic coefficients) satisfying
where B is a standard Brownian motion on
$\Bbb R$
. In other words,
$\boldsymbol {p}_{\mathbf {dif}}(x,\bullet )$
is just the transition density at time 1 of the Itô diffusion (6.10) started from
$Y(0)=x.$
This is an elliptic diffusion with generator
$Lf(a) = \frac 12(2-2(\phi *\phi )(a))f"(a).$
Since the diffusion matrix is bounded above and below by positive constants, both bullets in Item (6) of Assumption 1.2 will be satisfied by standard existence theory of smooth positive densities for uniformly elliptic diffusions [Reference StroockStr88]. Note in this case that we have reversibility of
$\boldsymbol {p}_{\mathbf {dif}}$
with respect to
$\pi ^{\mathrm {inv}}$
, that is, the generator L is self-adjoint on
$L^2(\pi ^{\mathrm {inv}})$
as is easily shown using integration by parts.
It remains to verify Item (5) of Assumption 1.2. To do this, we will again verify the stronger assumption (1.1). As noted in [Reference Brockington and WarrenBW22, Eq. (1.10)] or [Reference WarrenWar15, Eq. (4)], the annealed law of k independent particles
$X^i(t) (1\le i \le k)$
each satisfying (6.9) (with independent driving noises
$W^i$
but the same realization of V) is an Itô diffusion on
${\mathbb {R}}^k$
with generator
$$ \begin{align*}L^{(k)}f({\mathbf{a}} ) = \frac12 \sum_{i,j=1}^k \bigg( \theta \mathbf{1}_{\{i=j\}} + (\phi*\phi)(a_i-a_j) \bigg) \frac{\partial^2}{\partial a_i\partial a_j}f({\mathbf{a}}) ,\end{align*} $$
where we recall that
$\theta := 1- (\phi *\phi )(0).$
Let
$\Sigma ({\mathbf {a}})$
be a positive square root of that diffusion matrix, so that the eigenvalues of
$\Sigma ({\mathbf {a}})$
are bounded above and below independently of
${\mathbf {a}} \in {\mathbb {R}}^k$
. If
$\phi *\phi $
is supported on
$[-B,B]$
then note that
$\Sigma ({\mathbf {a}}) = \sigma \cdot \mathrm {Id}$
outside of the set
$G:=\{ {\mathbf {a}}: \min _{1\le i<j\le k} |a_i-a_j|>B\}.$
Then the diffusion with generator L has law given by the solution of
$d\vec X(t) = \Sigma (\vec X(t)) d\vec W(t)$
for a standard k-dimensional Brownian motion
$\vec W$
. Thus
$\boldsymbol p^{(k)}({\mathbf {x}},\bullet )$
is just the transition density at time 1 of this elliptic diffusion started from
$\vec X(0)={\mathbf {x}}$
. Since
$\Sigma ({\mathbf {a}}) = \sigma \cdot \mathrm {Id}$
outside of the set G, the diffusion
$\vec X$
simply evolves as standard k-dimensional Brownian motion outside G. Consequently the total variation bound (1.1) can be proved by a simple and explicit coupling construction. A simple union bound using the Gaussian tail decay of the Brownian maximum on
$[0,1]$
shows that, starting from
${\mathbf {x}}=(x_1,...,x_k)$
, the probability of the Brownian sample path
$({\mathbf {x}}+\sigma \vec W(t))_{t\in [0,1]}$
hitting G is bounded above by
$C\sum _{1\le i<j\le k} e^{-\frac 1{2\sigma ^2}(x_i-x_j)^2},$
where C is a large constant depending on the support value B. If
$\vec X(0) = {\mathbf {x}}$
, then the diffusion sample path
$(\vec X(t))_{t\in [0,1]}$
agrees with
$({\mathbf {x}}+\sigma \vec W(t))_{t\in [0,1]}$
on the event that
$({\mathbf {x}}+\sigma \vec W(t))_{t\in [0,1]}$
does not hit G, which provides the explicit coupling of
$\vec X(1)$
with the k-dimensional normal distribution
${\mathbf {x}}+\sigma \vec W(1)$
. This shows that the total variation distance in (1.1) is bounded above by
$\sum _{1\le i<j\le k} e^{-\frac 1{2\sigma ^2}(x_i-x_j)^2}$
. Using the brutal bound
$\sum _{1\le i<j\le k} e^{-\frac 1{2\sigma ^2}(x_i-x_j)^2} \leq \frac 12k(k-1) e^{-\frac 1{2\sigma ^2} \min _{1\le i<j\le k} (x_i-x_j)^2}$
and noting that
$(x_i-x_j)^2>-1+|x_i-x_j|$
then implies the desired bound (1.1), thus completing the verification of Assumption 1.2.Footnote
3
Thus Assumption 1.2 has been verified (with
$p=1$
in Item (4)), which means that Theorem 1.5 will hold for this family of stochastic kernels
$(K_r)_{r\ge 0}$
. Using (6.10), the invariant measure is given explicitly as
Thanks to this explicit representation of
$\pi ^{\mathrm {inv}}$
, one may ask if the coefficient
$\gamma _{\mathrm {ext}}$
in Theorem 1.5 could also be calculated explicitly. The authors of [Reference Brockington and WarrenBW22] observe that this is indeed possible, and in fact one has
Notice by the convolution property (6.8) that
$K_1\cdots K_r (0,\cdot )$
simply agrees with
$u(r,\cdot )$
where u solves (6.7). Consequently in this context, the statement of Theorem 1.5 is saying that the field
is converging to the solution of (1.5). Of course this leads to a number of other questions, such as what happens if we do not integrate against a test function
$\phi $
but rather just consider the “un-coarsened” or “pointwise” field
$\mathfrak L^N(t,x):=N^{1/2}D_{N,t,x} u(Nt,N^{3/4}t+N^{1/2}x).$
This question remains open, and we expect a nontrivial answer, specifically we expect that there are additional nontrivial pointwise fluctuations in the limit that vanish under the test function integrated formulation as in
$\mathfrak H^N$
. It is not difficult to show impossibility of convergence in a topology of continuous functions, and we do this for the model considered in [Reference Das, Drillick and ParekhDDP24b]. But specifically gauging the exact nature of these additional pointwise fluctuations is more difficult and would require new ideas. This type of “local limit” question was explored in [Reference Corwin and GuCG17] in a different regime; perhaps a similar idea could be used to gauge these pointwise fluctuations.
A related question is that if we define the rescaled noise field
$V^N(t,x):= N^{3/4} V(Nt, N^{3/4}t+N^{1/2}x) $
, and we take a joint limit point of
$(V^N,\mathfrak H^N)$
, say
$(\xi , \mathcal U)$
, then is it true that the space-time white noise
$\xi $
is simply the driving noise of
$\mathcal U$
? This question was answered recently in [Reference Das, Drillick and ParekhDDP24b, Theorem 1.3] for a related model, and the answer is that
$\xi $
will not be the driving noise of
$\mathcal U$
. Rather the driving noise of
$\mathcal U$
will consist of
$\xi $
plus another independent space-time white noise. This is because, as pointed out in [Reference Brockington and WarrenBW22], the rescaled field
$z_N(t,x):= N^{1/2} D_{N,t,x} u(Nt,N^{3/4}t+N^{1/2}x)$
satisfies the Itô SPDE given for
$(t,x) \in \Bbb R_+\times \Bbb R$
by
Interestingly, the latter family of SPDEs is no longer “scaling subcritical” in the sense of regularity structures [Reference HairerHai14], but it is instead critical. Nonetheless, Theorem 1.5 clearly indicates that there is an explicit SPDE limit for
$z_N$
, and the final term causes the creation of additional multiplicative noise in the SPDE. A recent result of Hairer [Reference HairerHai25] on variance blowup of the tree diagrams may be relevant in describing more explicitly where the extra noise comes from.
This phenomenon of “creating extra noise in the limit” is not explored in the generalized context of the present work, as it is unclear to us how to define the prelimiting analogue of the driving noise
$V^N$
for the generalized model considered in Assumption 1.2.
6.4 Sticky Brownian motion
Here we discuss a continuum model of random walks in a random environment that has nice exact solvability properties that were explored in [Reference Brockington and WarrenBW23, Reference Barraquand and RychnovskyBR20, Reference Barraquand and Le DoussalBLD20]. We already prove KPZ convergence for this model in [Reference Das, Drillick and ParekhDDP24b], but we now discuss how to recover that result in the present context.
In order to define the random kernels for this model, we first need to introduce the concept of continuum stochastic flows of kernels. For
$s\le t$
, a random probability kernel, denoted
$L_{s,t}(x, A)$
, is a measurable function defined on some underlying probability space
$\Omega $
, such that it defines a probability measure on
$\mathbb {R}$
for each
$(x, \omega ) \in \mathbb {R} \times \Omega $
. The quantity
$L_{s,t}(x, A)$
is interpreted as the (random) probability of a particle to to arrive in A at time t starting at x at time s.
A family of random probability kernels
$(L_{s,t})_{s\le t}$
on
$\mathbb {R}$
is called a (continuum) stochastic flow of kernels if:
-
1. For any
$s\le t\le u$
and
$x\in \mathbb {R}$
, almost surely one has that
$K_{s,s}(x,A)=\delta _x(A)$
and for all A in the Borel
$$ \begin{align*} \int_{\mathbb{R}} L_{t,u}(y,A)L_{s,t}(x,dy)=L_{s,u}(x,A) \end{align*} $$
$\sigma $
-algebra of
$\mathbb {R}$
.
-
2. For any
$t_1\le t_2 \le \cdots \le t_k$
, the
$(L_{t_i,t_{i+1}})_{i=1}^{k-1}$
are independent. -
3. For any
$s\le u$
and
$t\in \mathbb {R}$
,
$L_{s,u}$
and
$L_{s+t,u+t}$
have the same finite dimensional distributions.
For the continuum stochastic flow that we consider in this section, one can ensure that the set of probability
$1$
on which 1 holds is independent of
$x\in \mathbb {R}$
and
$s\le t\le u$
, see [Reference Schertzer, Sun and SwartSSS14] and the more general result of [Reference Raimond and RiabovRR24]. This allows us to interpret
$(L_{s,t})_{s\le t}$
as bona fide transition kernels of a random motion in a continuum random environment (using, e.g., Kolmogorov extension theorem). The annealed law of such a motion is called the
$1$
-point motion associated to
$(L_{s,t})_{s\le t}$
. More generally, the k-point motion of a continuum stochastic flow of kernels is defined as the
$\mathbb {R}^k$
valued stochastic process
$\mathbf {X}=(X^1,\ldots ,X^k)$
with transition probabilities given by
$$ \begin{align*} P_t(\vec{x},d\vec{y})=\mathbb{E}\left[\prod_{i=1}^k L_{0,t}(x_i,dy_i)\right]. \end{align*} $$
We will be interested in a particular random motion in a continuum random environment originating from the Howitt-Warren flow of kernels. Its corresponding k-point motion solves a well-posed martingale problem that was first studied by Howitt and Warren in [Reference Howitt and WarrenHW09]. Below, we introduce the k-point motion by stating the martingale problem as formulated in [Reference Schertzer, Sun and SwartSSS14].
Let
$\nu $
be a finite and nonnegative measure on
$[0,1]$
. We say an
$\mathbb {R}^k$
-valued process
$\mathbf {X}_t=(X_t^1,\ldots ,X_t^k)$
solves the Howitt-Warren martingale problem with characteristic measure
$\nu $
if
$\mathbf {X}$
is a continuous, square-integrable martingale with the covariance process between
$X^i$
and
$X^j$
given by
$$ \begin{align*} \langle X^i,X^j\rangle_t=\int_0^t \mathbf{1}_{\{X_s^i=X_s^j\}}ds, \end{align*} $$
and furthermore it satisfies the following condition: consider any nonempty
$\Delta \subset \{1,2,\ldots ,k\}$
. For
$\mathbf {x}\in \mathbb {R}^k$
, let
Then the process
$f_{\Delta }(\mathbf {X}_t)-\int _0^t \beta _+\big ( g_{\Delta }(\mathbf {X}_s)\big )ds$
is a martingale with respect to the filtration generated by
$\mathbf {X}$
, where
$\beta _+(1):=0$
and
$$ \begin{align*} \beta_+(m):={\frac12}\int \sum_{k=0}^{m-2} (1-y)^k \nu(dy), \quad m\ge 2. \end{align*} $$
Howitt and Warren [Reference Howitt and WarrenHW09] proved existence and uniqueness in law for this martingale problem. From the covariation formula above, we see that each
$X^i$
is marginally a Brownian motion. The k particles
$X^i$
can thus be interpreted as Brownian motions evolving independently of each other when apart, but when they meet there is some “stickiness” or tendency to move as a single particle, and the intersection set will have positive measure (albeit being nowhere dense). Consequently the process
$\mathbf X$
is referred to as k-point sticky Brownian motion with characteristic measure
$\nu $
in the literature.
By a remarkable result of Le Jan and Raimond [Reference Le Jan and RaimondLJR04a, Theorem 2.1], any consistent family of Feller processes can be viewed as a k-point motion of some stochastic flow of kernels, unique in finite-dimensional distributions. Thus, in particular, the solution of the Howitt-Warren martingale problem can be viewed as the k-point motion of some stochastic flow of kernels. We call this stochastic flow of kernels the Howitt-Warren flow. In other words, these kernels are not explicitly specified in terms of some simpler object, but rather we are using the abstract result of [Reference Le Jan and RaimondLJR04a] to say that there is a unique family of stochastic kernels whose k-point motion is precisely k-point sticky Brownian motion for every
$k\in \mathbb N.$
We refer to [Reference Schertzer, Sun and SwartSSS09, Reference Schertzer, Sun and SwartSSS14, Reference Schertzer, Sun and SwartSSS17] for a more concrete description in terms of the Brownian web and net.
Proposition 6.6. Let
$L_{s,t}(x,{\mathrm {d}} y)$
denote the Howitt-Warren flow, with any choice of finite and nonzero characteristic measure
$\nu $
, and let us define our kernels
Then Assumption 1.2 holds true for this family of kernels.
Proof. We note that the independence and stationarity in space-time is clear from the definitions. The fact that
$\mu $
has exponential moments is clear from the fact that the
$1$
-point motion is just Brownian motion, hence
$\mu $
is Normal
$(0,1)$
. The verification of Item (5) can be done by noting that the k-point motion behaves as Brownian motion away from the boundary of the Weyl chamber
$\bigcup _{1\le i<j\le k} \{a_i=a_j\}\subset {\mathbb {R}}^k$
, consequently the total variation distance in (1.1) can be controlled using a similar explicit coupling construction that was used for the transport SPDE models of the previous subsection.
It remains to verify Item (6) of Assumption 1.2. To do this, we note that the annealed difference process behaves as Brownian motion sticky at zero. This is a process with pregenerator
$Lf(a) =\frac 12 f"(a)$
but with domain given by only those functions, smooth away from
$0$
, satisfying the relation
$f'(0+)-f'(0-) = \frac 1{2\nu [0,1]} f"(0\pm )$
, see [Reference Engelbert and PeskirEP14]. Alternatively it may be described as the unique-in-law process
$(Y(t))_{t\ge 0}$
such that all three of the processes
$$ \begin{align*}Y(t),\;\;\;\; Y(t)^2 - \int_0^t\mathbf{1}_{\{Y_s\ne 0\}}ds,\;\;\;\; |Y(t)| - \frac1{2\nu[0,1]} \int_0^t \mathbf{1}_{\{Y_s=0\}}ds\end{align*} $$
are continuous martingales. Since the characteristic measure
$\nu $
is assumed to be nonzero, the result of [Reference Brockington and WarrenBW23, Theorem 1.2] implies that for every starting point
$x\in I$
, this Markov process has an absolutely continuous (and everywhere positive) part to its time
$t=1$
transition kernel
$\boldsymbol {p}_{\mathbf {dif}}(x,\bullet )$
, thus the irreducibility of Item (6) in Assumption 1.2 is satisfied. The total variation continuity can also be deduced from the same result in [Reference Brockington and WarrenBW23], noting that both the singular part at the origin and the absolutely continuous part appearing in the formula are continuous as a function of the starting value, with the absolutely continuous part decaying like a Gaussian.
Thus Assumption 1.2 has been verified (with
$p=1$
in Item (4)), which means that Theorem 1.5 will hold for this family of stochastic kernels
$(K_r)_{r\ge 0}$
. We remark that the invariant measure
$\pi ^{\mathrm {inv}}$
is explicit in this case, given by Lebesgue measure plus a Dirac mass of weight
$\frac 1{2\nu [0,1]}$
at the origin (and
$\boldsymbol {p}_{\mathbf {dif}}$
is reversible with respect to
$\pi ^{\mathrm {inv}}$
). Thus, as we proved in [Reference Das, Drillick and ParekhDDP24a], the coefficient
$\gamma _{\mathrm {ext}}$
is explicitly computable in this case and satisfies
$\gamma _{\mathrm {ext}}^2 = \frac 1{2\nu [0,1]}.$
Acknowledgments
We thank the two anonymous referees for an extremely detailed reading of the paper. We thank Ivan Corwin for suggesting the problem of proving very general invariance principles for these random walk models in random environments, and for interesting discussions about the landscape model in Section 6.2. We thank Hindy Drillick and Sayan Das for many interesting conversations over the years related to the ideas presented herein. We thank Jacob Hass for suggesting that nontrivial behavior could be found with higher-order symmetry parameters
$p>1$
. We thank Yu Gu and Bálint Virág for interesting discussions surrounding the question of chaos and pointwise fluctuations, left for future work. We thank Jeremy Quastel for discussions about the hierarchy of nonequivalent instances of the KPZ equation arising in [Reference Hairer and QuastelHQ18, Theorem 1.1].
Competing interests
The author has no competing interests to declare.
Funding statement
This research was supported by grant DMS-2401884 from the NSF.
Data availability statement
Due to the theoretical nature of the work, there was no data produced or analyzed as part of this work.
Ethical standards
The research meets all ethical guidelines, including adherence to the legal requirements of the study country.















