1. Introduction
In this paper, we consider non-square-integrable observables $\phi : [0,1]\to {\mathbb R}$ on two simple classes of random dynamical system. One consists of randomly choosing in an independent and identically distributed (i.i.d.) manner from a finite set of maps which are strictly polynomially mixing with an indifferent fixed point at the origin, the other consisting of randomly choosing from a finite set of maps which are uniformly expanding and exponentially mixing. The main type of observable we consider is of the form ${\phi (x) = |x-x_0|^{-{1}/{\alpha }}}$ , $\alpha \in (0,2)$ which, in the i.i.d. case, lies in the domain of attraction of a stable law of index $\alpha $ . For certain results, the point $x_0$ has to satisfy some non-genericity conditions and, in particular, not be a periodic point for almost every realization of the random system (see Definition 2.3). Some of our results, particularly those involving convergence to exponential and Poisson laws, hold for general observables that are regularly varying with index $\alpha $ .
The settings for investigations on stable limit laws for observables on dynamical systems tend to be of two broad types: (1) ‘good observables’ (typically Hölder) on slowly mixing non-uniformly hyperbolic systems; and (2) ‘bad’ observables (unbounded with fat tails) on fast mixing dynamical systems. As illustrative examples of both settings, we give two results.
Example of type (1): The LSV intermittent map $T_{\gamma }: [0,1]\to [0,1]$ , $\gamma \in (0,1)$ , is defined by
The map $T_{\gamma }$ has a unique absolutely continuous invariant measure $\mu _{\gamma }$ .
Gouëzel [Reference GouëzelGou04, Theorem 1.3] showed that if $\gamma>\tfrac 12$ and $\phi : [0,1]\to {\mathbb R}$ is Hölder continuous with $\phi (0)\not =0$ , $E_{\mu _{\gamma }}(\phi )=0$ , then for $\alpha ={1}/{\gamma }$ ,
( $\beta $ has a complicated expression).
Example of type (2): Gouëzel [Reference GouëzelGou, Theorem 2.1] showed that if $T:[0,1] \to [0,1]$ is the doubling map $T(x)=2x$ (mod $1$ ) with invariant measure m, Lebesgue, and ${\phi (x) =x^{-{{1}/{\alpha }}}}$ , $\alpha \in (0,2)$ , then there exists a sequence $c_n$ such that
For further results on the first type, we refer to the influential papers [Reference GouëzelGou04, Reference GouëzelGou07] and [Reference Melbourne and ZweimüllerMZ15]. In the setting of ‘good observables’ (typically Hölder) on slowly mixing non-uniformly hyperbolic systems, the technique of inducing on a subset of phase space and constructing a Young tower has been used with some success. ‘Good’ observables lift to well-behaved observables lying in a suitable Banach space on the Young tower. This is not the case with unbounded observables with fat tails, though in [Reference GouëzelGou04], the induction technique allows an observable to be unbounded at the fixed point in a family of intermittent maps.
For further results on the second type, we refer to the papers by Marta Tyran-Kaminska [Reference Tyran-KamińskaTK10a, Reference Tyran-KamińskaTK10b]. In the setting of Gibbs–Markov maps, she shows, among other results, that functions which are measurable with respect to the Gibbs–Markov partition and in the domain of attraction of a stable law with index $\alpha $ converge (under the appropriate scaling) in the $J_1$ topology to a Lévy process of index $\alpha $ [Reference Tyran-KamińskaTK10b, Theorem 3.3, Corollaries 4.1 and 4.2].
For recent results on limit laws, though not stable laws, in the setting of skew-products with an ergodic base map and uniformly hyperbolic fiber maps, see also [Reference Dragičević, Froyland, González-Tokman and VaientiDFGTV20a]. For a still very useful survey of techniques and ideas in random dynamical systems, we refer to [Reference KiferKif98].
Our main results are given in §2. An introduction to stable laws and a discussion of modes of convergence are given in §§3 and 4. The Poisson point approach and its application to our random setting are detailed in §5. Results on convergence of return times to an exponential law and our point processes to a Poisson process are given in §6 (though the proofs of these results are delayed until §§8.1, 8.2, 9.1, and 9.2). The proofs of the main results are given in §10. We conclude in §11 with results on stable laws for the corresponding annealed systems.
2. Main results
For the sake of concreteness, we restrict ourselves to observables of the form
where $x_0$ is a non-recurrent point (see Definition 2.3) and $\alpha \in (0,2)$ , but it is possible to consider more general regularly varying observables $\phi $ which are piecewise monotonic with finitely many branches, see for instance [Reference Tyran-KamińskaTK10b, §4.2] in the deterministic case. Note that $\phi _{x_0}$ is regularity varying with index $\alpha $ .
We will be considering the following set-up with $(\Omega , \sigma )$ the full two-sided shift on finitely many symbols. In most of our settings, we take $Y=[0,1]$ .
Let $\sigma : \Omega \to \Omega $ be an invertible ergodic measure-preserving transformation on a probability space $(\Omega , {\mathcal F}, {\mathbb P})$ . For a measurable space $(Y, {\mathcal B})$ , let $\sigma :\Omega \to \Omega $ be the usual full shift and define
by
We assume F preserves a probability measure $\nu $ on $\Omega \times Y$ . We assume that $\nu $ admits a disintegration given by $\nu (d \omega , dx) = {\mathbb P}(d \omega ) \nu ^\omega (dx)$ . For all $n \ge 1$ , we have
where
which satisfies the equivariance relations $(T_\omega ^n)_{*} \nu ^\omega = \nu ^{\sigma ^n \omega }$ for ${\mathbb P}$ -almost every (a.e.) ${\omega \in \Omega} $ .
For each $\omega \in \Omega $ , we denote by $P_\omega $ the transfer operator of $T_\omega $ with respect to the Lebesgue measure m: for all $\phi \in L^\infty (m)$ and $\psi \in L^1(m)$ ,
We can then form, for $\omega \in \Omega $ and $n \ge 1$ , the cocycle
Definition 2.1. (Scaling constants)
We consider a sequence $(b_n)_{n \ge 1}$ of positive real numbers such that
Definition 2.2. (Centering constants)
We define the centering sequence $(c_n)_{n \ge 1}$ by
We now introduce two classes of random dynamical system (RDS) for which we are able to establish stable limit laws.
2.1. Random uniformly expanding maps
We consider random i.i.d. compositions with additional assumptions of uniform expansion. Let ${\mathcal S}$ be a finite collection of m piecewise $C^2$ uniformly expanding maps of the unit interval $[0,1]$ . More precisely, we assume that for each $T \in {\mathcal S}$ , there exist a finite partition ${\mathcal A}_T$ of $[0,1]$ into intervals, such that for each $I \in {\mathcal A}_T$ , T can be continuously extended as a strictly monotonic $C^2$ function on $\bar {I}$ and
The maps $T_\omega $ (determined by the zeroth coordinate of $\omega $ ) are chosen from ${\mathcal S}$ in an i.i.d. fashion according to a Bernoulli probability measure ${\mathbb P}$ on $\Omega := \{1, \ldots , m\}^{{\mathbb Z}}$ . We will denote by ${\mathcal A}_\omega $ the partition of monotonicity of $T_\omega $ , and by ${\mathcal A}_\omega ^n = \vee _{k=0}^{n-1} (T_\omega ^k)^{-1}({\mathcal A}_{\sigma ^k \omega })$ the partition associated to $T_\omega ^n$ . We introduce
the set of discontinuities of all the maps $T_\omega ^n$ . Note that ${\mathcal D}$ is at most a countable set.
In the uniformly expanding case, we also assume the conditions (LY), (Dec), and (Min). Condition (LY) is the usual Lasota–Yorke inequality while conditions (Dec) and (Min) were introduced by Conze and Raugi [Reference Conze, Raugi and AssaniCR07].
-
(LY) There exist $r \ge 1$ , $M>0$ and $D> 0$ and $\rho \in (0,1)$ such that for all $\omega \in \Omega $ and all $f \in \mathrm {BV}$ ,
$$ \begin{align*} \|P_\omega f\|_{\mathrm{BV}} \le M \|f\|_{\mathrm{BV}} \end{align*} $$and$$ \begin{align*} \mathrm{Var}(P_\omega^r f) \le \rho \mathrm{Var}(f) + D \|f\|_{L^1(m)}. \end{align*} $$
-
(Dec) There exist $C>0$ and $\theta \in (0,1)$ such that for all $n \ge 1$ , all $\omega \in \Omega $ , and all ${f \in \mathrm {BV}}$ with ${\mathbb E}_m(f) = 0$ :
$$ \begin{align*} \| P_\omega^n f \|_{\mathrm{BV}} \le C \theta^n \|f\|_{\mathrm{BV}}. \end{align*} $$
-
(Min) There exist $c>0$ such that for all $n\ge 1$ and all $\omega \in \Omega $ ,
Definition 2.3. We say that $x_0$ is non-recurrent if $x_0$ satisfies the condition $T^n_{\omega } (x_0)\not =x_0$ for all $n\ge 1$ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
Theorem 2.4. In the setting of expanding maps, assume conditions (LY), (Min), and (Dec). Suppose that $x_0 \notin {\mathcal D}$ is non-recurrent and consider the observable $\phi _{x_0}$ .
If $\alpha \in (0,1)$ , then for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , the functional stable limit holds:
in the $J_1$ topology under the probability measure $\nu ^\omega $ , where $X_{(\alpha )}(t)$ is the $\alpha $ -stable process with Lévy measure $d\Pi _\alpha (dx) = \alpha |x|^{-(\alpha + 1)}$ on $[0,\infty )$ .
If $\alpha \in [1,2)$ , then the same result holds for m-a.e. $x_0$ .
Example 2.5. ( $\beta $ -transformations)
A simple example of a class of maps satisfying conditions (LY), (Dec), and (Min) [Reference Conze, Raugi and AssaniCR07] is to take m $\beta $ -maps of the unit interval, $T_{\beta _i } (x)=\beta _i x$ (mod $1$ ). We suppose $\beta _i>1+a$ , $a>0$ , for all $\beta _i$ , $i=1,\ldots ,m$ .
2.2. Random intermittent maps
Now we consider a simple class of intermittent type maps.
Liverani, Saussol, and Vaienti [Reference Liverani, Saussol and VaientiLSV99] introduced the map $T_{\gamma }$ as a simple model for intermittent dynamics:
If $0\le \gamma <1$ , then $T_{\gamma }$ has an absolutely continuous invariant measure $\mu _{\gamma }$ with density $h_{\gamma }$ bounded away from zero and satisfying $h_{\gamma } (x)\sim C x^{-\gamma }$ for x near zero.
We form a random dynamical system by selecting $\gamma _i\in (0,1)$ , $i=1,\ldots ,m$ in an i.i.d. fashion and setting $T_i:=T_{\gamma _i}$ . The associated Markov process on $[0,1]$ has a stationary invariant measure $\nu $ which is absolutely continuous, with density h bounded away from zero.
We denote $\gamma _{\mathrm {max}}:=\max _{1\le i \le m} \{\gamma _i\}$ and $\gamma _{\mathrm {min}}:=\min _{1\le i \le m} \{\gamma _i\}$ .
Theorem 2.6. In the setting of an i.i.d. random composition of intermittent maps, suppose $\alpha \in (0,1)$ and $\gamma _{\mathrm {max}}<\tfrac 13$ . Then, for m-a.e. $x_0$ , $({1}/{b_n})\sum _{j=0}^{n-1}\phi _{x_0}\circ T_{\omega }^j \, {\stackrel {d}{\to } \,} X_{(\alpha )}(1)$ under the probability measure $\nu ^{\omega }$ for ${\mathbb P}$ -a.e. $\omega $ (recall that $c_n=0$ for $\alpha \in (0,1)$ ).
Remark 2.7. (Convergence with respect to Lebesgue measure)
We state our limiting theorems with respect to the fiberwise measures $\nu ^{\omega }$ , but by general results of Eagleson [Reference EaglesonEag76] (see also [Reference ZweimüllerZwe07]), the convergence holds with respect to any measure $\mu $ for which $\mu \ll \nu ^{\omega }$ , in particular, our convergence results hold with respect to Lebesgue measure m. Further details are given in the Appendix.
Our proofs are based on a Poisson process approach developed for dynamical systems by Tyran-Kaminska [Reference Tyran-KamińskaTK10a, Reference Tyran-KamińskaTK10b].
3. Probabilistic tools
In this section, we review some topics from probability theory.
3.1. Regularly varying functions and domains of attraction
We refer to Feller [Reference FellerFel71] or Bingham, Goldie, and Teugels [Reference Bingham, Goldie and TeugelsBGT87] for the relations between domains of attraction of stable laws and regularly varying functions. For $\phi $ regularly varying, we define the constants $b_n$ and $c_n$ as in the case of $\phi _{x_0}$ .
Remark 3.1. When $\alpha \in (0,1)$ , then $\phi $ is not integrable and one can choose the centering sequence $(c_n)$ to be identically $0$ . When $\alpha = 1$ , it might happen that $\phi $ is not integrable, and it is then necessary to define $c_n$ with suitably truncated moments as above. If $\phi $ is integrable, then center by $c_n = n{\mathbb E}_\nu (\phi )$ .
We will use the following asymptotics for truncated moments, which can be deduced from Karamata’s results concerning the tail behavior of regularly varying functions. Define p by $ \lim _{x \to \infty } ({\nu (\phi>x)}/{\nu (|\phi |>x)}) = p$ .
Proposition 3.2. (Karamata)
Let $\phi $ be regularly varying with index $\alpha \in (0,2)$ . Then, setting $\beta := 2p -1$ and, for $\epsilon> 0$ ,
and the following hold for all $\epsilon>0$ :
-
(a)
-
(b) if $\alpha \in (0,1)$ ,
-
(c) if $\alpha \in (1,2)$ ,
-
(d) if $\alpha = 1$ ,
-
(e) if $\alpha = 1$ ,
for a slowly varying function $\widetilde {L}$ .
3.2. Lévy $\alpha $ -stable processes
A helpful and more detailed discussion can be found, e.g., in [Reference Tyran-KamińskaTK10a, Reference Tyran-KamińskaTK10b].
The $X(t)$ is a Lévy stable process if $X(0)=0$ , X has stationary independent increments, and $X(1)$ has an $\alpha $ -stable distribution.
The Lévy–Khintchine representation for the characteristic function of an $\alpha $ -stable random variable $X_{\alpha ,\beta }$ with index $\alpha \in (0,2)$ and parameter $\beta \in [-1,1]$ has the form:
where:
-
• $ a_{\alpha }= \begin {cases} \beta ({\alpha }/({1-\alpha })) & \alpha \not = 1;\\ 0 &\alpha =1;\end {cases}$
-
• $\Pi _{\alpha }$ is a Lévy measure given by
$$ \begin{align*} d \Pi_{\alpha} =\alpha (p1_{(0,\infty)} (x)+(1-p)1_{(-\infty,0)} (x) ) |x|^{-\alpha -1}\,dx; \end{align*} $$ -
• $ p=({\beta +1})/{2}.$
Note that p and $\beta $ may equally serve as parameters for $X_{\alpha ,\beta }$ . We will drop the $\beta $ from $X_{\alpha ,\beta }$ , as is common in the literature, for simplicity of notation and when it plays no essential role.
3.3. Poisson point processes
Let $(T_n)_{n \ge 1}$ be a sequence of measurable transformations on a probability space $(Y, {\mathcal B}, \mu )$ . For $n \ge 1$ , we denote
Given $\phi : Y \to {\mathbb R}$ measurable, recall that we define the scaled Birkhoff sum by
for some real constants $b_n> 0$ , $c_n$ and the scaled random process $X_n(t)$ , $n\ge 1$ , by
For $X_{\alpha }(t)$ , a Lévy $\alpha $ -stable process, and $B\in \mathcal {B} ((0,\infty ) \times ({\mathbb R} \setminus \{0\}))$ , define
where $\Delta X_{\alpha }(t):=X_{\alpha }(t)-X_{\alpha }(t^-)$ .
The random variable $N_{(\alpha )}(B)$ , which counts the jumps (and their time) of the Lévy process that lie in B, is finite almost surely (a.s.) if and only if $(m \times \Pi _\alpha ) (B)< \infty $ . In that case, $N_{(\alpha )}(B)$ has a Poisson distribution with mean $(m \times \Pi _\alpha ) (B)$ .
Similarly define
$N_n(B)$ counts the jumps of the process in equation (3.4) that lie in B. When a realization $\omega \in \Omega $ is fixed, we define
Definition 3.3. We say $N_n$ converges in distribution to $N_{(\alpha )}$ and write
if and only if $N_n(B) \, {\stackrel {d}{\to } \,} N_{(\alpha )}(B)$ for all $B \in B((0, \infty ) \kern1.3pt{\times}\kern1.3pt ({\mathbb R} \setminus \{ 0 \}))$ with ${(m \kern1.3pt{\times} \Pi _\alpha ) (B) \kern1.3pt{<}\kern1.3pt \infty }$ and $(m \times \Pi _\alpha ) (\partial B) = 0$ .
4. Modes of convergence
Consider the process $X_{\alpha }$ determined by the observable $\phi $ (that is, an i.i.d. version of $\phi $ which regularly varies with the same index $\alpha $ and parameter p). We are interested in the following limits.
-
(A) Poisson point process convergence.
$$ \begin{align*} ~N^{\omega}_n \, {\stackrel{d}{\to} \,} N_{(\alpha)} \end{align*} $$with respect to $\nu ^{\omega }$ for ${\mathbb P}$ a.e. $\omega $ , where $N_{(\alpha )}$ is the Poisson point process of an $\alpha $ -stable process with parameter determined by $\nu $ , the annealed measure. -
(B) Stable law convergence.
$$ \begin{align*} ~S_n^{\omega}:=\frac{1}{b_n} \bigg[\sum_{j=0}^{n-1} \phi\circ T_{\omega}^j -c_n\bigg] \, {\stackrel{d}{\to} \,} X_{\alpha}(1) \end{align*} $$for ${\mathbb P}$ -a.e. $\omega $ , with respect to $\nu ^{\omega }$ , for $\phi $ regularly varying with index $\alpha $ and $X_{\alpha }(t)$ the corresponding $\alpha $ -stable process, for suitable scaling and centering constants $b_n$ and $c_n$ . -
(C) Functional stable law convergence.
$$ \begin{align*} ~X_n^{\omega} (t) := \frac{1}{b_n}\bigg[ \sum_{j=0}^{\lfloor nt \rfloor-1} \phi \circ T_{\omega}^j - tc_n\bigg] \, {\stackrel{d}{\to} \,} X_{\alpha} (t) \end{align*} $$in ${\mathbb D}[0,\infty )$ in the $J_1$ topology ${\mathbb P}$ -a.e. $\omega $ , with respect to $\nu ^{\omega }$ for $\phi $ regularly varying with index $\alpha $ and $X_{\alpha }(t)$ the corresponding $\alpha $ -stable process.
For the cases we are considering, the scaling constants $b_n$ are given by equation (2.2) in Definition 2.1, and the centering constants $c_n$ are given in Definition 2.2 (see also Remark 3.1).
Remark 4.1. In the limit laws for quenched systems that we obtain of type (B) and (C), the centering sequence $c_n$ does not depend on the realization $\omega $ . This is in contrast to the case of the central limit theorem (CLT), where a random centering is necessary; see [Reference Abdelkader and AiminoAA16, Theorem 9] and [Reference Nicol, Pereira and TörökNPT21, Theorem 5.3].
5. A Poisson point process approach to random and sequential dynamical systems
Our results are based on the Poisson point process approach developed by Tyran-Kamińska [Reference Tyran-KamińskaTK10a, Reference Tyran-KamińskaTK10b] adapted to our random setting (see Theorems 5.1 and 5.3). Namely, convergence to a stable law or a Lévy process follows from the convergence of the corresponding (Poisson) jump processes, and control of the small jumps.
A key role is played by Kallenberg’s theorem [Reference KallenbergKal76, Theorem 4.7] to check convergence of the Poisson point processes, $N_n\, {\stackrel {d}{\to } \,} N_{(\alpha )} $ . Kallenberg’s theorem does not assume stationarity and hence we may use it in our setting.
In this section, we provide general conditions ensuring weak convergence to Lévy stable processes for non-stationary dynamical systems, following closely the approach of Tyran-Kamińska [Reference Tyran-KamińskaTK10b]. We start from the very general setting of non-autonomous sequential dynamics and then specialize to the case of quenched random dynamical systems, which will be useful to treat i.i.d. random compositions in the later sections.
5.1. Sequential transformations
Recall the notation introduced in §3.3. Here, $(T_n)_{n \ge 1}$ is a sequence of measurable transformations on a probability space $(Y, {\mathcal B}, \mu )$ . For $n \ge 1$ , recall we define
The proof of the following statement is essentially the same as the proof of [Reference Tyran-KamińskaTK10b, Theorem 1.1].
Note that the measure $\mu $ does not have to be invariant. Moreover (see [Reference Tyran-KamińskaTK10b, Remark 2.1]), the convergence $X_n \, {\stackrel {d}{\to } \,} X_{(\alpha )}$ holds even without the condition $\mu (\phi \circ T_1^j \neq 0) = 1$ , which is used only for the converse implication of the ‘if and only if’.
Theorem 5.1. (Functional stable limit law, [Reference Tyran-KamińskaTK10b, Theorem 1.1])
Let $\alpha \in (0,2)$ and suppose that $\mu (\phi \circ T_1^j \neq 0) = 1$ for all $j\ge 0$ . Then $X_n \, {\stackrel {d}{\to } \,} X_{(\alpha )}$ in ${\mathbb D}[0, \infty )$ under the probability measure $\mu $ for some constants $b_n>0$ and $c_n$ if and only if:
-
• $N_n \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ and
-
• for all $\delta> 0$ , $\ell \ge 1$ , with $c_\alpha (\epsilon )$ given by equation (3.1),
(5.1)
Remark 5.2. In some cases, the convergence $N_n \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ does not hold, but one has convergence of the marginals, $N_n((0,1] \times \cdot ) \, {\stackrel {d}{\to } \,} N_{(\alpha )}((0,1] \times \cdot )$ . In this case, although unable to obtain a functional stable law convergence of type (C), we can in some settings prove the convergence to a stable law for the Birkhoff sums (convergence of type (B)).
In particular, we are unable to prove $N_n^\omega \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ for the case of random intermittent maps. However, in the setting of random uniformly expanding maps, we use the spectral gap to show that $N_n^\omega \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ , and then obtain the functional stable limit law.
The next statement is [Reference Tyran-KamińskaTK10b, Lemma 2.2, part (2)], which follows from [Reference Tyran-KamińskaTK10a, Theorem 3.2]. Again, the measure does not have to be invariant.
Theorem 5.3. (Stable limit law, [Reference Tyran-KamińskaTK10b, Lemma 2.2])
For $\alpha \in (0,2)$ , consider an observable $\phi $ on the probability measure $\mu $ , and $c_\alpha (\epsilon )$ given by equation (3.1).
If
and, for all $\delta> 0$ ,
then
under the probability measure $\mu $ .
5.2. Random dynamical systems
Let $\phi : Y \to {\mathbb R}$ be a measurable function such that $\nu ^\omega ( \phi \neq 0) = 1$ .
Proposition 5.4. [Reference Tyran-KamińskaTK10b, proof of Theorem 1.2]
Let $\alpha \in (0,1)$ . With $b_n$ as in Definition 2.1 and $c_n=0$ , suppose that for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
and
Then $X_n^\omega \, {\stackrel {d}{\to } \,} X_{(\alpha )}$ in ${\mathbb D}[0, \infty )$ under the probability measure $\nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
Proof. We will check that the hypothesis of Theorem 5.1 is met for ${\mathbb P}$ -a.e. $\omega $ with ${T_n = T_{\sigma ^{n-1} \omega }}$ , $\mu = \nu ^\omega $ . Recall that $c_n=c_\alpha (\epsilon ) = 0$ when $\alpha \in (0,1)$ . Using [Reference Kounias and WengKW69, Theorem 1] (see Theorem 5.6) and the equivariance of the family of measures $\{ \nu ^\omega \}_{\omega \in \Omega }$ , we have
which shows that the condition in equation (5.3) implies the condition in equation (5.1) for all $\delta> 0$ and $\ell \ge 1$ .
Remark 5.5. One could replace the condition in equation (5.3) by one similar to that in equation (5.5), and use the argument in the proof of Proposition 5.7.
Theorem 5.6. (Kounias and Weng [Reference Kounias and WengKW69, special case of Theorem 1 therein])
Assume the random variables $X_k$ are in $L^1(\mu )$ . Then
Proposition 5.7. Let $\alpha \in [1, 2)$ .
With $b_n$ and $c_n$ as in Definitions 2.1 and 2.2, and $c_\alpha (\epsilon )$ as in equation (3.1), suppose that for all $\epsilon> 0$ and all $\ell \ge 1$ ,
and that for all $\delta> 0$ ,
If $N_n^\omega \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , then $X_n^\omega \, {\stackrel {d}{\to } \,} X_{(\alpha )}$ in ${\mathbb D}[0, \infty )$ under the probability measure $\nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
Proof. As in the proof of Proposition 5.4, we check the hypothesis of Theorem 5.1 with $T_n = T_{\sigma ^{n-1} \omega }$ , $\mu = \nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ . We will see that equation (5.1) follows from equations (5.4) and (5.5).
Using the equivariance of $\{ \nu ^\omega \}_{\omega \in \Omega }$ , we see that the condition in equation (5.1) is implied by the equation (5.4) and
We next show that the condition in equation (5.5) implies equation (5.6).
Since
we obtain that, using again the equivariance, for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
Thus, the condition in equation (5.5) implies equation (5.6), which concludes the proof.
The analog for the convergence to a stable law is the following proposition.
Proposition 5.8. Suppose that for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , we have
If $\alpha \in (0,1)$ (so $c_n = 0$ ), we require in addition that
If $\alpha \in [1,2)$ , we require instead of equation (5.7) that for all $\epsilon>0$ ,
and
Then
under the probability measure $\nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
5.3. The annealed transfer operator
We assume that the random dynamical system $F:\Omega \times [0,1]\to \Omega \times [0,1]$ ,
which can also be viewed as a Markov process on $[0,1]$ , has a stationary measure $\nu $ with density h. The map $F: \Omega \times [0,1] \to \Omega \times [0,1]$ will preserve ${\mathbb P}\times \nu $ . Recall that ${{\mathbb P}:=\{(p_1, \ldots , p_m)\}^{\mathbb {Z}}}$ .
We use the notation $P_{\mu ,i}$ for the transfer operator of $T_i:[0,1]\to [0,1]$ with respect to a measure $\mu $ on $[0,1]$ , that is,
The annealed transfer operator is defined by
with adjoint
which satisfies the duality relation
As above, we assume there are sample measures $d \nu ^\omega = h_{\omega }\,d x $ on each fiber $[0,1]$ of the skew product such that
where $P_{\omega }$ is the transfer operator of $T_{\omega _0}$ with respect to the Lebesgue measure.
Therefore,
for all Borel sets $A\subset [0,1]$ .
5.4. Decay of correlations
We now consider the decay of correlation properties of the annealed systems associated to maps satisfying conditions (LY), (Dec), and (Min) and intermittent maps.
By [Reference Aimino, Nicol and VaientiANV15, Proposition 3.1] in the setting of maps satisfying conditions (LY), (Dec), and (Min), we have exponential decay in BV against $L^1$ : there are $C>0$ , $0<\unicode{x3bb} <1$ such that
In the setting of intermittent maps, by [Reference Bahsoun and BoseBB16, Theorem 1.2], we have polynomial decay in Hölder against $L^{\infty }$ : there exists $C>0$ such that
We now consider a useful property satisfied by our class of random uniformly expanding maps.
Definition 5.9. (Condition U)
We assume that almost each $\nu ^\omega $ is absolutely continuous with respect to the Lebesgue measure m, and
Consequently, the stationary measure $\nu $ is also absolutely continuous with respect to m, with density $h \in L^\infty (m)$ given by $h(x) = \int _\Omega h_\omega (x) {\mathbb P}(d\omega )$ and satisfying equation (5.8).
Lemma 5.10. Properties (LY), (Min), and (Dec) imply Condition U. Namely, there exists a unique Hölder map $\omega \in \Omega \mapsto h_\omega \in \mathrm {BV}$ such that $P_\omega h_\omega = h_{\sigma \omega }$ and equations (5.8) and (5.9) are satisfied by [Reference Aimino, Nicol and VaientiANV15].
Proof. By condition (Dec), and as all the operators $P_\omega $ are Markov with respect to m, we have
which proves that is a Cauchy sequence in BV converging to a unique limit $h_\omega \in BV$ satisfying $P_\omega h_\omega = h_{\sigma \omega }$ for all $\omega $ . The lower bound in equation (5.8) follows from the condition (Min), while the upper bound is a consequence of the uniform Lasota–Yorke inequality of condition (LY), as actually the family $\{h_\omega \}_{\omega \in \Omega }$ is bounded in BV. To prove the Hölder continuity of $\omega \mapsto h_\omega $ with respect to the distance $d_\theta $ , we remark that if $\omega $ and $\omega '$ agree in coordinates $|k| \le n$ , then
Remark 5.11. Note that the density h of the stationary measure $\nu $ also belongs to BV and is uniformly bounded from above and below, as the average of $h_\omega $ over $\Omega $ .
5.4.1. The sample measures $h_\omega $
The regularity properties of the sample measures $h_{\omega }$ , both as functions of $\omega $ and as functions of x on $[0,1]$ , play a key role in our estimates. We will first recall how the sample measures are constructed. Suppose ${\omega :=(\ldots , \omega _{-1}, \omega _0,\omega _1,\ldots , \omega _n,\ldots ,)}$ and define $h_n (\omega )=P_{\omega _{-1}}\ldots P_{\omega _{-n} }1$ as a sequence of functions on the fiber I above $\omega $ . In the setting both of random uniformly expanding maps and of intermittent maps $\{h_n (\omega ) \}$ is a Cauchy sequence and has a limit $h_{\omega }$ .
In the setting of random expanding maps, $h_\omega $ is uniformly BV in $\omega $ as
In the setting of intermittent maps with $\gamma _{\mathrm {max}}=\max _{1\le i \le m} \{\gamma _i\}$ , the densities $h_\omega $ lie in the cone
where $X(x) = x$ is the identity function and $m(f)$ is the integral of f with respect to m. In [Reference Aimino, Hu, Nicol, Török and VaientiAHN+ 15 ], it is proven that for a fixed value of $\gamma _{\mathrm {max}} \in (0,1)$ , provided that the constant a is big enough, the cone L is invariant under the action of all transfer operators $P_{\gamma _i}$ with $0<\gamma _i \leq \gamma _{\mathrm {max}}$ and so (see e.g. [Reference Nicol, Pereira and TörökNPT21, Proposition 3.3], which summarizes results of [Reference Nicol, Török and VaientiNTV18])
whence $h_\omega \in L^1(m)$ . In later arguments, we will use the approximation
We mention also the recent paper [Reference Korepanov and LeppänenKL21], where the logarithm term in equation (5.11) is shown to be unnecessary and moment estimates are given.
We now show that $h_{\omega }$ is a Hölder function of $\omega $ on $(\Omega , d_{\theta })$ in the setting of random expanding maps.
For $\theta \in (0,1)$ , we introduce on $\Omega $ the symbolic metric
where $s(\omega , \omega ') = \inf \{k \ge 0 : \omega _\ell \neq \omega _\ell ' \text { for some } |\ell |\le k\}$ .
Suppose $\omega $ , $\omega '$ agree in coordinates $|k| \le n$ (that is, backwards and forwards in time) so that $d_{\theta } (\omega , \omega ^{\prime }) \le \theta ^n$ in the symbolic metric on $\Omega $ . Then
Recall that $\|f\|_{\infty } \le C \|f\|_{\mathrm {BV}}$ , see e.g. [Reference Boyarsky and GóraBG97, Lemma 2.3.1].
That is, Condition U (see Definition 5.9) holds for random expanding maps.
The map $\omega \mapsto h_{\omega }$ is not Hölder in the setting of intermittent maps; in several arguments, we will use the regularity properties of the approximation $h_n(\omega )$ for $h_\omega $ .
However, on intervals that stay away from zero, all functions in the cone L are comparable to their mean. Therefore, on sets that are uniformly away from zero, all the above densities/measures ( $d \nu = h\,dx$ , $h_\omega $ , $h_n(\omega )$ ) are still comparable.
Namely,
Indeed, $h/m(h)$ is bounded below by [Reference Liverani, Saussol and VaientiLSV99, Lemma 2.4], and the upper bound follows from the definition of the cone.
6. Ancilliary results
Let $x_0 \in [0,1]$ , and, for $\alpha \in (0,2)$ , recall we define the function $\phi _{x_0}(x) = |x - x_0|^{ - {1}/{\alpha }}$ . It is easy to see that $\phi _{x_0}$ is regularity varying with index $\alpha $ and that $p=1$ .
6.1. Exponential law and point process results
We denote by ${\mathcal J}$ the family of all finite unions of intervals of the form $(x, y]$ , where $-\infty \le x < y \le \infty $ and $0 \notin [x,y]$ .
For a measurable subset $U \subset [0,1]$ , we define the hitting time of $(\omega , x) \in \Omega \times [0,1]$ to U by
Recall that $\phi _{x_0}(x):=d(x,x_0)^{-{1}/{\alpha }}$ depends on the choice of $x_0 \in [0,1]$ . Recall also that
the set of discontinuities of all the maps $T_\omega ^n$ .
Theorem 6.1. In the setting of §2.1, assume conditions (LY), (Min), and (Dec). If $x_0 \notin {\mathcal D}$ is non-recurrent, then, for ${\mathbb P}$ -a.e. $\omega \in \Omega $ and all $0 \le s < t$ ,
where $A_n := \phi _{x_0}^{-1}(b_n J)$ , $J \in {\mathcal J}$ .
Theorem 6.2. In the setting of intermittent maps, assume that $\gamma _{\mathrm {max}} <\tfrac 13$ . Then for m-a.e. $x_0$ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ and all $0 \le s < t$ ,
where $A_n := \phi _{x_0}^{-1}(b_n J)$ , $J \in {\mathcal J}$ .
Theorem 6.3. In the setting of §2.1, assume conditions (LY), (Min), and (Dec). If $x_0 \notin {\mathcal D}$ is non-recurrent, then for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
under the probability $\nu ^\omega $ .
Theorem 6.4. In the setting of intermittent maps for m-a.e. $x_0$ for ${\mathbb P}$ -a.e. $\omega $ ,
After some preliminary lemmas and results, Theorem 6.1 is proved in §8.1, Theorem 6.2 in §8.2, Theorem 6.3 in §9.1, and Theorem 6.4 in §9.2.
7. Scheme of proofs
7.1. Two useful lemmas
We now proceed to the proofs of the main results. We will use the following technical propositions which are a form of spatial ergodic theorem which allows us to prove exponential and Poisson limit laws.
Lemma 7.1. Assume Condition U and let $\chi _n : Y \to {\mathbb R}$ be a sequence of functions in $L^1(m)$ such that ${\mathbb E}_m(| \chi _n |) = {\mathcal O}(n^{-1} \widetilde {L}(n))$ for some slowly varying function $\widetilde {L}$ . Then, for ${\mathbb P}$ -a.e. $\omega \in \Omega $ and for all $\ell \ge 1$ ,
Therefore, given $( s, t] \subset [0, \infty )$ and $\epsilon> 0$ , for ${\mathbb P}$ -a.e. $\omega $ , there exists $N(\omega )$ such that
for all $n \ge N(\omega )$ .
Proof. We obtain the second claim by taking the difference between two values of $\ell $ in the first claim.
Fix $\ell \ge 1$ . For $\delta>0$ , let
and
Note that
We define $f_n(\omega ) = {\mathbb E}_{\nu ^\omega }(\chi _n)$ and $\overline {f}_n = {\mathbb E}_{{\mathbb P}}(f_n)$ . We claim that $f_n:\Omega \to \mathbb {R}$ is Hölder with norm $\|f_n \|_\theta = {\mathcal O}(n^{-1}\widetilde {L}(n))$ . Indeed, for $\omega \in \Omega $ , we have
and for $\omega , \omega ' \in \Omega $ , we have
since $\omega \in \Omega \mapsto h_\omega \in L^\infty (m)$ is Hölder continuous. In particular, we also have that ${\overline {f}_n = {\mathcal O}(n^{-1}\widetilde {L}(n))}$ .
We have, using Chebyshev’s inequality,
By the $\sigma $ -invariance of ${\mathbb P}$ , we have
and, since $(\Omega , {\mathbb P}, \sigma )$ admits exponential decay of correlations for Hölder observables, there exist $\unicode{x3bb} \in (0,1)$ and $C>0$ such that
We then obtain that
which implies that
Let $\eta>0$ . By the Borel–Cantelli lemma, it follows that for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , there exists $N(\omega , \delta ) \ge 1$ such that $\omega \notin B^{\lfloor p^{1+\eta }\rfloor } (\delta )$ for all $p \ge N(\omega , \delta )$ .
Let now $P:= \lfloor p^{1+\eta } \rfloor < n \le P' = \lfloor (p+1)^{1+\eta } \rfloor $ for p large enough. Let $0 \le k \le \ell $ . Then, since $\|f_n\|_\infty = {\mathcal O}(n^{-1}\widetilde {L}(n))$ ,
because on the one hand,
and on the other hand, by Potter’s bounds, for $\tau> 0$ ,
Since
for all $0 \le k \le \ell $ , it follows that for ${\mathbb P}$ -a.e. $\omega $ , there exists $N(\omega , \delta )$ such that $\omega \notin B^n(2 \delta )$ for all $n \ge N(\omega , \delta )$ , which concludes the proof.
We now consider a corresponding result to Lemma 7.1 in the setting of intermittent maps.
Lemma 7.2. Assume that $\gamma _{\mathrm {max}} < 1/2$ , and that $\chi _n \in L^1(m)$ is such that ${\mathbb E}_m( |\chi _n|) = {\mathcal O}(n^{-1})$ , $\|\chi _n\|_\infty = {\mathcal O}(1)$ and there is $\delta>0$ such that $\mathrm {supp}(\chi _n) \subset [\delta ,1]$ for all n.
Then, for ${\mathbb P}$ -a.e. $\omega \in \Omega $ and for all $\ell \ge 1$ ,
Proof. In the setting of intermittent maps, we must modify the argument of Lemma 7.1 slightly as $h_\omega $ is not a Hölder function of $\omega $ . Instead, we consider and use that, by equation (5.11),
Note that $h^i_{\omega }$ is the ith approximate to $h_{\omega }$ in the pullback construction of $h_{\omega }$ . Let $\nu _\omega ^i$ be the measure such that ${d \nu _\omega ^i}/{dm} = h_\omega ^i$ .
Consider
By equation (5.12), on the set $[\delta , 1]$ , the densities involved ( $h_\omega ^k, h_\omega , h=d \nu /d m$ ) are uniformly bounded above and away from zero. Thus, $\|f_n^i \|_\infty = {\mathcal O}(n^{-1})$ .
Pick $0<a<1$ such that $\beta :=(({1}/{\gamma _{\mathrm {max}}})-1)a -1>0$ .
For a given n, take $i=i_n=n^a$ . By equation (7.1), for all $\omega $ , n, and $i=n^a$ ,
Then
and
Given $\epsilon $ , choose n large enough that for all $0 \le k \le \ell $ ,
By Chebyshev,
We bound
and note that if $|r-u|>n^{a}$ , then by independence,
and hence, we may bound
Thus, for n large enough,
The rest of the argument proceeds as in the case of Lemma 7.1 using a speedup along a sequence $n=p^{1+\eta }$ , where $\eta>{a}/({1-a})$ , since $\|f_n\|_\infty = {\mathcal O}(n^{-1})$ still holds.
7.2. Criteria for stable laws and functional limit laws
The next theorem shows that for regularly varying observables, Poisson convergence and Condition U imply convergence in the $J_1$ topology if $\alpha \in (0,1)$ and gives an additional condition to be verified in the case $\alpha \in [1,2)$ .
Note that equation (7.2) is essentially the condition in equation (5.5) of Proposition 5.7.
Theorem 7.3. Assume $\phi $ is regularly varying, Condition U holds, and that
for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
If $\alpha \in [1, 2)$ , assume furthermore that for all $\delta> 0$ and ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
Then $X_n^\omega \, {\stackrel {d}{\to } \,} X_{(\alpha )}$ in ${\mathbb D}[0, \infty )$ under the probability measure $\nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ .
Remark 7.4. From equation (5.8) and Theorem 5.1, it follows that the convergence of $X_n^\omega $ also holds under the probability measure $\nu $ .
Proof of Theorem 7.3
When $\alpha \in (0,1)$ , we check the hypothesis of Proposition 5.4. Using equation (5.8), we have
Using Proposition 3.2, we see that the condition in equation (5.3) is satisfied since $\alpha <1$ , thus proving the theorem in this case.
When $\alpha \in [1,2)$ , we consider instead Proposition 5.7. First, we remark that the condition in equation (5.5) is implied by equations (7.2) and (5.8). It remains to check the condition in equation (5.4), which constitutes the rest of the proof.
If $\alpha \in (1,2)$ , we have
with
and
Since $\phi $ is regularity varying with index $\alpha>1$ , it is integrable and the function ${\omega \mapsto {\mathbb E}_{\nu ^\omega }(\phi )}$ is Hölder. Hence, it satisfies the law of the iterated logarithm, and we have for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
Thus, we have
As a consequence, we can deduce that $\lim _{n \to \infty } \sup _{0 \le t \le \ell } A_n^\omega (t) = 0 $ since $b_n = n^{\frac {1}{\alpha }} \widetilde {L}(n)$ for a slowly varying function $\widetilde {L}$ , with $\alpha <2$ .
By Proposition 3.2, we also have
In particular, we have
This also implies that ${\mathbb E}_m(| \chi _n| ) = {\mathcal O}(n^{-1})$ if we define . From Lemma 7.1, it follows that $\lim _{n \to \infty } \sup _{0 \le t \le \ell } B_{n, \epsilon }^\omega (t) = 0$ .
Putting all these estimates together concludes the proof when $\alpha \in (1,2)$ .
When $\alpha = 1$ , we estimate the right-hand side of equation (7.3) by $A_{n, \epsilon }^\omega (t) + B_{n, \epsilon }^\omega (t)$ with
and
We define . By Proposition 3.2, we have ${\mathbb E}_m(| \chi _n |) = {\mathcal O}(n^{-1} \widetilde {L}(n))$ for some slowly varying function $\widetilde {L}$ , and so by Lemma 7.1,
However, by Proposition 3.2, we have
and so $\lim _{n \to \infty } \sup _{0 \le t \le \ell } B_{n, \epsilon }^\omega (t) = 0$ which completes the proof.
8. An exponential law
We denote by ${\mathcal J}$ the family of all finite unions of intervals of the form $(x, y]$ , where ${- \infty \le x < y \le \infty }$ and $0 \notin [x,y]$ . For $J \in {\mathcal J}$ , we will establish a quenched exponential law for the sequence of sets $A_n = (\phi _{x_0})^{-1}(b_n J)$ . Similar results were obtained in [Reference Crimmins and FroylandCF20, Reference Freitas, Freitas and VaientiFFV17, Reference Haydn, Rousseau and YangHRY20, Reference Rousseau, Saussol and VarandasRSV14, Reference Rousseau and ToddRT15].
Since $\phi $ is regularly varying, it is easy to verify that
In particular, $ m(A_n) = {\mathcal O}(n^{-1})$ .
Lemma 8.1. Assume Condition U and that $\phi $ is regularly varying with index $\alpha $ .
If $A_n \subset [0,1]$ is a sequence of measurable subsets such that $m(A_n) = {\mathcal O}(n^{-1})$ , then for all $0 \le s < t$ ,
The same result holds in the setting of intermittent maps if $A_n \subset [\delta , 1]$ for some $\delta>0$ with $m(A_n) = {\mathcal O}(n^{-1})$ . In particular, if $A_n = \phi _{x_0}^{-1}(b_n J)$ for $J \in {\mathcal J}$ and $x_0 \neq 0$ , then for all $0 \le s < t$ ,
Proof. For the first statement, it suffices to apply Lemma 7.1 or Lemma 7.2 with . The second statement immediately follows since $\lim _n n \nu (A_n) = \Pi _\alpha (J)$ .
Corollary 8.2. Assume the hypothesis of Lemma 8.1.
Let $J \in {\mathcal J}$ and set $A_n = \phi ^{-1}(b_n J)$ . Then for ${\mathbb P}$ -a.e. $ \omega \in \Omega $ and all $0 \le s < t$ ,
Proof. Since $\nu ^{\omega }(A_n)$ is of order at most $n^{-1}$ uniformly in $\omega \in \Omega $ , it follows that
By Lemma 8.1,
which yields the conclusion.
Definition 8.3. For a measurable subset $U \subset Y = [0,1]$ , we define the hitting time of $(\omega , x) \in \Omega \times Y$ to U by
and the induced measure by $\nu $ on U by
To establish our exponential law, we will first obtain a few estimates, based on the proof of [Reference Hirata, Saussol and VaientiHSV99, Theorem 2.1], to relate $\nu ^\omega (R_{A_n}(\omega )> \lfloor nt \rfloor )$ to $\sum _{j=0}^{\lfloor nt \rfloor - 1} \nu ^{\sigma ^j \omega }(A_n)$ so that we are able to invoke Corollary 8.2.
The next lemma is basically [Reference Rousseau, Saussol and VarandasRSV14, Lemma 6].
Lemma 8.4. For every measurable set $U \subset [0,1]$ , we have the bound
where
and
Proof. Note that $\{R_U(\omega )> k\}=[T_{\omega }^{1}]^{-1} (U^c \cap \{R_U(\sigma \omega ) > k-1\})$ and so, using the equivariance of $\{\nu ^\omega \}_{\omega \in \Omega }$ ,
Hence,
We note that
Iterating we obtain, using the fact that for ${\mathbb P}$ -a.e. $\omega $ , $\nu ^{\omega } (R_U(\omega )\ge 1)=1$ ,
which yields the conclusion.
We will estimate now the coefficients $c_\omega (U)$ .
Lemma 8.5. For any measurable subset $U \subset Y$ such that , we have, for all N,
and
Proof. The estimates in equation (8.2) follow from
and therefore
For equation (8.1), note that
If $j \le N$ , then
If $j> N$ , we write
To bound terms (a) and (c), note that
so
and similarly for $\nu _U^{\omega }$ .
To bound term (b), we use the decay of $P_\omega ^k$ . Setting $V = \{ R_U(\sigma ^N \omega ) \le j - N \}$ , we have
8.1. Exponential law: proof of Theorem 6.1
We can now prove the exponential law for $A_n = \phi ^{-1}(b_n J)$ , $J \in {\mathcal J}$ .
Proof of Theorem 6.1
Due to rounding errors when taking the integer parts, we have
and it is thus enough to prove the convergence of $\nu ^{\sigma ^{\lfloor ns \rfloor }\omega } (R_{A_n}(\sigma ^{\lfloor ns \rfloor }\omega )> \lfloor nt \rfloor - \lfloor ns \rfloor )$ .
By Lemmas 8.4 and 8.5, for all $N \ge 1$ , we have
with
and
To estimate (I), we choose $\epsilon> 0$ such that $J \subset \{ |x|> \epsilon \}$ and we introduce ${V_n = \{ | \phi |> \epsilon b_n \}}$ . For a measurable subset $V \subset Y$ , we also define the shortest return to V by
and we set
We have
It follows from condition (Dec) that
as BV is a Banach algebra, and both and $\|h_{\sigma ^j \omega } \|_{\mathrm { BV}}$ are uniformly bounded. (Recall that, from the definition of $\phi $ , it follows that $V_n$ is an open interval, and thus has a uniformly bounded BV norm.)
Consequently,
However, we have by equation (8.2),
and it follows from condition (Dec) that
since $\{h_\omega \}_{\omega \in \Omega }$ is a bounded family in BV, $A_n$ is the union of at most two intervals, and thus is uniformly bounded. We can thus bound equation (8.4) by
and, assuming for the moment that $r(V_n) \to + \infty $ , we obtain the conclusion by choosing $N = N(n) = 2 \log n$ and letting $n \to \infty $ .
It thus remains to show that $r(V_n) \to + \infty $ . Recall that $V_n$ is the ball of center $x_0$ and radius $b^{-1} \epsilon ^{-\alpha } n^{-1}$ . Let $R \ge 1$ be a positive integer. Since $x_0$ is assumed to be non-recurrent, and that the collection of maps $T_\omega ^j$ for $\omega \in \Omega $ and $0 \le j < R$ is finite, we have that
is positive. Since all the maps $T_\omega ^j$ are continuous at $x_0$ by assumption, there exists $n_R \ge 1$ such that for all $n \ge n_R$ , $j < R$ , and $\omega \in \Omega $ ,
Increasing $n_R$ if necessary, we can assume that $b^{-1} \epsilon ^{-\alpha } n^{-1} < ({\delta _R}/{2})$ for all $n \ge n_R$ .
Then, for all $n \ge n_R$ , $\omega \in \Omega $ , $j < R$ , and $x \in V_n$ , we have
and thus $T_\omega ^j(x) \notin V_n$ .
This implies that $r(V_n)> R$ for all $n \ge n_R$ , which concludes the proof as R is arbitrary.
Remark 8.6. A quenched exponential law for random piecewise expanding maps of the interval is proved in [Reference Haydn, Rousseau and YangHRY20, Theorem 7.1]. Our proof follows the same standard approach. We are able to specify that Theorem 6.1 holds for non-recurrent $x_0$ , since our assumptions imply decay of correlations against $L^1$ observables, which is known to be necessary for this purpose, see [Reference Aytaç, Freitas and VaientiAFV15, §3.1]. Our proof is shorter, as we consider the simpler setting of finitely many maps, which are all uniformly expanding. In addition, we use the exponential law in the intermittent case of [Reference Haydn, Rousseau and YangHRY20, Theorem 7.2] to establish the short returns condition of Lemma 8.7 below.
8.2. Exponential law: proof of Theorem 6.2
To prove the exponential law in the intermittent setting, Theorem 6.2, we need a genericity condition on the point $x_0$ in the definition (2.1) of $\phi _{x_0}$ .
Lemma 8.7. If $\gamma _{\mathrm {max}}<\tfrac 13$ , for m-a.e. $x_0$ and for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
for all $c>0$ and all $0 \le s < t$ .
Proof. Let $N = \lfloor n (\log n)^{-1} \rfloor $ an $V_n = B_{c n^{-1}}(x_0)$ . First, we remark that for m-a.e. $x_0$ and ${\mathbb P}$ -a.e. $\omega $ ,
This is a consequence of [Reference Haydn, Rousseau and YangHRY20, Theorem 7.2]. Their result is stated for two intermittent LSV maps both with $\gamma <\tfrac 13$ but generalizes immediately to a finite collection of maps with a uniform bound of $\gamma _{\mathrm {max}}<\tfrac 13$ . The exponential law for return times to nested balls implies that for a fixed t, for m-a.e $x_0$ , and ${\mathbb P}$ -a.e. $\omega $ ,
which shows in particular, since $\{R_{V_n}(\omega ) \le N \} \subset \{R_{V_n}(\omega ) \le nt \}$ for all n large enough, that for all $t>0$ , m-a.e $x_0$ , and ${\mathbb P}$ -a.e. $\omega $ ,
Using equation (5.12), taking the limit $t \to 0$ proves equation (8.5). Note that, even though the set of full measure of $x_0$ and $\omega $ such that equation (8.6) holds may depend on t, it is enough to consider only a sequence $t_k \to 0$ .
Now, for $k \ge 0$ and $n_0 \ge 1$ , we introduce the set
According to equation (8.5), we have for all $k \ge 0$ ,
By the Birkhoff ergodic theorem, for all $k \ge 0$ , $n_0 \ge 1$ , and ${\mathbb P}$ -a.e. $\omega $ ,
which implies that for all $0 \le s < t$ ,
Let $n_0 = n_0(\omega , k)$ such that ${\mathbb P}(\Omega _k^{n_0}) \ge 1 - 2^{-k}$ , and for all $n \ge n_0$ ,
Then, for all $n \ge n_0(\omega , k)$ , we have
Consequently,
This proves that
and the result follows by taking the limit $k \to \infty $ .
Note that the set of $x_0$ and $\omega $ for which the lemma holds depends a priori on $c>0$ , but it is enough to consider a countable and dense set of c, since for $c < c'$ ,
The exponential law for random intermittent maps follows from Lemma 8.7.
Proof of Theorem 6.2
We consider the three terms in equation (8.4) with ${N = \lfloor n (\log n)^{-1} \rfloor }$ .
Let $V_n = \{ | \phi |> \epsilon b_n\}$ , where $\epsilon>0$ is such that $A_n \subset V_n$ for all $n\ge 1$ . Since $V_n$ is a ball of center $x_0$ and radius $b^{-1} \epsilon ^{-\alpha } n^{-1}$ , and since $V_n \subset [\delta , 1]$ , the term
tends to zero by Lemma 8.7 for m-a.e $x_0$ .
The term
also tends to zero since $N=o(n)$ . Lastly, we consider
We approximate by a $C^1$ function g such that $\|g\|_{C^1}\le n^{\tau }$ , on $A_n$ , and (recall $A_n$ is two intervals of length roughly ${1}/{n}$ so a simple smoothing at the endpoints of the intervals allows us to find such a function g). Later we will specify $\tau>1$ as needed. By [Reference Nicol, Pereira and TörökNPT21, Lemma 3.4] with $h = h_\omega $ and ${\varphi = g - m(g h_\omega )}$ , for all $\omega $ ,
Using the decomposition , we estimate, leaving out the log term,
where the value of C may change line to line. Taking $\gamma _{\mathrm {max}}<\tfrac 13$ and $1<\tau <({1}/{\gamma _{\mathrm {max}}})-2$ suffices.
9. Point process results
We now proceed to the proof of the Poisson convergence. In §11, we will consider an annealed version of our results.
9.1. Uniformly expanding maps: proof of Theorem 6.3
Recall Theorem 6.3: under the conditions of §2.1, in particular conditions (LY), (Min), and (Dec), if $x_0 \notin {\mathcal D}$ is non-recurrent, then for ${\mathbb P}$ -a.e. $\omega \in \Omega $ ,
under the probability measure $\nu ^\omega $ .
Our proof of Theorem 6.3 uses the existence of a spectral gap for the associated transfer operators $P^n_{\omega }$ , and breaks down in the setting of intermittent maps. The use of the spectral gap is encapsulated in the following lemma.
Lemma 9.1. Assume condition (LY). Then there exists $C>0$ such that for all $\omega \in \Omega $ , all $f, f_n \in \mathrm {BV}$ with
we have
Proof. We proceed in four steps.
Step 1. We define
where we have set . We observe that for all $n \ge 0$ , there exists $C_n> 0$ such that for all $\omega \in \Omega $ ,
The first estimate is immediate, and the second follows, because
and so we can define by induction $C_{n+1} = C_n + \sup _{\omega \in \Omega } \# {\mathcal A}_\omega ^{n+1}$ which is finite, as there are only finitely many maps in ${\mathcal S}$ .
Step 2. We first prove the lemma in the case where $r=1$ in the condition (LY). Before, we claim that for $f \in \mathrm {BV}$ and sequences $(f_j) \subset \mathrm {BV}$ as in the statement, we have
This implies the lemma when $r=1$ , since
and
We prove the claim by induction on $n \ge 0$ . It is immediate for $n=0$ , and for the induction step, we have, using condition (LY),
which proves equation (9.2) for $n+1$ , assuming it holds for n.
Step 3. Now, we consider the general case $r \ge 1$ and we assume that n is of the particular form $n = pr$ , with $p \ge 0$ . We note that the random system defined with ${\mathcal T} = \{ T_\omega ^r \}_{\omega \in \Omega }$ satisfies the condition (LY) with $r = 1$ . Consequently, by the second step and equation (9.1), we have
Step 4. Finally, if $n = pr + q$ , with $p \ge 0$ and $q \in \{0, \ldots , r-1\}$ , as an immediate consequence of condition (LY), we obtain
However, from Step 3, we have
and, using equation (9.1),
which concludes the proof of the lemma.
Proof of Theorem 6.3
We denote by ${\mathcal R}$ the family of finite unions of rectangles R of the form $R = (s,t] \times J$ with $J \in {\mathcal J}$ . By Kallenberg’s theorem, see [Reference KallenbergKal76, Theorem 4.7] or [Reference ResnickRes87, Proposition 3.22], $N_n^\omega \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ if for any $R \in {\mathcal R}$ ,
and
We first prove equation (b). We write
with $R_i = (s_i, t_i] \times J_i$ disjoint.
Then
and
By Lemma 8.1, for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , we have
which proves equation (b).
We next establish equation (a). We will use induction on the number of ‘time’ intervals $(s_i,t_i] \subset (0,\infty ]$ . Let $R=(s_1,t_1]\times J_1$ , where $J_1\in \mathcal {J}$ . Define
Since
we have that
because, due to rounding when taking integer parts, $\lfloor n t_1 \rfloor - \lfloor n s_1 \rfloor -1$ is either equal to $\lfloor n( t_1 - s_1) \rfloor -1$ or to $\lfloor n(t_1 - s_1) \rfloor $ . By Theorem 6.1,
as desired.
Now let $R=\bigcup _{j=1}^k (s_i, t_i]\times J_i$ with $0\le s_1 < t_1<\cdots < s_k < t_k$ and $J_i \in \mathcal {J}$ . Furthermore, define $s_i'=s_i-s_1$ and $t_i'=t_i-s_1$ .
Observe that, accounting for the rounding errors when taking integer parts as for equation (9.3), we get
so, after replacing $\omega $ by $\sigma ^{\lfloor n s_1 \rfloor }\omega $ , we can assume that $s_1=0$ . Let
Then, with $A_n =\phi _{x_0}^{-1} (b_n J_1)$ ,
as $n\to \infty $ , uniformly in $\eta \in \Omega $ , as in equation (9.4). Moreover, as we check below,
as $n\to \infty $ , uniformly in $\eta \in \Omega $ . Therefore, setting $\eta = \sigma ^{\lfloor n s_2 \rfloor } \omega $ in equations (9.5) and (9.6), we have, by Theorem 6.1,
which gives the induction step in the proof of equation (a).
We prove now equation (9.6). Our proof uses the spectral gap for $P^{n}_{\omega }$ and breaks down for random intermittent maps.
Similarly to equation (9.4),
We have, using the notation
that
where the last inequality follows from the decay, uniform in $\eta $ , of $\{P_{\eta }^k\}_{k}$ in BV (condition (Dec)).
However,
which proves equation (9.6). This follows from Lemma 9.1 applied to $f = h_\eta $ and , because
and both $\|h_\eta \|_{\mathrm {BV}}$ and are uniformly bounded. Note that for the stationary case, the estimate in equation (9.7) is used in the proof of [Reference Tyran-KamińskaTK10b, Theorem 4.4], which refers to [Reference Aaronson, Denker, Sarig and ZweimüllerADSZ04, Proposition 4].
9.2. Intermittent maps: proof of Theorem 6.4
We prove a weaker form of convergence in the setting of intermittent maps, which suffices to establish stable limit laws but not functional limit laws.
In the setting of intermittent maps, we will show that for ${\mathbb P}$ -a.e. $\omega $ ,
Proof of Theorem 6.4
We will show that for ${\mathbb P}$ -a.e. $\omega \in \Omega $ , the assumptions of Kallenberg’s theorem [Reference KallenbergKal76, Theorem 4.7] hold.
Recall that $\mathcal {J}$ denotes the set of all finite unions of intervals of the form $(x,y]$ , where $x <y$ and $0\not \in [x,y]$ .
By Kallenberg’s theorem [Reference KallenbergKal76, Theorem 4.7], $N^{\omega }_n[(0,1]\times \cdot ) \to ^{d} N_{(\alpha )} ((0,1] \times \cdot )$ if for all $J \in \mathcal {J}$ ,
and
We prove first equation (b) following [Reference Tyran-KamińskaTK10b, p. 12]. Write
with $J_i=(x_i,y_i]$ disjoint.
Then
and
We check that
for $J=\bigcup _{i=1}^k J_i$ .
Write $A_n:=\phi _{x_0}^{-1} (b_n J)$ . Then
hence
by Lemma 7.2.
Now we prove equation (a), that is,
for all $J \in \mathcal {J}$ .
Let $J \in \mathcal {J}$ and denote as above $A_n := \phi _{x_0}^{-1} (b_n J)\subset X =[0,1]$ . Then
Hence,
and by Theorem 6.2, for m-a.e. $x_0$
This proves equation (a).
10. Stable laws and functional limit laws
10.1. Uniformly expanding maps: proof of Theorem 2.4
In this section, we prove Theorem 2.4, under the conditions given in §2.1, in particular, conditions (LY), (Dec), and (Min).
For this purpose, we consider first some technical lemmas regarding short returns. For $\omega \in \Omega $ , $n \ge 1$ , and $\epsilon>0$ , let
Lemma 10.1. There exists $C>0$ such that for all $\omega \in \Omega $ , $n \ge 1$ , and $\epsilon>0$ ,
Proof. We follow the proof of [Reference Holland, Nicol and TörökHNT12, Lemma 3.4], conveniently adapted to our setting of random non-Markov maps. Recall that ${\mathcal A}_\omega ^n$ is the partition of monotonicity associated to the map $T_\omega ^n$ . Consider $I \in {\mathcal A}_\omega ^n$ . Since $\inf _I |(T_\omega ^n)'| \ge \unicode{x3bb} ^n> 1$ , there exists at most one solution $x_I^\pm \in I$ to the equation
and since there is no sign change of $(T_\omega ^n)'$ on I, we have
We have
and by the mean value theorem,
Consequently,
Note that if there is no solutions to equation (10.1), then the estimate in equation (10.3) is actually improved. Rearranging equation (10.3) and summing over $I\in {\mathcal A}_\omega ^n$ , we obtain, thanks to equation (10.2),
The fact that
for a constant $C>0$ independent from $\omega $ and n follows from a standard distortion argument for one-dimensional maps that can be found in the proof of [Reference Aimino, Nicol and VaientiANV15, Lemma 8.5(3)] (see also [Reference Aimino and RousseauAR16, Lemma 7]), where finitely many piecewise $C^2$ uniformly expanding maps with finitely many discontinuities are also considered. Since it follows from condition (LY) that $\| P_\omega ^n f \|_{\mathrm {BV}} \le C \|f\|_{\mathrm {BV}}$ for some uniform $C>0$ , we do not have to average equation (10.4) over $\omega $ as in [Reference Aimino, Nicol and VaientiANV15], but instead, we can simply have an estimate that holds uniformly in $\omega $ .
Recall that, for a measurable subset U, $R_U^\omega (x) \ge 1$ is the hitting time of $(\omega , x)$ to U defined by equation (6.1).
Lemma 10.2. Let $a>0$ , $\frac 2 3 < \psi < 1$ , and $0 < \kappa < 3 \psi - 2$ . Then there exist sequences $(\gamma _1(n))_{n \ge 1}$ and $(\gamma _2(n))_{n \ge 1}$ with $\gamma _1(n) = {\mathcal O}(n^{- \kappa })$ and $\gamma _2(n) = o(1)$ , and for all $\omega \in \Omega $ , a sequence of measurable subsets $(A_n^\omega )_{n \ge 1}$ of $[0,1]$ with $m(A_n^\omega ) \le \gamma _1(n)$ and such that for all $x_0 \notin A_n^\omega $ ,
Proof. Let
Since $ B_{n^{-\psi }}(x_0) \cap \{R_{ B_{n^{-\psi }}(x_0)}^{\sigma ^i \omega } \le \lfloor a \log n \rfloor \} \subset B_{n^{-\psi }}(x_0) \cap E_n^{\sigma ^i \omega }$ , it is enough to consider
According to Lemma 10.1, we have
We introduce the maximal function
By [Reference RudinRud87, Equation (5), p. 138], for all $\unicode{x3bb}> 0$ , we have
Let $\rho>0$ and $\xi>0$ to be determined later. We define
so that we have
By definition of the maximal function $M_n^\omega $ , this implies that
from which it follows, by equation (10.5) with $\unicode{x3bb} = ( \log n) n^{\xi - \psi \rho }$ ,
where
If $x_0 \notin A_n^\omega $ , then
Since $\tfrac 23 < \psi < 1$ and $0 < \kappa < 3 \psi - 2$ , it is possible to choose $\rho>0$ and $\xi>0$ such that $\kappa = \xi + (1-\rho )\psi - 1$ , $\psi> \xi $ , and $\psi (1+\rho )>1$ (for instance, take $\xi = \psi - \delta $ and $\rho = \psi ^{-1} - 1 + \delta \psi ^{-1}$ with $\delta = ({3 \psi - 2 - \kappa })/{2}$ ), which concludes the proof.
Lemma 10.3. Suppose that $a>0$ and $\frac 3 4 < \psi < 1$ . Then for m-a.e. $x_0 \in [0,1]$ and ${\mathbb P}$ -a.e. $\omega \in \Omega $ and, we have
Proof. Let $0 < \kappa < 3 \psi - 2$ to be determined later. Consider the sets $(A_n^\omega )_{n \ge 1}$ given by Lemma 10.2, with $m(A_n^\omega ) \le \gamma _1(n) = {\mathcal O}(n^{- \kappa })$ . Since $\kappa < 1$ , we need to consider a subsequence $(n_k)_{k \ge 1}$ such that $\sum _{k \ge 1} \gamma _1(n_k) < \infty $ . For such a subsequence, by the Borel–Cantelli lemma, for m-a.e. $x_0$ , there exists $K = K(x_0, \omega )$ such that for all $k \ge K$ , $x_0 \notin A_{n_k}^\omega $ . Since $\lim _{k \to \infty } \gamma _2(n_k) = 0$ , this implies
We take $n_k = \lfloor k^\zeta \rfloor $ for some $\zeta>0$ to be determined later. To have $\sum _{k \ge 1} \gamma _1(n_k) < \infty $ , we need to require that $\kappa \zeta> 1$ . Set $U_n^\omega (x_0) = B_{n^{-\psi }}(x_0) \cap \{R_{ B_{n^{-\psi }}(x_0)}^\omega \le \lfloor a \log n \rfloor \}$ . To obtain the convergence to $0$ of the whole sequence, we need to prove that
For this purpose, we estimate
where
Before proceeding to estimate each term, we note that $|n_{k+1} - n_k| = {\mathcal O}(k^{-(1- \zeta )})$ , $|n_{k+1}^{- \psi } - n_k^{-\psi }| = {\mathcal O}(k^{-(1 + \zeta \psi )})$ , $| \log n_{k+1} - \log n_k | = {\mathcal O}(k^{-1})$ and $m(U_n^\omega (x_0)) \le m(B_{n^{-\psi }}(x_0)) = {\mathcal O}(k^{-\zeta \psi })$ .
From these observations, it follows
and
To obtain equation (10.6), it is thus sufficient to choose $\kappa>0$ and $\zeta>0$ such that $\kappa < 3 \psi - 2$ , $\kappa \zeta> 1$ , and $(1 - \psi ) \zeta < 1$ , which is possible if $\psi> \frac 3 4$ .
We can now prove the functional convergence to a Lévy stable process for i.i.d. uniformly expanding maps.
Proof of Theorem 2.4
We apply Theorem 7.3. By Theorem 6.3, we have $N_n^\omega \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ under the probability $\nu ^\omega $ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ . It thus remains to check that equation (7.2) holds for m-a.e. $x_0$ when $\alpha \in [1,2)$ to complete the proof. For this purpose, we will use a reverse martingale argument from [Reference Nicol, Török and VaientiNTV18] (see also [Reference Aimino and RousseauAR16, Proposition 13]). Because of equation (5.8), it is enough to work on the probability space $([0,1], \nu ^\omega )$ for ${\mathbb P}$ -a.e. $\omega \in \Omega $ . Let $\mathcal {B}$ denote the $\sigma $ -algebra of Borel sets on $[0,1]$ and
To simplify notation a bit, let
From equation (5.8), it follows that ${\mathbb E}_m(|f_{\omega , j, n}|) \le C \epsilon b_n$ , and from the explicit definition of $\phi $ , we can estimate the total variation of $f_{\omega ,j,n}$ and obtain the existence of $C> 0$ , independent of $\omega $ , $\epsilon $ , n, and j, such that
We define
and
Hence, $H_{\omega , 1, n}=0$ and an explicit formula for $H_{\omega , k, n}$ is
From the explicit formula, the exponential decay in the BV norm of $P_{\sigma ^j \omega }^{n-j}$ from condition (Dec), equations (5.8) and (10.7), we see that $\|H_{\omega , k, n}\|_{\mathrm {BV}}\le C \epsilon b_n$ , where the constant C may be taken as constant over $\omega \in \Omega $ . If we define
then the sequence $\{ M_{\omega , k, n}\}_{k \ge 1}$ is a reverse martingale difference for the decreasing filtration $\mathcal {B}_{\omega , k}=(T_{\omega }^n)^{-1}( \mathcal {B})$ as
The martingale reverse differences are
where
We see from the $L^{\infty }$ bounds on $\| H_{\omega , k, n}\|_{\infty } \le C b_n \epsilon $ and the telescoping sum that
By Doob’s martingale maximal inequality,
Note that
by pairwise orthogonality of martingale reverse differences.
As in [Reference Haydn, Nicol, Török and VaientiHNTV17, Lemma 6],
So we see that
where we have used $\| H^2_{\omega , j, n}\|_{\infty } \le C^2 b_n^2 \epsilon ^2$ .
Now we estimate
Using the equivariance of the measures $\{ \nu ^\omega \}_{\omega \in \Omega }$ and equation (5.8), we have
by Proposition 3.2 and that
since $\phi _{x_0}$ is regularly varying.
However, we are going to show that for m-a.e. $x_0$ ,
The first observation is that, due to condition (Dec),
where $\theta <1$ . Hence, there exists $a> 0$ independently of n and $\epsilon $ such that
and it is enough to prove that for $\epsilon> 0$ ,
By construction, the term ${\mathbb E}_{\nu ^\omega }[f_{\omega , i, n} \circ T_\omega ^i \cdot f_{\omega , j, n} \circ T_\omega ^j]$ is a covariance, and since $\phi $ is positive, we can bound this quantity by ${\mathbb E}_{\nu ^\omega }[f \circ T_\omega ^i \cdot f \circ T_\omega ^j] = {\mathbb E}_{\nu ^{\sigma ^i \omega }}[f_n \cdot f_n \circ T_{\sigma ^i \omega }^{j-i}]$ , where . Then, since the densities are uniformly bounded by equation (5.8), we are left to estimate
Let $\tfrac 3 4 < \psi < 1$ and $U_n = B_{n^{-\psi }}(x_0)$ . We bound equation (10.14) by $\mathrm {(I)} + \mathrm {(II)} + \mathrm {(III)}$ , where
and
Since $\|f_n\|_\infty \le \epsilon b_n$ , it follows that
which by Lemma 10.3 is a $o(b_n^2)$ as $n \to \infty $ for m-a.e. $x_0$ .
To estimate terms (II) and (III), we will use Hölder’s inequality. We first observe by a direct computation that
We consider term (III) first. Let $ A = U_n^c$ . We have
By equation (10.15), $( \int _A \phi _{x_0}^2\,dm )^{ 1/ 2} \le C n^{{\psi }/{2} ({2}/{\alpha } - 1 )}$ and by Proposition 3.2, $( \int f_n^2\,dm )^{ 1/ 2} \le C n^{{1}/{\alpha } - 1 /2}$ . Hence, we may bound equation (10.16) by $C n^{(1+\psi )({1}/{\alpha } - 1 /2)}$ .
To bound term (II), let $B= U_n \cap (T_{\sigma ^i \omega }^{j-i})^{-1}(U_n^c)$ . Then,
As before, $( \int f_n^2 dm )^{ 1 /2} \le C n^{{1}/{\alpha } - 1 /2}$ and
by equation (10.15), and so equation (10.18) is bounded by $C n^{(1+\psi )({1}/{\alpha } - 1 /2)}$ .
It follows that $\mathrm {(II)} + \mathrm {(III)} \le C (\log n) n^{1 + (1+\psi )({1}/{\alpha } - 1/ 2)} = o(n^{ 2/ \alpha })$ , since $ \psi < 1$ . This proves that equation (10.14) is $o(b_n^2)$ and concludes the proof of equation (10.13).
Finally, from equations (10.11), (10.12), and (10.13), we obtain
which gives the result by taking the limit first in n and then in $\epsilon $ in equation (10.10).
10.2. Intermittent maps: proof of Theorem 2.6
We prove convergence to a stable law in the setting of intermittent maps when $\alpha \in (0,1)$ .
Proof of Theorem 2.6
We apply Proposition 5.8. By Theorem 6.4, it remains to prove equation (5.7), since $\alpha \in (0,1)$ . We will need an estimate for which is independent of $\omega $ . For this purpose, we introduce the absolutely continuous probability measure $\nu _{\mathrm {\max }}$ whose density is given by $h_{\mathrm {\max }}(x) = \kappa x^{- \gamma _{\mathrm {\max }}}$ . Since all densities $h_\omega $ belong to the cone L, we have that $h_\omega \le ({a}/{\kappa }) h_{\mathrm {\max }}$ for all $\omega $ . Thus,
We can easily verify that $\phi _{x_0}$ is regularly varying of index $\alpha $ with respect to $\nu _{\mathrm {\max }}$ , with scaling sequence equal to $(b_n)_{n \ge 1}$ up to a multiplicative constant factor. Consequently, by Proposition 3.2, we have that, for some constant $c>0$ ,
which implies equation (5.7).
11. The annealed case
In this section, we consider the annealed counterparts of our results. Even though the annealed versions do not seem to follow immediately from the quenched version, it is easy to obtain them from our proofs in the quenched case. We take $\phi _{x_0} (x)=d(x,x_0)^{-{1}/{\alpha }}$ as before we consider the convergence on the measure space $\Omega \times [0,1]$ with respect to $\nu (d\omega , dx)={\mathbb P} (d \omega ) \nu ^{\omega } (dx)$ . We give precise annealed results in the case of Theorems 2.4 and 2.6, where we consider
viewed as a random process defined on the probability space $(\Omega \times [0,1], \nu )$ .
Theorem 11.1. Under the same assumptions as Theorem 2.4, the random process $X^a_n (t)$ converges in the $J_1$ topology to the Lévy $\alpha $ -stable process $X_{(\alpha )}(t)$ under the probability measure $\nu $ .
Proof. We apply [Reference Tyran-KamińskaTK10b, Theorem 1.2] to the skew-product system $(\Omega \times [0,1], F, \nu )$ and the observable $\phi _{x_0}$ naturally extended to $\Omega \times [0,1]$ . Recall that $\nu $ is given by the disintegration $\nu (d \omega , dx) = {\mathbb P}(d \omega ) \nu ^\omega (dx)$ .
We have to prove that:
-
(a) $N_n \, {\stackrel {d}{\to } \,} N_{(\alpha )}$ ;
-
(b) if $\alpha \in [1, 2)$ , for all $\delta> 0$ ,
where
To prove item (a), we take $f \in C_K^+((0, \infty ) \times ({\mathbb R} \setminus \{ 0 \}))$ arbitrary. Then, by Theorem 6.3, we have for ${\mathbb P}$ -a.e. $\omega $ ,
Integrating with respect to ${\mathbb P}$ and using the dominated convergence theorem yields
which proves item (a).
To prove item (b), we simply have to integrate with respect to ${\mathbb P}$ in the estimates in the proof of Theorem 2.4, which hold uniformly in $\omega \in \Omega $ , and then to take the limits as $n \to \infty $ and $\epsilon \to 0$ .
Similarly, we have the following theorem.
Theorem 11.2. Under the same assumptions as Theorem 2.6, $X_n^a(1) \, {\stackrel {d}{\to } \,} X_{(\alpha )}(1)$ under the probability measure $\nu $ .
Proof. We can proceed as for Theorem 11.1 to check the assumptions of [Reference Tyran-KamińskaTK10b, Theorem 1.3] for the skew-product system $(\Omega \times [0,1], F, \nu )$ and the observable $\phi _{x_0}$ .
Acknowledgements
R.A. was partially supported by FCT project PTDC/MAT-PUR/ 4048/2021, with national funds, and by CMUP, which is financed by national funds through FCT – Fundação para a Ciência e a Tecnologia, I.P., under the project with reference UIDB/00144/2020. M.N. was supported in part by NSF Grants DMS 1600780 and DMS 2009923. A.T. was supported in part by NSF Grant DMS 1816315. R.A. would like to thank Jorge Freitas for several very insightful discussions about return time statistics and point processes. We wish to thank an anonymous referee for helpful comments. Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
A. Appendix
The observation that our distributional limit theorems hold for any measures $\mu \ll \nu ^{\omega }$ follows from Zweimüller’s work [Reference ZweimüllerZwe07, Theorem 1, Corollary 1, and Corollary 3].
Let
and suppose
where Y is a Lévy random variable.
We consider first the set-up of intermittent maps. We will show that for any measure $\nu $ with density h that is $d\nu =h\,dm$ in the cone L, in particular Lebesgue measure m with $h=1$ ,
We focus on m. According to [Reference ZweimüllerZwe07, Theorem 1], it is enough to show that
for any $\psi : {\mathbb R} \rightarrow {\mathbb R}$ which is bounded and uniformly Lipschitz.
Fix such a $\psi $ and consider
Since $\|P_{\omega }^k (h_{\omega } -1)\|_{L^{1}_m}\to 0$ in case of Example 2.2 and maps satisfying conditions (LY), (Dec), and (Min), the assertion is proved. By [Reference ZweimüllerZwe07, Corollary 3], the proof for continuous time distributional limits follows immediately.