1. Introduction
Stochastic lattice systems can be used to model many practical systems with discrete character and random fluctuation. The long-time dynamics for stochastic lattice systems with or without delays have been investigated extensively in the literature. For stochastic lattice systems without delays, we refer the reader to [Reference Bates, Lu and Wang2, Reference Bessaih, Garrido-Atienza, Han and Schmalfuss3, Reference Caraballo, Morillas and Valero9, Reference Han, Shen and Zhou20] for pathwise random attractors and stability, [Reference Wang31, Reference Wang and Wang32, Reference Wang34, Reference Wang, Kloeden and Han36] for weak mean random attractors and invariant measures. Since the current states of the practical systems often depend on their past history, stochastic lattice systems with delays have been investigated; see e.g., [Reference Chen, Li and Wang12, Reference Chen and Wang15, Reference Chen, Wang and Yang16, Reference Lin and Li24] for invariant measures and weak mean random attractors, and [Reference Li, Wang and Wang23] for periodic measures. Recently, regime-switching was taken account into stochastic lattice systems, and invariant measures of such systems were studied in [Reference Chen, Sun and Yang13, Reference Li, Lin and Pu22].
In this paper we consider the following stochastic Schrödinger delay lattice system defined on the integer set $\mathbb {Z}$:
where $n\in \mathbb {Z}$; $\varepsilon \in (0,\,1)$; $\lambda$ and $\rho$ are positive constants; $g=(g_n)_{n\in \mathbb {Z}}$ and $h_{k}=(h_{k,n})_{n\in \mathbb {Z}}$ are deterministic complex-valued sequences for each $k \in \mathbb {N}$; $f_n$ and $\sigma _{k,n}$ are locally Lipschitz continuous functions for every $k\in \mathbb {N}$ and $n \in \mathbb {Z}$, and $\{W_k\}_{k\in \mathbb {N}}$ are independent two-sided real-valued standard Wiener processes on a complete filtered probability space $(\Omega,\,\mathcal {F},\,\{\mathcal {F}_{t}\}_{t\geq 0},\,\mathbb {P})$.
The first goal of this paper is to investigate the existence of invariant measures of the stochastic Schrödinger delay lattice system (1.1) in $l^2\times L^2((-\rho,\,0);l^2)$. To that end, we need to establish the tightness of a family of distributions of the solution and its segment process of (1.1) in $l^2\times L^2((-\rho,\,0);l^2)$. Actually, such tightness can be obtained by proving the uniform tail-estimates, the uniform estimates of higher-order moments and the Hölder continuity of the solution as in [Reference Chen and Wang15]. Note that the derivation of uniform estimates of higher-order moments requires not only sophisticated calculations, but also strong dissipativeness assumptions on the nonlinear terms. In order to relax the strong dissipativeness restrictions and prove the existence of invariant measures under weaker conditions on the nonlinear terms, in the present paper, we will employ the equicontinuity of the segment of the solution in probability, instead of uniform estimates of higher-order moments, to establish the tightness of distributions of the solution and its segment process. The idea of equicontinuity in probability was used for proving the tightness of the segment of the solution in [Reference Billingsley4, Reference Wang, Wu and Mao37] for finite-dimensional stochastic ordinary differential equations and in [Reference Chen and Wang14] for fractional stochastic partial differential equations. In the present paper, we will use this method to deal with the infinite-dimensional lattice system (1.1).
The second goal of the paper is to investigate the large deviation principle (LDP) of the solutions of (1.1) on a finite interval $[0,\,T]$ with $T>0$ by the weak convergence method. The weak convergence method is based on the variational representation of certain functionals of Brownian motion [Reference Boué and Dupuis5, Reference Budhiraja and Dupuis7, Reference Budhiraja, Dupuis and Maroulas8] as well as the equivalence of large deviation principles and Laplace principles. Compared with the classical discretization method as introduced in [Reference Freidlin and Wentzell19], the weak convergence method does not require any exponential-type probability estimates which are usually difficult to derive for infinite-dimensional models. The weak convergence method has been successfully applied to establish the LDP for many infinite-dimensional stochastic systems, see e.g. [Reference Budhiraja, Chen and Dupuis6, Reference Budhiraja, Dupuis and Maroulas8, Reference Cerrai and Debussche10, Reference Chen and Gao11, Reference Liu25, Reference Röckner, Zhang and Zhang28, Reference Sritharan and Sundar29, Reference Wang and Duan35] for stochastic partial differential equations, and [Reference Wang33] for stochastic reaction-diffusion lattice systems without delay. We refer the reader to [Reference Freidlin and Wentzell19] and [Reference Dupuis and Ellis18] for more details on the discretization method and the weak convergence method for LDPs, respectively.
Note that the LDPs of finite-dimensional stochastic delay differential equations have been studied by many authors, see e.g. [Reference Bao and Yuan1, Reference Jin, Chen and Zhou21, Reference Mohammed and Zhang27] for constant delay and [Reference Lipshutz26, Reference Suo and Yuan30] for general delay. However, to the best of our knowledge, there is no result available regarding the LDPs of infinite-dimensional delay lattice systems. We will close this gap and prove the LDP for the infinite-dimensional delay lattice system (1.1) in the last section of the paper. Compared with the finite-dimensional stochastic delay differential equations [Reference Jin, Chen and Zhou21] the main difficulty to verify the conditions of the weak convergence for system (1.1) lies in the fact that bounded subsets of $\ell ^2$ are not precompact. To deal with this issue, we adopt the idea of the finite-dimensional projection and the uniform tail-ends estimates to establish the precompactness of a family of solutions to the controlled system (4.5). The argument of the present paper can be extended to the path-dependent lattice systems driven by superlinear noise under certain conditions.
The paper is organized as follows. In Section 2, we discuss the assumptions on the nonlinear terms and present our main results. In the last two sections, we prove the existence of invariant measures and the LDP of (1.1), respectively.
For convenience, we will use $L^2(I;H)$ to denote the space of all square-integrable functions from an interval $I$ to a separable Hilbert space $H$ equipped with norm $\|\cdot \|_{L^2(I;H)}$. We also use $C(I;H)$ for the space of all continuous functions from $I$ to $H$ equipped with supremum norm $\|\cdot \|_{C(I;H)}$. As usual, we reserve $l^2$ for the space of all complex-valued square-summable sequences with inner product $(\cdot,\,\cdot )$ and norm $\|\cdot \|$, respectively.
2. Assumptions and main results
In this section, we discuss the assumptions on the nonlinear terms in (1.1), and present the main results of the paper. First, we define the linear operators $A,\,B,\,B^{\ast }:l^2 \rightarrow l^2$ by:
for any $n\in \mathbb {Z}$ and $u=(u_n)_{n\in \mathbb {Z}} \in l^2$. Then we have
Throughout the paper we make the following assumptions.
(A1) For any bounded subset $\mathcal {K}$ of $\mathbb {C}$, there exists a positive constant $L_{\mathcal {K}}$ such that
\[ |f_n(z_1)-f_n(z_2)| \leq L_{\mathcal{K}}|z_1-z_2|,\]for any $z_1,\,z_2\in \mathcal {K}$ and $n\in \mathbb {Z}$.(A2) For every $k\in \mathbb {N}$, $n\in \mathbb {Z}$ and every bounded subset $\mathcal {K}$ of $\mathbb {C}$, there exists a positive constant $L_{k,n,\mathcal {K}}$ such that for any $z_1,\,z_2\in \mathcal {K}$,
\[ |\sigma_{k,n}(z_1)-\sigma_{k,n}(z_2)| \leq L_{k,n,\mathcal{K}}|z_1-z_2|,\]where $L_{\mathcal {K}}=(L_{k,n,\mathcal {K}})_{k \in \mathbb {N},\,n \in \mathbb {Z}} \in l^2$.(A3) For any $n\in \mathbb {Z}$, there exist positive constants $\alpha _{n}$ and $\beta _0$ such that
\[ |f_n(z)|\leq \beta_0 |z|+\alpha_{n},\quad \forall\ z\in \mathbb{C},\]where $\|\alpha \|^2:=\sum _{n\in \mathbb {Z}}|\alpha _n|^2<\infty$.(A4) For every $k\in \mathbb {N}$, $n\in \mathbb {Z}$, there exist positive constants $\delta _{k,n}$ and $\beta _k$ such that
\[ |\sigma_{k,n}(z)|\leq \delta_{k,n}+\beta_k|z|, \quad \forall\ z \in \mathbb{C},\]where $\|\delta \|^2:=\sum _{k\in \mathbb {N}}\sum _{n \in \mathbb {Z}}|\delta _{k,n}|^2<\infty,\,\ \|\beta \|^2:=\sum _{k\in \mathbb {N}}|\beta _k|^2<\infty$.(A5)
(2.1)\begin{equation} \|g\|^2:=\sum_{n \in \mathbb{Z}}|g_n|^2 < \infty, \quad \|h\|^2:=\sum_{k \in \mathbb{N}}\sum_{n \in \mathbb{Z}}|h_{k,n}|^2<\infty. \end{equation}Consider operators $f,\,\sigma _k :l^2 \rightarrow l^2$ defined by
\[ f(u)=(f_n(u_n))_{n\in \mathbb{Z}}, \quad \sigma_k(u)=(\sigma_{k,n}(u_n))_{n\in \mathbb{Z}}, \quad \forall\ u=(u_n)_{n\in \mathbb{Z}}\in l^2.\]
Then by assumptions (A1)–(A4), we have:
(i) $f$ is well-defined, and
(2.2)\begin{equation} \|f(u)\|^2 \leq 2\beta_0^2 \|u\|^2+2\|\alpha\|^2, \quad \forall\ u\in l^2. \end{equation}(ii) $f$ is locally Lipschitz continuous; that is, for every $R>0$, there exists a positive constant $L^{f}_{R}$ such that for all $u,\,v \in l^2$ with $\|u\|\vee \|v\|\leq R$,
(2.3)\begin{equation} \|f(u)-f(v)\|^2 \leq L^{f}_{R}\|u-v\|^2. \end{equation}(iii) $\sigma _k$ is well-defined and
(2.4)\begin{equation} \sum_{k\in \mathbb{N}}\|\sigma_k(u)\|^2 \leq 2\|\beta\|^2\|u\|^2+2\|\delta\|^2, \quad \forall\ u\in l^2. \end{equation}(iv) $\sigma _k$ is locally Lipschitz continuous; more precisely, for every $R>0$, there exists a positive constant $L^{\sigma }_{R}$ such that for all $u,\,v \in l^2$ with $\|u\|\vee \|v\|\leq R$,
(2.5)\begin{equation} \sum_{k\in \mathbb{N}}\|\sigma_k(u)-\sigma_k(v)\|^2 \leq L^{\sigma}_{R}\|u-v\|^2. \end{equation}
With the above notation, problem (1.1) can be rewritten as the following form in $l^2$:
where $u^{0}=(u_n^{0})_{n\in \mathbb {Z}}$, $|u^{\varepsilon }(t)|^2u^{\varepsilon }(t) =(|u_{n}^{\varepsilon }(t)|^2u_{n}^{\varepsilon }(t))_{n \in \mathbb {Z}}$, $g=(g_{n})_{n\in \mathbb {Z}}$, $h_{k}=(h_{k,n})_{n\in \mathbb {Z}}$ and $\xi =(\xi _{n})_{n \in \mathbb {Z}}$.
From now on, we denote the segment of $u^{\varepsilon }$ by $u^{\varepsilon }_{t}$ which is defined by
Under conditions (A1)–(A5), for every $u^0 \in L^2(\Omega,\,\mathcal {F}_0;l^2)$ and $\xi \in L^2(\Omega,\, \mathcal {F}_0; L^2 ((-\rho,\,0);l^2))$, system (2.6) admits a unique solution $u^\varepsilon$ (see [Reference Chen and Wang15, Theorem 2.2]) in the sense that $u^\varepsilon (t)$, $t\ge -\rho$, is an $l^2$-valued stochastic process such that
• $u^{\varepsilon }(t)$ for $t\ge 0$ is pathwise continuous and $\mathcal {F}_t$-adapted.
• $u^{\varepsilon } (0) =u^0$, $u^{\varepsilon }_0 =\xi$ and $u^{\varepsilon } \in L^2(\Omega ; C([0,\,T];l^2) )$ for all $T>0$.
• For $t\geq 0$, $\mathbb {P}$-almost surely,
\begin{align*} u^{\varepsilon}(t) & = u^{0} + \int_{0}^{t} \left({-}iAu^{\varepsilon}(s) - i|u^{\varepsilon}(s)|^2 u^{\varepsilon}(s) - \lambda u^{\varepsilon}(s) + f(u^{\varepsilon}(s-\rho)) + g \right) \,{\rm d}s \\ & \quad + \sqrt{\varepsilon} \sum\limits_{k \in \mathbb{N}} \int_{0}^{t} \left( h_{k} + \sigma_{k}(u^{\varepsilon}(s-\rho)) \right) {\rm d}W_k(s) \ \ \ {\rm in} \ l^2. \end{align*}
Moreover, one can verify that for every $T>0$,
where $M_0$ is a positive constant independent of $u^{0}$, $\xi$ and $T$.
Based on the well-posedness of solutions of (2.6), we will prove the existence of invariant measures. For this purpose, we need an additional assumption as follows: (H) $\sqrt {2}\beta _0+2\|\beta \|^2< \lambda$.
Theorem 2.1 Suppose that (A1)–(A5) and (H) hold. Then (2.6) has an invariant measure on $l^2 \times L^{2}((-\rho,\,0);l^2)$.
Remark 2.2 Compared with [Reference Chen and Wang15, Theorem 4.1], the conditions on the nonlinear drift and the nonlinear diffusion terms are relaxed due to the fact that the uniform estimates of higher-order moments of solutions are not required in this paper.
Given $(u^{0},\,\xi ) \in l^2 \times L^2((-\rho,\,0);l^2)$ and a positive constant $T$, we will prove the LDP for the family of solutions $\{u^{\varepsilon }\}$ of (2.6) on the finite time interval $[0,\,T]$ as $\varepsilon \rightarrow 0$, which is given below.
Theorem 2.3 Suppose that (A1)–(A5) hold. Then the family of solutions $\{u^{\varepsilon }\}$ of system (2.6) on $[0,\,T]$, as $\varepsilon \rightarrow 0$, satisfies the large deviation principle on $C([0,\,T];l^2)$ with the good rate function $I:C([0,\,T];l^2)\rightarrow [0,\,\infty ]$ defined by (4.1).
Remark 2.4 We point out that theorem 2.1 and theorem 2.3 still hold with minor changes in the proofs if we replace the cubic term $i|u_n|^2u_n$ in (1.1) by a more general nonlinear term $\pm iF(|u_n|)u_n$, where $F:[0,\,\infty ]\rightarrow \mathbb {R}$ is continuous, $F(0)=0$, and there exist $L_F>0$ and $v\geq 0$ such that
3. Invariant measures
In this section, we prove the existence of invariant measures of (2.6). To that end, we need to derive the uniform estimates of the solution as well as its segment process in the next subsection.
3.1. Uniform estimates
In this subsection, we firstly establish the uniform estimates of the solution of (2.6). Note that by (H), there exist constants $\alpha _1>0$ and $\gamma >0$ such that
Lemma 3.1 Suppose that (A1)–(A5) and (H) hold. Then for any $(u^0,\,\xi ) \in L^2(\Omega ;l^2)\times L^2(\Omega ;L^2((-\rho,\,0);l^2))$, the solution $u^{\varepsilon }$ of (2.6) satisfies that for all $t\geq 0$ and $0<\varepsilon < 1$,
where $M_1$ is a positive constant independent of $u^{0}$, $\xi$ and $\varepsilon$.
Proof. Applying Itô's formula to (2.6), we obtain for all $t\geq 0$ and $0<\varepsilon < 1$,
Let $\gamma >0$ be the positive constant satisfying (3.1). Then we get from (3.3) that
We now deal with the right-hand side of (3.4). For the third term on the right-hand side of (3.4), by Young's inequality and (2.2) we have
Let $\alpha _1>0$ be a constant satisfying (3.1). By Young's inequality we get
For the last term on the right-hand side of (3.4), by (2.4) we obtain
It follows from (3.4)–(3.7) that for all $t\geq 0$,
which along with (3.1) indicates that for all $t\geq 0$,
This implies (3.2), and thus completes the proof.
Next we give the uniform estimates of the solution in probability.
Lemma 3.2 Suppose that (A1)–(A5) and (H) hold. If $(u^0\!,\,\xi )\! \in L^2(\Omega ;l^2\!) \!\times \!L^2\!(\Omega ;\! L^2\!((-\rho,\,0);l^2\!))$ satisfies that $\mathbb {E}[\|u^{0}\|^2] \vee \int _{-\rho }^{0}\mathbb {E}[\|\xi (s)\|^2]\,{\rm d}s \leq R$ for some $R>0$, then for any $T>0$ and $\varepsilon '>0$, there exists a positive constant $M_2=M_2(\varepsilon ',\,T,\,R)$, independent of $\varepsilon \in (0,\,1)$, such that
Proof. For any $t\geq \rho$ and $m \in \mathbb {N}$, let
and we set $\tau _{m}^{t}=\infty$ if $\{s\geq t:\|u^{\varepsilon }(s)\|> m\}=\emptyset$.
For any $T> 0$, applying Itô's formula to (2.6), we have
For the third term on the right-hand side of (3.8), by (2.2) we obtain
For the fourth term on the right-hand side of (3.8), by Young's inequality we have
For the last term on the right-hand side of (3.8), by (2.4) we get
From (3.8)–(3.11) and lemma 3.1, it follows that there exists a positive constant $C_{T,R}$ depending only on $R$ and $T$ such that for all $t\geq \rho$,
Recalling the definition of $\tau _{m}^{t}$, we obtain by (3.12) that for all $t\geq \rho$,
Then, we have
By (3.13) we find that for every $\varepsilon '>0$, $T>0$ and $R>0$, there exists $m_1=m_1(\varepsilon ',\,T,\,R)>0$ such that for $m\geq m_1$,
which implies that
On the other hand, making use of (2.7) and the Chebyshev inequality, we obtain that there exists $m_2=m_2(\varepsilon ',\,R)>0$ such that
which along with (3.14) implies that there exists $M_2=M_2(\varepsilon ',\,T,\,R)>0$ such that
as desired.
By lemma 3.2, we have the uniform estimates of the segment of the solution in probability as follows.
Remark 3.3 If $T=2\rho$ in lemma 3.2, then we obtain
Moreover, if $(u^0,\,\xi ) \in L^2(\Omega ;l^2)\times L^2(\Omega ;C([-\rho,\,0];l^2))$, from the proof of lemma 3.2, we can proceed to obtain that for any $T>0$ and $\varepsilon '>0$, there exists a positive constant $M_2=M_2(\varepsilon ',\,T,\,R)$, independent of $\varepsilon \in (0,\,1)$, such that
when $\mathbb {E}[\|u^{0}\|^2] \vee \mathbb {E}[\sup _{s\in [-\rho,0]}\|\xi (s)\|^2] \leq R$ for some $R>0$.
Lemma 3.4 Suppose (A1)–(A5) and (H) hold. If $(u^0\!,\,\xi )\! \in L^2(\Omega ;l^2) \times L^2(\Omega ; L^2 ((-\rho,\,0);l^2))$ satisfies that $\mathbb {E}[\|u^{0}\|^2] \vee \int _{-\rho }^{0}\mathbb {E}[\|\xi (s)\|^2]\,{\rm d}s \leq R$ for some $R>0$, then for any $\varepsilon '>0$ and $\delta _1>0$, there exists $\eta =\eta (\varepsilon ',\,\delta _1,\,R) \in (0,\,\rho )$, independent of $\varepsilon \in (0,\,1)$, such that
Proof. For any $\varepsilon '>0$, it follows from remark 3.3 that there exists $m_3=m_3(\varepsilon ',\,R)>0$ such that for any $t \geq \rho$,
For each $r\geq \rho$, define a stopping time $\tau _r$ by
and we set $\tau _r=\infty$ if $\{s\geq r: \|u^{\varepsilon }_s\|_{C([-\rho,0];l^2)}> m_3\}=\emptyset$. By (3.15) we know that
By (2.6) we have for any $\rho \leq r \leq t$,
and hence for any $r\geq \rho$, $0<\eta <\rho$ and $p>1$, we get
By (2.2), (2.4), the Hölder inequality and the Burkholder–Davis–Gundy (BDG) inequality, it follows from (3.17) that for any $r\geq \rho$, $0<\eta < \min \{1,\, \rho \}$,
where $C_p$ is the coefficient of the BDG inequality and $C_{0}$ is a positive constant independent of $\eta$, $r$ and $\varepsilon \in (0,\,1)$. From (3.16) and (3.18), we can derive that for any $\delta _1>0$ and $t \geq 2\rho$,
Let
Then by (3.19) we obtain that for any $t\geq 2\rho$,
On the other hand, since $u^{\varepsilon }\in L^2(\Omega ;C([0,\,2\rho ];l^2))$, we find that there exists a constant $\eta _2=\eta _2(\delta _1,\,\varepsilon ')>0$ such that for any $\rho \leq t \leq 2\rho$,
which along with (3.20) yields that
where $\eta =\eta _1\wedge \eta _2$, as desired.
Remark 3.5 If $(u^0,\,\xi ) \in L^2(\Omega ;l^2) \times L^2(\Omega ;C([-\rho,\,0];l^2))$, from the proof of lemma 3.4, we can further obtain that for any $\varepsilon '>0$ and $\delta _1>0$, there exists $\eta =\eta (\varepsilon ',\,\delta _1,\,R) \in (0,\,\rho )$, independent of $\varepsilon \in (0,\,1)$, such that
when $\mathbb {E}[\|u^{0}\|^2] \vee \mathbb {E}[\sup _{s\in [-\rho,0]}\|\xi (s)\|^2] \leq R$ for some $R>0$.
Lemma 3.6 Suppose that (A1)–(A5) and (H) hold. Then for every compact subset $E$ of $L^2(\Omega ;l^2) \times L^2(\Omega ;L^2((-\rho,\,0);l^2))$ and $\varepsilon '>0$, there exists a positive integer $N_1=N_1(\varepsilon ',\,E)$ such that for all $m\geq N_1$, $\varepsilon \in (0,\,1)$ and $t\geq 0$, the solution $u^{\varepsilon }(t)$ of (2.6) with $(u^{0},\,\xi )\in E$ satisfies
Proof. Hereafter, we denote by $C$ a generic positive constant independent of $E$, $T$ and $\varepsilon '$. Consider a smooth function $\theta :\mathbb {R}\rightarrow [0,\,1]$ satisfying
Fixed $m \in \mathbb {N}$, denote by $\theta _m=(\theta ({n}/{m}))_{n\in \mathbb {Z}}$ and $\theta _m u = (\theta ({n}/{m})u_n)_{n\in \mathbb {Z}}$ for $u=(u_n)_{n\in \mathbb {Z}}\in l^2$. Then by (2.6) we have
Similar to (3.4), by (3.22) we get that for all $t\geq 0$,
By the argument of (4.4)–(4.6) in [Reference Wang31], we have
For the fifth term on the right-hand side of (3.23), by assumption (A3) we obtain
By Young's inequality we have
For the last term on the right-hand side of (3.23), by assumption (A4) we get
It follows from (3.23)–(3.27) that for all $t\geq 0$,
Since $\|h\|^2 \vee \|\delta \|^2 \vee \|\alpha \|^2 \vee \|g\|^2< \infty$, we infer that there exists $m_4=m_4(\varepsilon ')\geq 0$ such that for all $m\geq m_4$,
For any $\varepsilon '>0$, since $E$ is compact in $L^2(\Omega ;l^2)\times L^2(\Omega ;L^2((-\rho,\,0);l^2))$, then it has a finite open cover of balls with radius ${\sqrt {\varepsilon '}}/{2}$, which is denoted by $\left \{B\left( (u^{j},\,\xi ^{j}) ,\, {\sqrt {\varepsilon '}}/{2}\right) \right \}_{j=1}^{l}$. Since $(u^{j},\,\xi ^{j}) \in L^2(\Omega ;l^2) \times L^2(\Omega ;L^2((-\rho,\,0);l^2))$ for $j=1,\,2,\,\cdots,\,l$, there exists $m_5=m_5(\varepsilon ',\,E)\geq m_4$ such that for all $m\geq m_5$ and $j=1,\,2,\,\cdots,\,l$,
which implies for all $m\geq m_5$ and $(u^{0},\,\xi )\in E$,
We therefore obtain that for all $m\geq m_5$ and $(u^{0},\,\xi )\in E$,
and
On the other hand, by lemma 3.1 we know that there exists $m_6=m_6(\varepsilon ',\,E)\geq m_5$ such that for all $m\geq m_6$ and $t\geq 0$,
where $C_1>0$ is a constant depending only on $E$. Substituting (3.29)-(3.32) into (3.28), we obtain for all $m\geq m_6$ and $t\geq 0$,
which implies that for all $m\geq m_6$ and $t\geq 0$,
as desired.
Lemma 3.7 Suppose that (A1)–(A5) and (H) hold. Then for every compact subset $E$ of $L^2(\Omega ;l^2)\times L^2(\Omega ;L^2((-\rho,\,0);l^2))$, the solution $u^{\varepsilon }(t)$ of (2.6) with $(u^{0},\,\xi )\in E$ satisfies
Proof. Let $\theta$ be the smooth cut-off function as given by (3.21). It follows from lemma 3.6 that for every $\varepsilon '>0$, there exists $N_1=N_1(\varepsilon ',\, E)>0$ such that for any $t\geq 0$ and $m \geq N_1$,
Applying Itô's formula to (3.22), we obtain for all $t\geq 0$, $r\in [t,\,t+T]$, $\varepsilon \in (0,\,1)$ and $m\geq N_1$,
For the second term on the right-hand side of (3.34), similar to (3.24), we have
For the second term on the right-hand side of (3.36), by lemma 3.1 we know that there exists $N_2=N_2(\varepsilon ',\,E) \geq N_1$ such that for $m\geq N_2$,
For the third term on the right-hand side of (3.36), by (A3) and (3.33) we have for $m\geq N_2$,
For the fourth term on the right-hand side of (3.36), by (3.33) and Young's inequality we get for $m\geq N_2$,
For the fifth term on the right-hand side of (3.36), by (A4) we have for $m\geq N_2$,
For the last term on the right-hand side of (3.36), by (A4) and the BDG inequality we obtain for $m\geq N_2$,
It follows from (3.36)-(3.41) that
Similar to (3.29)–(3.30), we obtain that there exists $N_3=N_3(\varepsilon ',\,E)\geq N_2$ such that for all $m\geq N_3$,
and
which along with (3.42) implies that for all $t \geq 0$ and $m\geq N_3$,
where $C_T>0$ depends only on T but not on $\varepsilon ',\, m$ or $E$. This completes this proof.
As an immediate consequence of lemma 3.7, we have the following result.
Corollary 3.8 Suppose (A1)–(A5) and (H) hold. If $(u^0\!,\,\xi )\! \in L^2(\Omega ;l^2\!) \!\times \!L^2\!(\Omega ;\! L^2\!((-\rho,\,0);l^2\!))$, then the solution $u^{\varepsilon }$ of (2.6) satisfies that for every $\delta _2>0$ and $T>0$,
Proof. By the Chebyshev inequality, we obtain that
which together with lemma 3.7 completes the proof.
Remark 3.9 If $(u^0,\,\xi ) \in L^2(\Omega ;l^2) \times L^2(\Omega ;C([-\rho,\,0];l^2))$, from the proofs of lemma 3.7 and corollary 3.8, we can further obtain that for every $\delta _2>0$ and $T>0$,
3.2. Existence of invariant measures
3.2.1. Transition semigroup.
In this subsection, we first introduce the transition semigroup of (2.6), and then show the Feller property and the Markov property of the transition semigroup, which will play a crucial role in proving the existence of invariant measures on $l^2 \times L^2((-\rho,\,0);l^2)$.
For any initial time $t_0\geq 0$ and initial data $(u^{0},\,\xi ) \in L^2(\Omega ;l^2) \times L^2(\Omega ;L^2((-\rho,\,0); l^2))$, we know that (2.6) has a unique solution on $[t_0,\,\infty )$, which is denoted by $u^{\varepsilon }(t;t_0,\,u^{0},\,\xi )$. The segment of $u^{\varepsilon }(t;t_0,\,u^{0},\,\xi )$ on $(t-\rho,\,t)$ with $t\geq t_0$ is written as $u^{\varepsilon }_t(t_0,\,u^{0},\,\xi )$; that is,
Then we have $u^{\varepsilon }_t(t_0,\,u^{0},\,\xi ) \in L^2(\Omega ;L^2((-\rho,\,0);l^2))$ for all $t\geq t_0$.
If $\varphi :l^2 \times L^2((-\rho,\,0);l^2) \rightarrow \mathbb {C}$ is a bounded Borel function, then for $0 \leq r \leq t$ and $(u^{0},\,\xi ) \in l^2 \times L^2((-\rho,\,0);l^2)$, we set
The family $\{p^{\varepsilon }_{r,t}\}_{0\leq r \leq t}$ is called the transition semigroup of (2.6), and $p^{\varepsilon }_{0,t}$ is written as $p^{\varepsilon }_{t}$ for simplicity. In particular, for $\Gamma \in \mathcal {B}(l^2 \times L^2((-\rho,\,0);l^2))$, $0 \leq r\leq t$ and $(u^{0},\,\xi ) \in l^2 \times L^2((-\rho,\,0);l^2)$, we set
where $I_{\Gamma }$ is the characteristic function of $\Gamma$. Recall that a probability measure $\mu ^{\varepsilon }$ on $l^2 \times L^2((-\rho,\,0);l^2)$ is called an invariant measure of (2.6), if
for every bounded Borel function $\varphi :l^2 \times L^2((-\rho,\,0);l^2)\rightarrow \mathbb {C}$.
Given $(u^{n},\, \xi ^n),\, (u^{0},\,\xi )\in l^2 \times L^2((-\rho,\,0);l^2)$, $R>0$ and $r_0\ge 0$, define
Next we show the continuity of $(u^{\varepsilon }(t_0;r_0,\,u^{0},\,\xi ),\,u^{\varepsilon }_{t_0}(r_0,\,u^{0},\,\xi ))$ with respect to initial data in $l^2 \times L^2((-\rho,\,0);l^2)$, which is useful for proving the Feller property of $\{p^{\varepsilon }_{r,t}\}_{0\leq r \leq t}$.
Lemma 3.10 Suppose that (A1)–(A5) and (H) hold. If $(u^{n},\, \xi ^n)\rightarrow (u^{0},\,\xi )$ in $l^2 \times L^2((-\rho,\,0);l^2)$, then for every $0\le r_0\le t_0$,
Proof. For simplicity, we write $u^{\varepsilon }(t;r_0,\,u^{n},\,\xi ^n)$ as $u^{n,\varepsilon }(t)$ and $u^{\varepsilon }(t;r_0,\,u^{0},\,\xi )$ as $u^{\varepsilon }(t)$. By (2.6) and Itô's formula, we get for all $r_0 \leq t \leq t_0$,
For the second term on the right-hand side of (3.45), we know that there exists $C_{1,R}>0$ depending only on $R$ such that
For the third term on the right-hand side of (3.45), by (2.3) we have
where $C_{2,R}>0$ depends only on $R$. For the fourth term on the right-hand side of (3.45), by (2.5) we get
where $C_{3,R}>0$ depends only on $R$. It follows from (3.45)–(3.48) that
By (2.5) and the BDG inequality, we obtain that there exists a constant $C_{4,R}>0$ depending only on $R$ such that
By (3.49)–(3.50) we have for all $t \in [r_0,\,t_0]$,
By Gronwall's inequality and (3.51) we get
Since $(u^{n},\, \xi ^n)\rightarrow (u^{0},\,\xi )$ in $l^2 \times L^2((-\rho,\,0);l^2)$, by (2.2), (2.4) and the Vitali convergence theorem, we infer that $\int _{-\rho }^{0}\|f(\xi ^n(s))-f(\xi (s))\|^2\,{\rm d}s \rightarrow 0$ and $\small \sum _{k\in \mathbb {N}}\small \int _{-\rho }^{0}\|\sigma _{k}(\xi ^n(s)) -\sigma _{k}(\xi (s))\|^2\,{\rm d}s \rightarrow 0$. Thus by (3.52) we know that
Since $\|\xi ^{n}-\xi \|_{L^2((-\rho,0);l^2)}\rightarrow 0$, it follows from (3.53) that
as desired.
By lemmas 3.2 and 3.10 and the arguments of [Reference Da Prato and Zabczyk17, pp. 250-252], we can obtain the following properties of $\{p^{\varepsilon }_{r,t}\}_{0\leq r \leq t}$.
Lemma 3.11 Suppose that (A1)–(A5) and (H) hold. Then we have:
(1) $\{p^{\varepsilon }_{r,t}\}_{0 \leq r \leq t}$ is Feller; that is, if $\varphi :l^2 \times L^2((-\rho,\,0);l^2)\rightarrow \mathbb {C}$ is bounded and continuous, then for any $0 \leq r \leq t$, the function $p^{\varepsilon }_{r,t}\varphi :l^2 \times L^2((-\rho,\,0);l^2)\rightarrow \mathbb {C}$ is also bounded and continuous.
(2) $\{p^{\varepsilon }_{r,t}\}_{0 \leq r \leq t}$ is homogeneous; that is, for all $0 \leq r \leq t$,
\[ p^{\varepsilon}(r,(u^{0},\xi);t,\cdot)=p^{\varepsilon}(0,(u^{0},\xi);t-r,\cdot), \quad \forall\ (u^{0},\xi) \in l^2 \times L^2((-\rho,0);l^2).\](3) For any $0 \leq s \leq r \leq t$, the Chapman–Kolmogorov equation holds true:
\[ p^{\varepsilon}(s,(u^{0},\xi);t,\Gamma) =\int_{l^2 \times L^2((-\rho,0);l^2)}p^{\varepsilon}(s,(u^{0},\xi);r,dx)p^{\varepsilon}(r,x;t,\Gamma),\]where $(u^{0},\,\xi ) \in l^2 \times L^2((-\rho,\,0);l^2)$ and $\Gamma \in \mathcal {B}(l^2 \times L^2((-\rho,\,0);l^2))$.
3.2.2. Proof of theorem 2.1.
Now we are in a position to present the proof of theorem 2.1.
Proof. For simplicity, we now write $u^{\varepsilon }(t;0,\,\textbf {0},\,\textbf {0})$ as $u^{\varepsilon }(t)$ and $u^{\varepsilon }_t(0,\,\textbf {0},\,\textbf {0})$ as $u^{\varepsilon }_t$. By remark 3.3 we see that for given $\varepsilon '>0$, there exists $R_1=R_1(\varepsilon ')>0$ such that for all $t\geq 0$,
By remark 3.5, we know that for given $\varepsilon '>0$ and $m \in \mathbb {N}$, there exists $\eta _{m,\varepsilon '}>0$ depending only on $m$ and $\varepsilon '$ such that for all $t\geq 0$,
and thus
It follows from remark 3.9 that for given $\varepsilon '>0$ and $m \in \mathbb {N}$, there exists an integer $n_{m,\varepsilon '}>0$ depending only on $m$ and $\varepsilon '$ such that for all $t \geq 0$,
and hence we obtain for all $t \geq 0$,
Given $\varepsilon '>0$, denote by
and
It follows from (3.54)–(3.56) that for all $t\geq 0$,
By (3.57), (3.59)–(3.60), we know that the set $\{z(0):z \in Z_{\varepsilon '}\}$ is precompact in $l^2$. Moreover, according to the Ascoli–Arzalà theorem and (3.57)-(3.60), one can show that $Z_{\varepsilon '}$ is a precompact subset of $C([-\rho,\,0];l^2)$. Since the embedding $C([-\rho,\,0];l^2)\hookrightarrow L^2((-\rho,\,0);l^2)$ is continuous, $Z_{\varepsilon '}$ is precompact in $L^2((-\rho,\,0);l^2)$. Thus we conclude that $\tilde {Z}_{\varepsilon '}=\{(z(0),\,z):z \in Z_{\varepsilon '}\}$ is precompact in $l^2 \times L^2((-\rho,\,0);l^2)$.
On the other hand, by (3.61) we obtain that for all $t\geq 0$,
which along with the precompactness of $\tilde {Z}_{\varepsilon '}$ implies that the distributions of the family $\left \{(u^{\varepsilon }(t),\,u^{\varepsilon }_{t}):t\geq 0\right\}$ are tight on $l^2 \times L^2((-\rho,\,0);l^2)$.
We denote the distributions of the family $\left \{(u^{\varepsilon }(t),\,u^{\varepsilon }_{t}):t\geq 0\right \}$ by $\{L^{\varepsilon }_{t}\}_{t\geq 0}$ for simplicity. For given $k \in \mathbb {N}$, we set
By (3.61), we know that for all $k \in \mathbb {N}$,
Consequently, it follows from (3.62)–(3.63) that $\{\mu ^{\varepsilon }_k\}_{k=1}^{\infty }$ is tight, and hence there exists a probability measure $\mu ^{\varepsilon }$ on $l^2 \times L^2((-\rho,\,0);l^2)$ such that, up to a subsequence, $\mu ^{\varepsilon }_k$ weakly converges to $\mu ^{\varepsilon }$ as $k\rightarrow \infty$. Then by lemma 3.11, one can verify that $\mu ^{\varepsilon }$ is an invariant measure of (2.6) by the argument of [Reference Chen, Li and Wang12, Theorem 4.3].
4. The large deviation principle
In this section, we will investigate the LDP of the family $\{u^{\varepsilon }\}_{\varepsilon >0}$ by the weak convergence method. We first review the basic concepts of weak convergence theory in the next subsection.
4.1. Preliminaries
In this subsection, we recall some definitions and results from the theory of large deviations. Let $\mathcal {E}$ be a polish space, and $\{X^{\varepsilon }\}$ be a family of random variables defined on the space $(\Omega,\,\mathcal {F},\,\{\mathcal {F}_{t}\}_{t\geq 0},\,\mathbb {P})$ and taking values in $\mathcal {E}$.
Definition 4.1 A function $I: \mathcal {E}\rightarrow [0,\,\infty ]$ is called to be a rate function, if it is lower semicontinuous on $\mathcal {E}$. A rate function $I$ is called a good rate function, if for each $a \in [0,\,\infty )$, the level set $\{x \in \mathcal {E}:I(x)\leq a \}$ is a compact subset of $\mathcal {E}$.
Definition 4.2 Let $I$ be a rate function on $\mathcal {E}$. The family $\{X^{\varepsilon }\}$ is said to satisfy the LDP on $\mathcal {E}$ with rate function $I$ if the following two conditions hold:
(1) Large deviation upper bound. For each closed subset $F$ of $\mathcal {E}$,
\[ \limsup_{\varepsilon\rightarrow 0}\varepsilon \log \mathbb{P}(X^{\varepsilon} \in F) \leq{-}\inf_{x \in F}I(x),\](2) Large deviation lower bound. For each open subset $G$ of $\mathcal {E}$,
\[ \liminf_{\varepsilon\rightarrow 0}\varepsilon \log \mathbb{P}(X^{\varepsilon} \in G) \geq{-}\inf_{x \in G}I(x).\]
Definition 4.3 Let I be a rate function on $\mathcal {E}$. The family $\{X^{\varepsilon }\}$ is said to satisfy the Laplace principle on $\mathcal {E}$ with rate function I if for all bounded continuous functions $h:\mathcal {E}\rightarrow \mathbb {R}$,
Since $\mathcal {E}$ is a polish space, the family $\{X^{\varepsilon }\}$ satisfies the large deviation principle on $\mathcal {E}$ with a rate function $I$ if and only if the family $\{X^{\varepsilon }\}$ satisfies the Laplace principle on $\mathcal {E}$ with the same rate function. In view of this equivalent result, we will focus on the Laplace principle hereafter. In what follows, we introduce some notations and a criteria for the Laplace principle, which is useful for proving the Laplace principle for the family of solutions $\{u^{\varepsilon }\}$ of (2.6) on $t\in [0,\,T]$. Let
For every $k \in \mathbb {N}$, let $e_k=(\delta _{k,j})_{j=1}^{\infty }$ with $\delta _{k,j}=1$ for $j=k$ and $\delta _{k,j}=0$ otherwise. Then $\{e_k,\,k\in \mathbb {N}\}$ is an orthonormal basis of $H$. Let $W$ be the cylindrical Wiener process on $H$ (which does not take values in $H$), given by
where the series converges in $L^2(\Omega ;C([0,\,T];U))$ with $U$ being a larger separable Hilbert space such that the embedding $H\hookrightarrow U$ is Hilbert–Schmidt.
For each $a>0$, define
Then $S_{a}$ is a Polish space under the weak topology of $L^2([0,\,T]; H)$. Henceforth, wherever we refer to $S_{a}$, we will consider it endowed with this topology. Let $\mathcal {A}$ denote the class of $H$-valued $\mathcal {F}_t$-predictable processes $v$ which satisfy $\int _{0}^{T}\|v\|^2_{H}\,{\rm d}s<\infty$, $\mathbb {P}$-almost surely, and for each $a \in (0,\,\infty )$, we define
For each $\varepsilon \in (0,\,1)$, let $G^{\varepsilon }:C([0,\,T];U) \rightarrow C([0,\,T];l^2)$ be a measurable map. The following lemma gives sufficient conditions for the Laplace principle to hold for the family $\{G^{\varepsilon }(W)\}$ as $\varepsilon \rightarrow 0$.
Lemma 4.4 [Reference Budhiraja and Dupuis7], theorem 4.4
Suppose that there exists a measurable map $G:C([0,\,T];U) \rightarrow C([0,\,T];l^2)$ such that the following two conditions hold:
(H1) for each $a \in (0,\,\infty )$, the set $\Big \{ G ( \int _{0}^{.} v(s) \,{\rm d}s ) : v \in S_{a} \Big \}$ is a compact subset of $C([0,\,T];l^2)$,
(H2) if $\{v^{\varepsilon }\} \subset \mathcal {A}_{a}$ for some $a>0$, and $v^{\varepsilon }$ converges in distribution to $v$ as $S_a$-valued random variables, then $G^{\varepsilon } ( W + \varepsilon ^{-{1}/{2}} \int _{0}^{.} v^{\varepsilon }(t) \,{\rm d}t )$ converges in distribution to $G ( \int _{0}^{.} v(t) \,{\rm d}t )$.
Then $\{G^{\varepsilon }(W)\}$ satisfies the Laplace principle on $C([0,\,T];l^2)$ with rate function $I:C([0,\,T];l^2) \rightarrow [0,\,\infty ]$ defined by
where we use the usual convention $\inf (\emptyset )=\infty$.
4.2. The LDP for solution processes
This subsection is devoted to formulating the Laplace principle for the family of solutions $\{u^{\varepsilon }\}$ of system (2.6) on the finite time interval $[0,\,T]$ as $\varepsilon \rightarrow 0$, from which we can establish the LDP for the family $\{u^{\varepsilon }\}$. We first specify the maps $G^{\varepsilon }$ and $G$ in the context of system (2.6), and we then use lemma 4.4 to deduce an analogous criterion of the Laplace principle for the family $\{u^{\varepsilon }\}$.
Given $u \in l^2$, define $\sigma (u):H\rightarrow l^2$ by
We find that $\sigma (u)$ is well-defined by (2.1) and (2.4). Moreover, the operator is Hilbert–Schmidt and
where $L(H;l^2)$ denotes the space of bounded linear operators from $H$ to $l^2$ with norm $\|\cdot \|_{L(H;l^2)}$, and $L_2(H;l^2)$ denotes the space of Hilbert–Schmidt operators from $H$ to $l^2$ with norm ${\|\cdot \|}_{L_2(H;l^2)}$. In terms of (4.2), system (2.6) on the finite time interval $[0,\,T]$ can be reformulated as
Given $(u^{0},\,\xi ) \in l^2\times L^{2}((-\rho,\,0),\,l^2)$, $\varepsilon \in (0,\,1)$ and $T>0$, by the existence and uniqueness of solutions of system (2.6), we infer that there exists a Borel measurable map $G^{\varepsilon }:C([0,\,T];U)\rightarrow C([0,\,T];l^2)$ such that $u^{\varepsilon }=G^{\varepsilon }(W)$, $\mathbb {P}$-almost surely.
Moreover, for any $v \in \mathcal {A}_a$ with $a\in (0,\,\infty )$, the Girsanov theorem shows that the stochastic process
is a cylindrical Wiener process with identity covariance operator under the probability $\mathbb {P}_{v}^{\varepsilon }$ as given by
Let $u_{v}^{\varepsilon }=G^{\varepsilon }(\widetilde {W})$. Then $u_{v}^{\varepsilon }$ is the unique solution of (4.3) with $W$ replaced by $\widetilde {W}$, which implies that $u^{\varepsilon }_{v}$ is the unique solution of the following controlled stochastic delay system:
To define the map $G$, we introduce a controlled deterministic delay system associated with (4.3) as follows:
By a solution $u_v$ of (4.5), we mean $u_v$ is a map from $[-\rho,\, T]$ to $l^2$ such that $u_v(t)$ is continuous for $t\in [0,\,T]$, $u_v(0) =u^0$ and $u_v =\xi$ on $(-\rho,\, 0)$.
For any $v \in L^2([0,\,T];H)$ and $(u^{0},\,\xi ) \in l^2 \times L^{2}((-\rho,\,0);l^2)$, we will prove the existence and uniqueness of solutions of (4.5) in lemma 4.7 in Subsection 4.3. As a consequence of lemma 4.7, we will see that the solution of (4.5) is continuous in $C([0,\,T];\ell ^2)$ with respect to the control term $v$ in $L^2([0,\,T];H)$. Hence we can define $G:C([0,\,T];U)\rightarrow C([0,\, T];l^2)$ by
where $u_v$ is the unique solution of (4.5) corresponding to the control term $v$.
By lemma 4.4, we deduce the following result.
Corollary 4.5 If $G^{\varepsilon }$ and $G$ defined in this subsection satisfy conditions (H1) and (H2) presented in lemma 4.4, then the family $\{u^{\varepsilon }\}$ satisfies the Laplace principle on $C([0,\,T];l^2)$ with the rate function $I$ given by (4.1).
In the following, we will prove theorem 2.3 by verifying that $G^{\varepsilon }$ and $G$ defined in this subsection satisfy the conditions (H1) and (H2) in lemma 4.4.
4.3. Proof of theorem 2.3
To prove theorem 2.3, we need the following priori estimates for the solutions of (4.5).
Lemma 4.6 Suppose that (A1)–(A5) hold and $T>0$. If $(u^{0},\,\xi ) \in l^2 \times L^{2}((-\rho,\,0);l^2)$, $v \in L^2([0,\,T];H)$ and $u_{v}$ is a solution of system (4.5), then
where $C_T>0$ is a constant depending only on $T$.
Proof. By (4.5) we obtain that for all $t \in [0,\,T]$,
For the second term on the right-hand side of (4.7), by (2.2) we get
For the last term on the right-hand side of (4.7), by (2.4) we have
By (4.7)–(4.9) and Young's inequality we obtain that for all $t \in [0,\,T]$,
which implies
By (4.10) we have for all $t \in [0,\,T]$,
By Gronwall's inequality, it follows from (4.11) that for all $t \in [0,\,T]$,
which completes the proof.
Based on above priori estimates of solutions, we next prove the well-posedness of system (4.5).
Lemma 4.7 Suppose that (A1)–(A5) hold. Then for every $(u^{0},\,\xi ) \in l^2 \times L^{2}((-\rho,\,0);l^2)$ and $v \in L^2([0,\,T];H)$, system (4.5) has a unique solution $u_v$ in $C([0,\, T];l^2)$. Moreover, if $v_1,\,v_2\in L^2([0,\,T];H)$ with $\|v_1\|_{L^2([0,T];H)} \vee \|v_2\|_{L^2([0,T];H)} \leq R_1$ for some $R_1>0$ and $\|u^{0}\|^2 \vee \int _{-\rho }^{0} \|\xi (s)\|^2\,{\rm d}s \leq R_2$ for some $R_2>0$, then the solutions $u_{v_1}$ and $u_{v_2}$ of (4.5) with initial data $(u^{0},\,\xi )$ satisfy
where $C_1>0$ is a constant depending on $R_1$,$R_2$ and $T$.
Proof. Note that system (4.5) on $[0,\,\rho ]$ is equivalent to the following system without delay:
Let $F(t,\,u)=-iAu-i|u|^2u-\lambda u+f(\xi (t-\rho ))+g+\sigma (\xi (t-\rho ))v(t)$. By (2.2) and (2.4) we find that for every $R>0$, there exists $C_{R}>0$ depending only on $R$ such that for all $t \in [0,\,T]$ and $u \in l^2$ with $\|u\|\leq R$,
and for all $t \in [0,\,T]$ and $u_1,\,u_2 \in l^2$ with $\|u_1\|\vee \|u_2\| \leq R$,
Hence, by (4.14)–(4.15) and lemma 4.6, system (4.13) has a unique solution $u_v$ defined on $[0,\,\rho ]$. Repeating this argument, one can extend the solution $u_v$ to the whole interval $[0,\,T]$.
Next, we are going to prove (4.12). By lemma 4.6, for $v_1,\,v_2\in L^2([0,\,T];H)$ with $\|v_1\|_{L^2([0,T];H)} \vee \|v_2\|_{L^2([0,T];H)} \leq R_1$ for some $R_1>0$ and $\|u^{0}\|^2 \vee \int _{-\rho }^{0}\|\xi (s)\|^2\,{\rm d}s \leq R_2$ for some $R_2>0$, there exists $K=K(R_1,\,R_2,\,T)>0$ such that $\sup _{t\in [0,T]}(\|u_{v_1}(t)\| + \|u_{v_{2}}(t)\|) \leq K$. By (4.5) we get for all $t\in [0,\,T]$,
By (2.3) and Young's inequality we have
where $K_1>0$ depends on $R_1$, $R_2$ and $T$. For the last term on the right-hand side of (4.16), by (2.5) we get that there exists $K_2=K_2(R_1,\,R_2,\,T)>0$ such that
Since $\int _{-\rho }^{0}\|\xi (s)\|^2\,{\rm d}s \leq R_2$ and $\sup _{t\in [0,T]}\|u_{v_{2}}(t)\| \leq K$, we obtain that there exists $K_3=K_3(R_1,\,R_2,\,T)>0$ such that for all $t \in [0,\,T]$,
It follows from (4.16)–(4.19) that
Then by Gronwall's inequality we know that for all $t \in [0,\,T]$,
which completes the proof.
The following lemma shows the continuity of an integral operator.
Lemma 4.8 Suppose that (A1)–(A5) hold. For a fixed $\varphi \in L^{\infty }([0,\,T];l^2)\cap L^{2}((-\rho,\,T);l^2)$, define the operator $\Gamma : L^2([0,\,T];H)\rightarrow C([0,\,T];l^2)$ by
Then $\Gamma$ is continuous from the weak topology of $L^{2}([0,\,T];H)$ to the strong topology of $C([0,\,T];l^2)$.
Proof. Note that the operator $\Gamma : L^2([0,\,T];H)\rightarrow C([0,\,T];l^2)$ is well-defined. In fact, by (2.1) and (2.4) we get for every $v \in L^2([0,\,T];H)$,
which implies that $\Gamma (v)\in C([0,\,T];l^2)$ for all $v \in L^2([0,\,T];H)$. Moreover, by (4.21) we get that $\Gamma : L^2([0,\,T];H)\rightarrow C([0,\,T];l^2)$ is bounded. On the other hand, from (4.2) and (4.20) it is easy to see that $\Gamma : L^2([0,\,T];H)\rightarrow C([0,\,T];l^2)$ is linear. Since the operator $\Gamma : L^2([0,\,T];H)\rightarrow C([0,\,T];l^2)$ is strongly continuous and linear, one can deduce that $\Gamma$ is weakly continuous. Following the argument of [Reference Wang33, Lemma 4.3] with small modification and the Ascoli–Arzelà theorem, one can further show that $\Gamma$ is continuous from the weak topology of $L^{2}([0,\,T];H)$ to the strong topology of $C([0,\,T];l^2)$.
Thanks to this lemma, we can proceed to prove the continuity of $u_v$ in $C([0,\,T];l^2)$ with respect to $v \in L^{2}([0,\,T];H)$ in the weak topology of $L^{2}([0,\,T];H)$, which is crucial to verify condition (H1) for the Laplace principle of $\{u^{\varepsilon }\}$.
Lemma 4.9 Suppose that (A1)–(A5) hold. If $v_n\rightarrow v$ weakly in $L^{2}([0,\,T];H)$, then $u_{v_n}\rightarrow u_{v}$ strongly in $C([0,\,T];l^2)$, where $u_{v_n}$ and $u_{v}$ are solutions of (4.5) corresponding to $v_{n}$ and $v$, respectively.
Proof. Since $v_n\rightarrow v$ weakly in $L^{2}([0,\,T];H)$, there exists a constant $N_1>0$ such that $\|v\|_{L^{2}([0,\,T];H)}$ $\leq N_1$ and $\|v_{n}\|_{L^{2}([0,\,T];H)}\leq N_1$ for all $n \in \mathbb {N}$. Then by lemma 4.6 there exists a constant $N_2= N_2(N_1,\, T,\,u^0,\,\xi )>0$ such that
By (4.5) we have
We set
Since $v_n\rightarrow v$ weakly in $L^{2}([0,\,T];H)$, by lemma 4.8 we obtain
Then by (4.22) and (4.23) one can show $u_{v_n} \to u_{v}$ strongly in $C([0,\,T]; l^2)$. The details are similar to [Reference Wang33] and hence omitted here.
We now prove the map $G$ given by (4.6) fulfills condition (H1) in lemma 4.4.
Lemma 4.10 Suppose that (A1)–(A5) hold. Then for every $a \in (0,\,\infty )$, the set
is a compact subset in $C([0,\,T];l^2)$.
Proof. Let $\{u_{v_{n}}\!\}$ be any sequence in $\Xi _{a}$ and $\{v_{n}\} \subset S_{a} \subset L^2([0,\,T];H)$. Since $S_{a}$ is a polish space under the weak topology of $L^2([0,\,T]; H)$, there exist $v \in S_{a}$ and a subsequence $\{v_{n_{k}}\!\}$ such that $v_{n_{k}}\rightarrow v$ weakly. Then, by lemma 4.9 we know that $u_{v_{n_{k}}}\rightarrow u_{v}$ in $C([0,\,T];l^2)$, which implies that $\Xi _{a}$ is compact in $C([0,\,T];l^2)$.
Next, we derive the uniform estimates for the solutions of (4.4).
Lemma 4.11 Suppose that (A1)–(A5) hold, $v \in \mathcal {A}_a$ for some $a \in (0,\,\infty )$ and $(u^{0},\,\xi )\in l^2 \times L^2((-\rho,\,0);l^2)$ with $\|u^{0}\|^2\vee \int _{-\rho }^{0}\|\xi (s)\|^2\,{\rm d}s\leq R$ for some $R>0$. Let $u_{v}^{\varepsilon }$ be the unique solution of system (4.4) with $v$. Then there exists a constant $C_{2}>0$ depending only on $a$, $R$ and $T$ such that
Proof. Applying Itô's formula to (4.4), we obtain that for all $t \in [0,\,T]$,
By (2.4) and the Hölder inequality we obtain
It follows from (2.4), (4.8), (4.26)–(4.27) and Young's inequality that for all $t \in [0,\,T] \ {\rm and} \ \varepsilon \in (0,\,1),$
For the last term on the right-hand side of (4.28), by the Burkholder inequality we have for all $t \in [0,\,T] \ {\rm and} \ \varepsilon \in (0,\,1)$,
Then (4.25) follows from (4.28)–(4.29) and Gronwall's inequality.
We now prove $G$ and $G^{\varepsilon }$ satisfy condition (H2) in lemma 4.4.
Lemma 4.12 Suppose that (A1)–(A5) hold and $\{v^{\varepsilon }\}\subseteq \mathcal {A}_{a}$ for some $a \in (0,\,\infty )$. If $\{v^{\varepsilon }\}$ converges in distribution to $v$ as $S_a$-valued random variables, then $G^{\varepsilon }(W+{1}/{\sqrt {\varepsilon }}\int _{0}^{.}v^{\varepsilon }(t)\,{\rm d}t)$ converges to $G(\int _{0}^{.}v(t)\,{\rm d}t)$ in distribution.
Proof. Notice that $u_v=G(\int _{0}^{.}v(t)\,{\rm d}t)$ is the solution of (4.5) with the control $v$. Let $u^{\varepsilon }_{v^{\varepsilon }}= G^{\varepsilon }(W + {1}/{\sqrt {\varepsilon }}\int _{0}^{.} v^{\varepsilon }(t)\,{\rm d}t)$. Then $u^{\varepsilon }_{v^{\varepsilon }}$ is the solution of the following system:
In order to show that $u^{\varepsilon }_{v^{\varepsilon }}$ converges to $u_v$ in $C([0,\,T];l^2)$ in distribution, we first establish the convergence of $u^{\varepsilon }_{v^{\varepsilon }}-u_{v^{\varepsilon }}$, where $u_{v^{\varepsilon }}=G(\int _{0}^{.}v^{\varepsilon }(t)\,{\rm d}t)$ is the solution of the following system:
For a given constant $M>0$, we define a stopping time $\tau ^{\varepsilon }$ by
and the infimum of the empty set is taken to be $\infty$. Applying Itô's formula to (4.32) yields that for all $t \in [0,\,T]$,
For fixed $(u^{0},\,\xi )\in l^2\times L^2((-\rho,\,0);l^2)$ and $\{v^{\varepsilon }\}\subset \mathcal {A}_{a}$, by lemma 4.6 there exists a positive constant $C_3=C_3(a,\,u^{0},\,\xi,\,T)$ such that for all $\varepsilon \in (0,\,1)$, $\mathbb {P}$-almost surely,
For the first term on the right-hand side of (4.33), by (4.34) we get
where $C_4>0$ depends on $a,\,u^{0},\,\xi,\,T$ and $M$. For the second term on the right-hand side of (4.33), by (2.3) we get
where $C_5>0$ depends on $a,\,u^{0},\,\xi,\,T$ and $M$. For the third term on the right-hand side of (4.33), by (2.5) we obtain
where $C_6$ depends on $a,\,u^{0},\,\xi,\,T$ and $M$. For the last term on the right-hand side of (4.33), by (2.4) we get
It follows from (4.33)–(4.38) that for all $t \in [0,\,T]$,
By (4.39) and Gronwall's inequality, we obtain that for all $t \in [0,\,T]$,
where $C_7=e^{(C_4+C_5+C_6+1)T+a}$. By Doob's maximal inequality and (4.34), we now estimate the first term on the right-hand side of (4.40),
From (4.40)–(4.41), it follows that
Recalling the definition of $\tau ^{\varepsilon }$, by the Chebyshev inequality and lemma 4.11 we obtain
Hence it follows that
which implies that
Since $\{v^{\varepsilon }\}$ converges in distribution to $v$ as $S_a$-valued random variables, according to Skorokhod's representation theorem, there exist a probability space $(\widetilde {\Omega },\,\widetilde {F},\,\widetilde {P})$, and $S_a$-valued random variables $\{\widetilde {v}^{\varepsilon }\}$ and $\widetilde {v}$ with the same distribution as $\{v^{\varepsilon }\}$ and $v$, respectively, such that $\{\widetilde {v}^{\varepsilon }\}\rightarrow \widetilde {v}$ $\widetilde {P}$-almost surely in $S_a$. By lemma 4.9 we infer that $u_{\widetilde {v}^{\varepsilon }}\rightarrow u_{\widetilde {v}}$ $\widetilde {P}$-almost surely in $C([0,\,T];l^2)$. Then $u_{\widetilde {v}^{\varepsilon }}\rightarrow u_{\widetilde {v}}$ in $C([0,\,T];l^2)$ in distribution, and hence
which together with (4.44) implies the desired result.
By lemma 4.10, lemma 4.12 and corollary 4.5, we see that the family $\{u^{\varepsilon }\}$ satisfies the LDP provided (A1)–(A5) hold. This completes the proof of theorem 2.3.
Acknowledgements
The work is partially supported by the NNSF of China (11471190, 11971260). The authors would like to thank the editor and referees for their very valuable suggestions and comments.