Abstract
Consider N independent stochastic processes \((X_i(t), t\in [0,T])\), \(i=1,\ldots , N\), defined by a stochastic differential equation with random effects where the drift term depends linearly on a random vector \(\Phi _i\) and the diffusion coefficient depends on another linear random effect \(\Psi _i\). For these effects, we consider a joint parametric distribution. We propose and study two approximate likelihoods for estimating the parameters of this joint distribution based on discrete observations of the processes on a fixed time interval. Consistent and \(\sqrt{N}\)-asymptotically Gaussian estimators are obtained when both the number of individuals and the number of observations per individual tend to infinity. The estimation methods are investigated on simulated data and show good performances.
Similar content being viewed by others
References
Berglund M, Sunnaker M, Adiels M, Jirstrand M, Wennberg B (2001) Investigations of a compartmental model for leucine kinetics using non-linear mixed effects models with ordinary and stochastic differential equations. Math Med Biol. https://doi.org/10.1093/imammb/dqr021
Delattre M, Dion C (2017) MsdeParEst: Parametric estimation in mixed-effects stochastic differential equations. https://CRAN.R-project.org/package=MsdeParEst. Accessed 16 Sept 2017
Delattre M, Genon-Catalot V, Larédo C (2017) Parametric inference for discrete observations of diffusion with mixed effects in the drift or in the diffusion coefficient. Stoch Process Appl. https://doi.org/10.1016/j.spa.2017.08.016
Delattre M, Genon-Catalot V, Samson A (2013) Maximum likelihood estimation for stochastic differential equations with random effects. Scand J Stat 40:322–343
Delattre M, Genon-Catalot V, Samson A (2015) Estimation of population parameters in stochastic differential equations with random effects in the diffusion coefficient. ESAIM Probab Stat 19:671–688
Delattre M, Lavielle M (2013) Coupling the SAEM algorithm and the extended Kalman filter for maximum likelihood estimation in mixed-effects diffusion models. Stat Interface 6(4):519–532
Dion C (2016) Nonparametric estimation in a mixed-effect Ornstein–Uhlenbeck model. Metrika 79:919–951
Ditlevsen S, De Gaetano A (2005) Mixed effects in stochastic differential equation models. REVSTAT Stat J 3:137–153
Donnet S, Samson A (2008) Parametric inference for mixed models defined by stochastic differential equations. ESAIM P&S 12:196–218
Forman J, Picchini U (2016) Stochastic differential equation mixed effects models for tumor growth and response to treatment. Preprint arXiv:1607.02633
Grosse Ruse M, Samson A, Ditlevsen S (2017) Multivariate inhomogeneous diffusion models with covariates and mixed effects. Preprint arXiv:1701.08284
Jacod J, Protter P (2012) Discretization of processes, vol 67. Stochastic modelling and applied probability. Springer, Berlin
Kessler M, Lindner A, Sørensen ME (2012) Statistical methods for stochastic differential equations, vol 124. Monographs on statistics and applied probability. Chapman & Hall, London
Leander J, Almquist J, Ahlstrom C, Gabrielsson J, Jirstrand M (2015) Mixed effects modeling using stochastic differential equations: Illustrated by pharmacokinetic data of nicotinic acid in obese zucker rats. AAPS J. https://doi.org/10.1208/s12248-015-9718-8
Nie L (2006) Strong consistency of the maximum likelihood estimator in generalized linear and nonlinear mixed-effects models. Metrika 63:123–143
Nie L (2007) Convergence rate of the MLE in generalized linear and nonlinear mixed-effects models: theory and applications. J Stat Plan Inference 137:1787–1804
Nie L, Yang M (2005) Strong consistency of the MLE in nonlinear mixed-effects models with large cluster size. Sankhya Indian J Stat 67:736–763
Picchini U, De Gaetano A, Ditlevsen S (2010) Stochastic differential mixed-effects models. Scand J Stat 37:67–90
Picchini U, Ditlevsen S (2011) Practicle estimation of high dimensional stochastic differential mixed-effects models. Comput Stat Data Anal 55:1426–1444
Whitaker GA, Golightly A, Boys RJ, Chris S (2017) Bayesian inference for diffusion driven mixed-effects models. Bayesian Anal 12:435–463
Author information
Authors and Affiliations
Corresponding author
Electronic supplementary material
Below is the link to the electronic supplementary material.
Appendices
Appendix
1.1 Proof of Proposition 1
For the likelihood of the i-th vector of observations, we first compute the conditional likelihood given \(\Phi _i=\varphi , \Psi _i=\psi \), and then integrate the result with respect to the joint distribution of \((\Phi _i, \Psi _i)\). We replace the exact likelihood given fixed values \((\varphi , \psi )\), i.e. the likelihood of \((X_i^{\varphi , \psi }(t_j), j=1, \ldots ,n)\) by the Euler scheme likelihood of (8). Setting \(\psi = \gamma ^{-1/2}\), it is given up to constants by:
The unconditional approximate likelihood is obtained integrating with respect to the joint distribution \(\nu _{\vartheta }(d\gamma ,d\varphi )\) of the random effects \((\Gamma _i=\Psi _i^{-2},\Phi _i)\). For this, we first integrate \(L_n(X_i, \gamma ,\varphi )\) with respect to the Gaussian distribution \({{\mathcal {N}}}_d({\varvec{\mu }},\gamma ^{-1}{\varvec{\Omega }})\). Then, we integrate the result w.r.t. the distribution of \(\Gamma _i\). This second integration is only possible on the subset \(E_{i,n}(\vartheta )\) defined in (12).
Assume first that \({\varvec{\Omega }}\) is invertible. Integrating (33) with respect to the distribution \({{\mathcal {N}}}({\varvec{\mu }}, \gamma ^{-1}{\varvec{\Omega }})\) yields the expression:
where
and
Computations using matrices equalities and (5), (6), (10) yield that \(T_{i,n}({{\varvec{\mu }}},{{\varvec{\Omega }}})\) is equal to the expression given in (11), i.e.:
Noting that \( \frac{det({{\varvec{\Sigma }}}_{i,n})}{ det({\varvec{\Omega }})}= (det(I_d+V_{i,n}{\varvec{\Omega }}))^{-1} \), we get
Then, we multiply \(\Lambda _n(X_{i,n},\gamma , {\varvec{\mu }}, {\varvec{\Omega }})\) by the Gamma density \((\lambda ^a /\Gamma (a)) \gamma ^{a-1}\exp {(-\lambda \gamma )}\), and on the set \(E_{i,n}(\vartheta )\) (see 12), we can integrate w.r.t. to \(\gamma \) on \((0,+\infty )\). This gives \({{\mathcal {L}}}_n(X_{i,n}, \vartheta )\).
At this point, we observe that the formula (14) and the set \(E_{i,n}(\vartheta )\) are still well defined for non invertible \({\varvec{\Omega }}\). Consequently, we can consider \({{\mathcal {L}}}_n(X_{i,n}, \vartheta )\) as an approximate likelihood for non invertible \({\varvec{\Omega }}\). \(\square \)
1.2 Proof of Theorem 1
For simplicity of notations, we consider the case \(d=1\) (univariate random effect in the drift, i.e. \(\mu = {\varvec{\mu }}\), \(\omega ^2={\varvec{\Omega }}\)). The case \(d>1\) does not present additional difficulties.
Proof of Lemma 2
We omit the index n but keep the index i for the i-th sample path. We have (see 20):
Therefore, using definition (22) and the notations introduced in (H4) (bounds on the parameter set),
Note that, using (11):
Thus, using (H4) and the fact that \(x+x^2\le 2(1+x^2)\) for all \(x\ge 0\),
By the Hölder inequality, we obtain using (6), (36) and (37):
Now, we apply Lemmas (7) and (5) of Sect. 7.2 and (48) of Sect. 7.3 and this yields inequality (23) of Lemma 2.
For the second inequality, we write:
On \(F_{i}\), \(Z_i\ge c/\sqrt{n}\), so:
Consequently, by the Hölder inequality,
Now, we take conditional expectation w.r.t. \(\Psi _i=\psi , \Phi _i=\varphi \) and apply the Cauchy–Schwarz inequality. We use that, for n large enough, (see (48) of Sect. 7.3):
And, we apply inequality (23) to get (24). \(\square \)
Now, we start proving Theorem 1. Omitting the index n in \(A_{i,n}, B_{i,n}\), we have [see (21), (25), (26)]
The remainder terms are:
The most difficult remainder terms are \(R_1, R_2,R_3,R_4\). They are treated in Lemma 3 below. For the term \(R'_2\), we use that \((\psi (a+(n/2)) - \log {(a+(n/2))})= O(n^{-1})\) (see 47) and that, for \(a>4\), \({{\mathbb {P}}}_{\theta }(F_{i}^c) \lesssim n^{-2}\) (see 17) to get that \(R'_2= O_P(\sqrt{N}/n)\).
Using Lemma 5 in Sect. 7.2, it is easy to check that \(R'_3\) and \(R'_4\) are \(O_P(\sqrt{N/n})\).
Therefore, there remains to find the limiting distribution of:
The first two components are exactly the score function corresponding to the exact observation of \((\Gamma _i, i=1, \ldots , N)\) (see Sect. 7.4). Hence, the first part of Theorem 1 is proved.
The whole vector is ruled by the standard central limit theorem. To compute the limiting distribution, we use results from Delattre et al. (2013) which deals with the case of \(\Gamma _i=\gamma \) fixed and \(\Phi _i \sim {{\mathcal {N}}}(\mu , \gamma ^{-1}\omega ^2)\). It is proved in this paper that
This result is stated for \(\gamma =1\) in Proposition 5, p. 328 of Delattre et al. (2013) with (unfortunately) different notations: \(A_i(T;\mu , \omega ^2)\) is denoted \(\gamma _i(\theta )\) and \(B_i(T, \omega ^2) \) is denoted \(I_i(\omega ^2)\) (formula (11) of this paper). It can be extended for any value of \(\gamma \) using formula (35) and the regularity properties of the statistical model. Hence, the third and fourth component are centered and the covariances between the first two components and the last two ones are null. Moreover, it is also proved (in the same Proposition 5) that
Hence, the covariance matrix of the last two components is equal to \(J(\vartheta )\) defined in (29).
The proof of the last item (second order derivatives) relies on the same tools with more cumbersome computations but no additional difficulty. It is detailed in the electronic supplementary material. Note that this part only requires that N, n both tend to infinity without further constraint. So the proof is complete. \(\square \)
Lemma 3
Recall (20). Then, for \(a>4\), (see (16) for the definition of \(F_{i,n}\)), \(R_1, R_2\) are \(O_P(\max \{1/\sqrt{n},\frac{\sqrt{N}}{n}\})\), \(R_3, R_4\) are \(O_P(\sqrt{\frac{N}{n}})\).
Proof
The proof goes in several steps. We have introduced
We know the exact distribution of \(S_{i}^{(1)}\): \(C_{i}^{(1)}\) is independent of \(\Gamma _i\) and has distribution \(\chi ^2(n)=G(n/2, 1/2)\). By exact computations, using Gamma distributions (see Sect. 7), we obtain:
Thus,
Let us detail the first computation. Note that \((\Gamma _i, i=1,\ldots ,N)\) and \((C_{i}^{(1)}, i=1,\ldots ,N)\) are independent. Thus,
The second computation is similar.
Then, we have to study
We write:
For \(a>4\), \({{\mathbb {P}}}_{\vartheta }(F_{i}^c) \lesssim n^{-2}\) (see 17). We have, applying the Cauchy–Schwarz inequality:
Next, apply Lemma 2,
As noted above, \(\Psi _i=\Gamma _i^{-1/2}\), \({{\mathbb {E}}}_{\vartheta }(\Psi _i^{-q})< +\infty \) for all \(q\ge 0\). We can write \(\Phi _i=\mu +\omega \Psi _i \varepsilon _i\) with \(\varepsilon _i\) a standard Gaussian variable independent of \(\Psi _i\). We have
Thus, for \(a>4\),
Joining (39) (left)–(40) (left) and (41), we obtain that, for \(a>4\), \(R_1=O_P(\max \{n^{-1/2}, \sqrt{N}/n\})\).
Analogously,
As above, we can prove
And:
On \(F_{i}\),
Therefore,
Now, we take conditional expectation w.r.t. \(\Phi _i=\varphi , \Psi _i=\psi \), and apply first the Cauchy–Schwarz inequality and then Lemma 2. This yields:
We have to check that the expectation above is finite. The worst term is \({{\mathbb {E}}}_{\vartheta }\Psi _i^6= {{\mathbb {E}}}_{\vartheta }\Gamma _i^{-3}\) which requires the constraint \(a>3\). Thus, for \(a>3\), we have:
Therefore, we have proved that \(R_1, R_2\) are \(O_P(\max \{\frac{1}{\sqrt{n}}, \frac{\sqrt{N}}{n}\})\).
For \(R_3, R_4\), we proceed analogously but we have to deal with the terms \(A_{i}, A_{i}^2\). We write again:
Using Lemma 6 and the Cauchy–Schwarz inequality, we obtain:
Applying now Lemma 2, we obtain:
This requires the constraint \({{\mathbb {E}}}_{\vartheta }\Psi _i^3 {{\mathbb {E}}}_{\vartheta } \Gamma _i^{-3/2}<+\infty \), i.e. \(a>3/2\). Note that \({{\mathbb {E}}}_{\vartheta }(\Psi _i^{-4p})\) is finite for any p. This is why these moments are omitted.
We proceed analogously for \(R_4\) and find that \(R_4= O_P(\sqrt{N/n})\) for \(a>2\). The different constraint on a derives from the presence of \(A_i^2\) in \(R_4\) which requires higher moments for \(\Gamma _i^{-1}\). \(\square \)
1.3 Proof of Theorem 3
We only give a sketch of the proof and assume \(d=1\) for simplicity. We compute \({{\mathcal {H}}}_{N,n}(\vartheta )\) (see 32) and set \(G_{i,n}=\{S_{i,n}\ge k\sqrt{n}\} \). We have:
We can prove that, under (H1)–(H2), if \(a_0>2\), \({{\mathbb {P}}}_{\vartheta _0} (G_{i,n}^c)\lesssim n^{-2}\) under analogous and simpler tools as in Lemma 1. The result of Lemma 2 holds with \(\xi _{i,n}\) instead of \(Z_{i,n}\) and without \(1_{F_{i,n}}\). This allows to prove that:
where \(r_1\) and \(r_2\) are \(O_P(\sqrt{N}/n)\).
The result of Lemma 2 holds with \(S_{i,n}/n\) instead of \(Z_{i,n}\) and \(G_{i,n}\) instead of \(F_{i,n}\) (and the proof is much simpler). This implies that:
and we can prove that \(r_3\) and \(r_4\) are \(O_P(N/n)\).
Auxiliary results
In Sect. 7.1, we explain why Assumption (H2) is required. Results on discretizations are recalled in Sect. 7.2. Sections 7.3 and 7.4 contain results on Gamma distributions and estimation from direct observations of the random effects.
1.1 Preliminary results for SDEs with random effects
SDEMEs have specific features that differ from usual SDEs. The discussion below justifies our setup and assumptions. Consider \((X(t), t \ge 0)\) a stochastic process ruled by:
where the Wiener process (W(t)) and the r.v.’s \((\Phi ,\Psi )\) are defined on a common probability space \((\Omega , {{\mathcal {F}}}, {{\mathbb {P}}})\) and independent. We set \(({{\mathcal {F}}}_t=\sigma ( \Phi , \Psi , W(s), s\le t), t\ge 0)\). To understand the properties of (X(t)), let us introduce the system of stochastic differential equations:
Existence and uniqueness of strong solutions is therefore ensured by the classical assumptions:
-
(A1)
The real valued functions \((x, \varphi )\in {{\mathbb {R}}}\times {{\mathbb {R}}}^{d} \rightarrow b(x, \varphi )\) and \((x, \psi )\in {{\mathbb {R}}}\times {{\mathbb {R}}}^{d'} \rightarrow \sigma (x, \psi )\) are \(C^1\).
-
(A2)
There exists a constant K such that, \(\forall x, \varphi , \psi \in {{\mathbb {R}}}\times {{\mathbb {R}}}^{d}\times {{\mathbb {R}}}^{d'}\), (\(\Vert .\Vert \) is the Euclidian norm):
$$\begin{aligned} |b(x, \varphi )| \le K(1+|x| +\Vert \varphi \Vert ), \quad |\sigma (x, \psi )| \le K(1+|x| +\Vert \psi \Vert ) \end{aligned}$$If the following additional assumption holds:
-
(A3)
The r.v.’s \(\Phi ,\Psi \) satisfy \({{\mathbb {E}}}(\Vert \Phi \Vert ^{2p} + \Vert \Psi \Vert ^{2p})<+\infty \)
then, under (A1)–(A2), for all T, \(\sup _{t\le T}{{\mathbb {E}}}X^{2p}(t)<+\infty \).
To deal with discrete observations, moment properties are required to control error terms. Here, we consider a simple model for the drift term, \(b(x,\varphi )= \varphi ' b(x)\), and for the diffusion coefficient, \(\sigma (x,\psi )= \psi \sigma (x)\), with \(\varphi \in {{\mathbb {R}}}^{d}\), \(\psi \in {{\mathbb {R}}}\). The common assumptions on b and \( \sigma \) should be that \(b, \sigma \) are \(C^1({{\mathbb {R}}})\) and have linear growth w.r.t. x. Then, for all fixed \((\varphi , \psi ) \in {{\mathbb {R}}}^d\times (0,+\infty )\), the stochastic differential equation
admits a unique strong solution process \((X^{\varphi ,\psi }(t), t\ge 0)\) adapted to the filtration \(({{\mathcal {F}}}_t)\). Moreover, the stochastic differential equation with random effects (1) admits a unique strong solution adapted to \(({{\mathcal {F}}}_t)\) such that the joint process \((\Phi ,\Psi , X(t), t \ge 0)\) is strong Markov and the conditional distribution of (X(t)) given \(\Phi =\varphi , \Psi =\psi \) is identical to the distribution of (46). We need more than this property. Indeed, moment properties for X(t) are required. Let us stress that, with \(b(x, \varphi )=\varphi 'b(x), \sigma (x,\psi )=\psi \sigma (x)\), (A2) does not hold even if \(b,\sigma \) have linear growth. In particular, (A3) does not ensure that X(t) has moments of order 2p. Let us illustrate this point on an example.
Example
Consider the mixed effect Ornstein–Uhlenbeck process \(dX(t)= \Phi X(t) dt + \Psi dW(t), X(0)=x\). Then, \(\displaystyle X(t)= x \exp {(\Phi t)} + \Psi \exp {(\Phi t)}\smallint _0^t \exp {(-\Phi s)}\; dW(s) \). The first order moment of X(t) is finite iff \({{\mathbb {E}}}\exp {(\Phi t)}<+\infty \) and \({{\mathbb {E}}}\left[ |\Psi |((\exp {(2\Phi t)}\right. \left. -1)/2\Phi )^{1/2}\right] <+\infty \) which is much stronger than the existence of moments for \(\Phi , \Psi \). When \((\Phi , \Psi )\) has distribution (2), \(\displaystyle {{\mathbb {E}}}\exp {(\Phi t)}= \exp {(\mu t)} {{\mathbb {E}}}(\exp {(\omega ^2 t^2/\Gamma )})=+\infty \).
In a more general setup, the existence of moments is standardly proved by using the Gronwall lemma. In this case, it would lead to an even stronger condition such as \( {{\mathbb {E}}}(\exp {(\Phi ^2 t)}+ \exp {(\Psi ^2 t)})<+\infty \).
Hence, stronger assumptions than usual are necessary. Indeed, (A2) holds for model (1) either if \(\varphi \) and \(\psi \) belong to a bounded set, or if b and \(\sigma \) are uniformly bounded.
We could think of using a localisation device to get bounded b and \(\sigma \). The problem here is to deal with N sample paths \((X_i(t)), i=1, \ldots ,N\) which have possibly no moments. The localisation device is here complex.
1.2 Approximation results for discretizations
The following lemmas are proved in Delattre et al. (2017). In the first two lemmas, we set \(X_1(t)=X(t), \Phi _1=\Phi , \Psi _1=\Psi \).
Lemma 4
Under (H1)–(H2), for \(s\le t\) and \(t-s \le 1\), \(p\ge 1\),
For \(t\rightarrow H(t,X.)\) a predictable process, let \(V(H;T)= \int _0^T H(s,X.)ds\) and \(U(H;T)=\int _0^T H(s,X.)dX(s)\). The following results can be standardly proved.
Lemma 5
Assume (H1)–(H2) and \(p\ge 1\). If H is bounded, \( {{\mathbb {E}}}_{\vartheta }(|U(H;T)|^{p}|\Phi =\varphi , \Psi =\psi ) \lesssim | \varphi |^{p}+\psi ^{p}.\)
Consider \(f:{{\mathbb {R}}} \rightarrow {{\mathbb {R}}}\) and set \(H(s,X.)=f(X(s)), H_n(s,X.)= \sum _{j=1}^n f(X((j-1)\Delta ))1_{((j-1)\Delta ,j\Delta ]}(s)\). If f is Lipschitz,
If f is \(C^2\) with \(f',f''\) bounded
Lemma 6
Recall notations (25)–(26). Under (H1)–(H2),
Let
Lemma 7
Then, for all \(p\ge 1\),
1.3 Properties of the Gamma distribution
The digamma function \(\psi (a)= \Gamma '(a)/\Gamma (a)\) admits the following integral representation: \(\psi (z)= -\gamma +\int _0^1 (1-t^{z-1})/(1-t) dt\) (where \(-\gamma =\psi (1)=\Gamma '(1)\)). For all positive a, we have \( \psi '(a)= -\int _0^1 \frac{\log {t}}{1-t}\;t^{a-1} dt\). Consequently, using an integration by parts, \( -a\psi '(a)=-1- \int _0^1 t^a g(t)dt\), where \(g(t)=(\log {t}/(1-t))' \). A simple study yields that \(t^ag(t)\) is integrable on (0, 1) and positive except at \(t=1\). Thus, \(1-a\psi '(a) \ne 0\). The following asymptotic expansion as a tends to infinity holds:
If X has distribution \(G(a, \lambda )\), then \(\lambda X\) has distribution G(a, 1). For all integer k, \({{\mathbb {E}}}(\lambda X)^k=\frac{\Gamma (a+k)}{\Gamma (a)}\). For \(a>k\), \({{\mathbb {E}}}(\lambda X)^{-k}=\frac{\Gamma (a-k)}{\Gamma (a)}\). Moreover, \({{\mathbb {E}}}\log {(\lambda X)}= \psi (a)\), Var \([\log {(\lambda X)}]= \psi '(a)\).
In particular, if \(X=\sum _{j=1}^n \varepsilon _i^2\) where the \(\varepsilon _i\)’s are i.i.d. \({{\mathcal {N}}}(0,1)\), then \(X\sim \chi ^2(n)=G(n/2,1/2)\). Therefore, \({{\mathbb {E}}}X^{-p}<+\infty \) for \(n>2p\) and as \(n\rightarrow +\infty \),
1.4 Direct observation of the random effects
Assume that a sample \((\Phi _i, \Gamma _i), i=1, \ldots ,N\), is observed and that \(d=1\) for simplicity. The Gamma distribution with parameters \((a, \lambda )\) (\(a>0, \lambda >0\)) \(G(a, \lambda )\), has density \( \gamma _{a,\lambda }(x)=(\lambda ^a/\Gamma (a)) x^{a-1} e^{-\lambda x} \mathbb {1}_{(0, +\infty )}(x), \) where \(\Gamma (a)\) is the Gamma function. We set \(\psi (a)= \Gamma '(a)/\Gamma (a)\). The log-likelihood \(\ell _N(\vartheta )\) of the N-sample \((\Phi _i, \Gamma _i), i=1, \ldots ,N\) has score function \({{\mathcal {S}}}_N(\vartheta )= \left( \frac{\partial }{\partial \lambda } \ell _N(\vartheta ) \;\; \frac{\partial }{\partial a} \ell _N(\vartheta )\;\;\frac{\partial }{\partial \mu } \ell _N(\vartheta ) \;\; \frac{\partial }{\partial \omega ^2} \ell _N(\vartheta ) \right) '\) given by
By standard properties, we have, under \({{\mathbb {P}}}_{\vartheta }\), \(N^{-1/2}{{\mathcal {S}}}_N(\vartheta ) \rightarrow _{{{\mathcal {D}}}}{{\mathcal {N}}}_4(0, {{\mathcal {J}}}_0(\vartheta )),\) where
Using properties of the di-gamma function (\(a\psi '(a)-1\ne 0\)), \(I_0(a,\lambda )\) is invertible for all \((a,\lambda )\in (0,+\infty )^2\). The maximum likelihood estimator based on the observation of \((\Phi _i,\Gamma _i, i=1, \ldots , N)\), denoted \(\vartheta _N=\vartheta _N(\Phi _i,\Gamma _i, i=1, \ldots , N)\), is consistent and satisfies \(\sqrt{N}(\vartheta _N-\vartheta )\rightarrow _{{{\mathcal {D}}}}{{\mathcal {N}}}_4(0, {{\mathcal {J}}}_0^{-1}(\vartheta ))\) under \({{\mathbb {P}}}_{\vartheta }\) as N tends to infinity.
In the simulations presented in Sect. 4, we took \(a=8\) and observed that estimations of a are biased with a large standard deviation. This can be seen on \(I_0^{-1}(\lambda ,a)\):
If a large, \(a/ (a\psi '(a)-1)=O(a^2)\).
However, natural parameters for Gamma distributions are \(m=a/\lambda , t= \psi (a)-\log {\lambda }\) with unbiased estimators \(\hat{m}= N^{-1}\sum _{i=1}^N \Gamma _i\), \(\hat{t}= N^{-1} \sum _{i=1}^N \log {\Gamma _i}\) such that the vector \(\sqrt{N}(\hat{m}-m, \hat{t} -t)\) is asymptotically Gaussian with limiting covariance matrix
The asymptotic variance of \(\hat{t}\) is \(\psi '(a)=O(a^{-1})\) and both parameters (m, t) are well estimated.
Rights and permissions
About this article
Cite this article
Delattre, M., Genon-Catalot, V. & Larédo, C. Approximate maximum likelihood estimation for stochastic differential equations with random effects in the drift and the diffusion. Metrika 81, 953–983 (2018). https://doi.org/10.1007/s00184-018-0666-z
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00184-018-0666-z