1 Introduction

Empirical studies [2, 6, 8, 11] on high frequency financial data of stock markets that employ the continuous double auction method have revealed a time series of signs of market orders has long memory property. In contrast, a time series of stock returns is known to have short memory property. A time series of order signs is defined by changing transactions at the best ask price into + 1 and transactions at the best bid price into − 1. The auto-correlation function of the order signs decays as a power law of the lag and the exponent of the decay is less than 1, which is equivalent to a Hurst exponent of the time series is greater than one-half.

In this paper, we propose a new mathematical model which takes account of origin of the long memory in order signs. As a first step, we define a discrete time stochastic process of cumulative order signs in accordance with some explanation for the origin of the phenomenon. Subsequently, we verify increments of the process has the long memory property. In general, there are three ways to verify a discrete time process has some property. The first one runs computer simulations. The second calculates the distribution of the discrete time process directly. And the third, which we use in this paper, is to show that the scaled discrete time process converges to a continuous process which has that property.

There are various explanations for the origin of the long memory property of order sings [3]. A cogent one, which was proposed by Lillo et al. [9], is that investors tend to strategically split their large hidden orders into small pieces before execution to prevent the increase in the trading costs. Empirical findings partially support this explanation. A long memory phenomenon is found in a time series of order signs of transactions initiated by a single member of the stock market [3, 8]. Investors enter their orders into the market through one of its members.

Assuming the size of hidden orders distributes as a power law, Lillo et al. [9] considered a discrete time mathematical model with this explanation. Under an additional technical assumption that the number of hidden orders is fixed, they showed rigorously the model has a long memory property. However, this technical assumption does not seem natural.

Taking account of the bursty nature of human dynamics [1], Kuroda et al. [7] proposed another theoretical model with this explanation. They assumed that a time interval between order executions distributes as a power law, and that the power law exponent does not depend on the size of hidden order. Under an additional technical assumption that the size of hidden order is bounded above, they showed the scaled discrete time process converges to a superposition of a Brownian motion and a finite number of fractional Brownian motions with Hurst exponents greater than one-half. Moreover, the number of hidden orders is not fixed in their model, and it randomly varies. Although, the maximum Hurst exponent of obtained process depends on the largest hidden order.

The Hurst exponent of order signs expected by the theory of splitting large hidden order is smaller than the value of the empirical study [3]. About stocks with high liquidity, the fluctuation of Hurst exponents of order signs is small by a stock and a period [6]. These findings suggest that there might be some other cause about the long memory of order signs. We can pay attention not only to large hidden orders but also to small hidden orders. Vázquez et al. found two universality class in human dynamics [13]. On the other hand, Zhou et al. observed that in an online movie rating site, a power law exponent of the time interval between user’s postings depends on user’s activity [15].

In this paper, we propose a new mathematical model with an explanation for the origin of long memory of order signs that investors split their hidden order of any size into small pieces before execution. We assumed that the power law exponent of distribution of a time interval between order executions depends on the size of hidden order. We showed the scaled discrete time process converges to a superposition of a Brownian motion and countably infinite number of fractional Brownian motions with Hurst exponents greater than one-half. We note that the number of hidden orders randomly varies, that Hurst exponents are not bounded above and that the maximum Hurst exponent of obtained process depends on hidden order of medium size.

2 Model

In this section, we introduce a probability space (Ω n , P n ), where n is a natural number. In the next section, we will define a discrete time stochastic process in time interval \(\varLambda _{n} =\{ 1,2,\ldots,n\}\) which describes cumulative order signs on the probability space. And we will show the increment of the process has a long memory property. We note that in this paper we only study the order sign and do not consider the stock price.

All essential assumptions our model requires for the market is as follows:

  • Investors tend to split their hidden orders into small pieces before execution.

  • The distribution of a time interval between order executions obeys a power law.

  • The power law exponent of the inter-event distribution depends on the size of the hidden order.

A hidden order of one investor in a stock market and execution times of its small pieces is denoted by p. Namely, p has two quantities: the order sign s(p) = s and the set of times of executions b(p) = { u 1, , u m }, where m ≥ 1 is the number of small pieces of the hidden order split by the investor. We call p a polymer using the terminology of a mathematical method called the cluster expansion, which we will use to prove our main theorem. A method of the cluster expansion is developed in the study of the statistical physics and is applied for instance to convergence theorems of the phase separation line of the two dimensional Ising model [4, 12]. Since the cluster expansion is defined in an abstract setting [5], it can be applied to a financial model [7, 14].

It is known that a time series of trading volume in a stock market exhibits long memory [10]. However, we do not consider the memory of the trading volume in this paper; we suppose the volume of each piece is 1 just for the sake of simplicity, and we emphasize it is not technical assumption. Consequently, the number m of small pieces is equivalent to the size (or the total volume) of the hidden order. For any polymer p, we can also regard m as the amount of activity of a investor in time period of her holding the polymer. Meanwhile, investors often do not split their own orders and submit it in a stock market at once. This situation is also included in our model as m = 1. Although the model proposed by Kuroda et al. [7] assumed that m is bounded above, that is, the maximum value of m is finite, our model does not require any restriction on the maximum value of m. More precisely, m has an upper bound \(\log \log n\), and n tends to infinity in our main theorem.

The order sign s(p) is assigned to + 1 or − 1 according to whether the hidden order is a buy order or a sell order. For any polymer p, its order sign s(p) is a single value. Obviously, different valued order signs are possibly assigned to different polymers possessed by one investor. In our model, the distribution of order signs is symmetry:

$$\displaystyle{ P_{n}\left (s(\mathbf{p}) = +1\right ) = P_{n}\left (s(\mathbf{p}) = -1\right ) = \frac{1} {2}. }$$
(4.1)

Each element of the set of times of executions b(p) = { u 1, , u m } is an integer. Their magnitude relation is given as u 1 < u 2 < ⋯ < u m , that is, the first piece of the hidden order is executed at u 1, and the last one is executed at u m . For any distinct two polymers p 1 and p 2, their execution times do not overlap:

$$\displaystyle{ b(\mathbf{p}_{1}) \cap b(\mathbf{p}_{2}) =\emptyset. }$$
(4.2)

Since we will observe the discrete time stochastic process of cumulative order signs in time interval \(\varLambda _{n}\), it is enough to consider only polymers p which satisfy

$$\displaystyle{ b(\mathbf{p}) \cap \varLambda _{n}\neq \emptyset. }$$
(4.3)

Taking the bursty nature of the human activity patterns into account, we assume that the distribution of a time interval between order executions obeys a power law and that its exponents α(m) depends on the size (or the activity) of the polymer:

$$\displaystyle{ P_{n}\left (u_{i},u_{i-1} \in b(\mathbf{p}),\mbox{ $\mathbf{p}$ is a polymer of size m}\right ) \propto (u_{i} - u_{i-1})^{-\alpha (m)}. }$$
(4.4)

According to an empirical study on human dynamics [15], power law exponents of the inter-event times are increasing in a parameter of activity. Hence, it suggests that the exponent α(m) is increasing in m. Our model does not require that the exponent is increasing, though it requires some condition on the exponent.

In the following, we define our model using the mathematical terminology. Let n be a natural number and \(\varLambda _{n} =\{ 1,2,\ldots,n\}\) be an observation time of a discrete time process. We describe a hidden order and its execution times of small pieces by a polymer.

Let \(\mathcal{P}_{n,1}\) be the set of all polymers corresponds to hidden order of size 1:

$$\displaystyle{ \mathcal{P}_{n,1} = \left \{\mathbf{p} = (s,u)\,;\,s \in \{ +1,-1\},\,u \in \varLambda _{n}\right \}. }$$
(4.5)

For each polymer \(\mathbf{p} = (s,u) \in \mathcal{P}_{n,1}\), we denote the order sign by s(p) = s, the time of execution by b(p) = { u} and the size of hidden order by | p | = 1.

For any m, \((2 \leq m \leq \log \log n)\), we define the set of all polymers corresponds to hidden order of size m by

$$\displaystyle\begin{array}{rcl} \mathcal{P}_{n,m}& =& \left \{\mathbf{p} = (s,u_{1},\ldots,u_{m})\,;\,s \in \{ +1,-1\},\,\{u_{1},\ldots,u_{m}\} \cap \varLambda _{n}\neq \emptyset,\right. {}\\ & & \qquad \qquad \qquad \qquad \quad \left.1 \leq u_{i} - u_{i-1} \leq n\quad (i = 2,\ldots,m)\right \}. {}\\ \end{array}$$

For each polymer \(\mathbf{p} = (s,u_{1},\ldots,u_{m}) \in \mathcal{P}_{n,m}\), we denote the order sign by s(p) = s, the set of times of executions by b(p) = { u 1, , u m } and the size of hidden order by | p | = m. The set of all polymers is denoted by

$$\displaystyle{ \mathcal{P}_{n} =\bigcup _{ m=1}^{\log \log n}\mathcal{P}_{ n,m}. }$$
(4.6)

The configuration space is denoted by

$$\displaystyle{ \varOmega _{n} =\bigcup _{ k=1}^{n}\left \{\omega =\{ \mathbf{p}_{ 1},\ldots,\mathbf{p}_{k}\} \subset \mathcal{P}_{n}\,;\,b(\mathbf{p}_{i}) \cap b(\mathbf{p}_{j}) =\emptyset,\,(1 \leq i <j \leq k)\right \}. }$$
(4.7)

Example 1

We consider the case that n = 10 and ω = {p 1, p 2, p 3, p 4} ∈ Ω 10. Each polymer has order sings and the set of execution times as follows:

Polymer 

Order sign

The set of execution times 

Size

p 1

\(s(\mathbf{p}_{1}) = -1\)

\(b(\mathbf{p}_{1}) =\{ -1,3\}\)

m = 2

p 2

\(s(\mathbf{p}_{2}) = +1\)

b(p 2) = { 2, 5, 6}

m = 3

p 3

\(s(\mathbf{p}_{3}) = -1\)

b(p 3) = { 4, 9, 11, 13}

m = 4

p 4

\(s(\mathbf{p}_{4}) = +1\)

b(p 4) = { 8}

m = 1

We note that the sets of execution times intersect with \(\varLambda _{10}\) and do not intersect each other (see Fig. 4.1).

Fig. 4.1
figure 1

Configuration of Example 1. The configuration consists of four polymers p 1, p 2, p 3 and p 4. Numbers + 1 or − 1 in parentheses are order signs of polymers. Black circles are times of executions. A discrete time stochastic process of cumulative order signs will be defined in time interval \(\varLambda _{10}\)

The power law exponent α(m) of distribution of a time interval between order executions depends on the size m of hidden order, and satisfies

$$\displaystyle{ 1 - \frac{1} {m - 1} <\alpha (m) <1 - \frac{1} {m}\left (\frac{3} {4}\right )^{m}. }$$
(4.8)

A probability intensity function of a polymer \(\mathbf{p} \in \mathcal{P}_{n}\) is given by

$$\displaystyle{ \varphi (\mathbf{p}) = \left \{\begin{array}{ll} d(n,1) &\ \left (\mathbf{p} = (s,u) \in \mathcal{P}_{n,1}\right ) \\ d(n,m)\prod _{i=2}^{m}(u_{ i} - u_{i-1})^{-\alpha (m)} & \ (\mathbf{p} = (s,u_{ 1},\ldots,u_{m}) \in \mathcal{P}_{n,m}, \\ &\qquad\quad\ \ 2 \leq m \leq \log \log n), \end{array} \right. }$$
(4.9)

where scale factors d(n, 1) and d(n, m) are given by

$$\displaystyle\begin{array}{rcl} d(n,1)& =& c\left (\log n\right )^{-4}, {}\\ d(n,m)& =& d(n,1)\left \{\frac{c \cdot (1 -\alpha (m))} {\mathrm{e} \cdot n^{1-\alpha (m)}} \right \}^{m-1}, {}\\ \end{array}$$

and c (0 < c < 1) is a constant. We define a probability measure on Ω n by

$$\displaystyle\begin{array}{rcl} P_{n}(\omega )& =& \frac{1} {\varXi _{n}}\prod _{\mathbf{p}\in \omega }\varphi (\mathbf{p}),\quad (\omega \in \varOmega _{n}), {}\\ \end{array}$$

where \(\varXi _{n} =\sum _{\omega \in \varOmega _{n}}\prod _{\mathbf{p}\in \omega }\varphi (\mathbf{p})\) is a normalization constant.

3 Main Theorem

We define a discrete time stochastic process of cumulative order signs by

$$\displaystyle{ S_{u}(\omega ) =\sum _{\mathbf{p}\in \omega }s(\mathbf{p})\sum _{v=1}^{u}1_{\{ v\in b(\mathbf{p})\}},\quad (u \in \varLambda _{n},\,\omega \in \varOmega _{n}). }$$
(4.10)

We note that the increment \(S_{u}(\omega ) - S_{u-1}(\omega )\) of the process is order signs. In order to verify that the increment exhibits long memory, we show that the scaled process of the discrete time process converges to a continuous time stochastic process, the increment of which has a long memory property.

Example 2

Let n = 10 and the configuration ω = {p 1, p 2, p 3, p 4} be the same one given in Example 1. The discrete time stochastic process is as follows:

Time u

1

2

3

4

5

6

7

8

9

10

Order sign

 

+ 1

− 1

− 1

+ 1

+ 1

 

+ 1

− 1

 

S u (ω)

0

+ 1

0

− 1

0

+ 1

+ 1

+ 2

+ 1

+ 1

A scaled process of S u (ω) is given by

$$\displaystyle{ X_{t}^{(n)}(\omega ) = \frac{1} {c(n)}\sum _{\mathbf{p}\in \omega }S_{[nt]}(\mathbf{p}),\quad (0 \leq t \leq 1,\,\omega \in \varOmega _{n}), }$$
(4.11)

where \(c(n) = \sqrt{n \cdot d(n, 1)} = \sqrt{c}n^{1/2}\left (\log n\right )^{-2}\) is a scale function, and [nt] indicates the greatest integer less than or equal to nt.

Theorem 1

The distribution of X t (n) weakly converges to the distribution of

$$\displaystyle{ X_{t} = \sqrt{c_{1}}B_{t} +\sum _{ m=2}^{\infty }\sum _{ \ell=1}^{m-1}\sqrt{c_{ 2}(m,\ell)}B_{t}^{H_{m,\ell} },\qquad (0 \leq t \leq 1), }$$
(4.12)

where

$$\displaystyle\begin{array}{rcl} c_{1}& =& 2\sum _{m=1}^{\infty }m\left (\frac{c} {\mathrm{e}}\right )^{m-1}, {}\\ c_{2}(m,\ell)& =& \frac{4\left (1 -\alpha (m)\right )^{\ell-1}(m-\ell)B_{\ell}(\alpha (m))} {\ell\left \{\left (1 -\alpha (m)\right )\ell + 1\right \}} \left (\frac{c} {\mathrm{e}}\right )^{m-1}, {}\\ B_{\ell}(\alpha (m))& =& \frac{\varGamma (1 -\alpha (m))^{\ell}} {\varGamma (\ell(1 -\alpha (m)))}, {}\\ \end{array}$$

B t is a standard Brownian motion, \(B _ t ^ { H _ { m , \ell } }\) is a fractional Brownian motion with Hurst exponent

$$\displaystyle{ H_{m,\ell} = \frac{1} {2}\left \{(1 -\alpha (m))\ell + 1\right \} }$$
(4.13)

and \(\left \{B_{t},\,B_{t}^{H_{m,\ell}}\,;\,m \geq 2,1 \leq \ell\leq m - 1\right \}\) are independent.

Remark 1

For any m ≥ 2 and 1 ≤  ≤ m − 1, it follows from the condition

$$\displaystyle{ 1 - \frac{1} {m - 1} <\alpha (m) }$$
(4.14)

that H m,   < 1. And it follows from the condition

$$\displaystyle{ \alpha (m) <1 - \frac{1} {m}\left (\frac{3} {4}\right )^{m} }$$
(4.15)

that

$$\displaystyle{ \sum _{m=2}^{\infty }\sum _{ \ell=1}^{m-1}c_{ 2}(m,\ell) <\infty. }$$
(4.16)

Remark 2

Let us consider a continuous time stochastic process Z t of a superposition of two independent (fractional) Brownian motions B t H and \(B_{t}^{\tilde{H}}\) with Hurst exponents \(\frac{1} {2} \leq H <\tilde{ H} <1\):

$$\displaystyle{ Z_{t} = aB_{t}^{H} +\tilde{ a}B_{ t}^{\tilde{H}} }$$
(4.17)

where \(a,\tilde{a}\) are constants. We note that when \(H = \frac{1} {2}\), since B t 1∕2 is a Brownian motion, the process Z t is a superposition of a Brownian motion and a fractional Brownian motion. We define an increment of the process by \(\varDelta Z_{t} = Z_{t} - Z_{t-1}\). Since E[Δ Z t ] = 0 and \(V ar(\varDelta Z_{t}) = a^{2} +\tilde{ a}^{2}\), the auto-correlation function of the increment is

$$\displaystyle\begin{array}{rcl} \rho _{Z}(\tau )& =& \frac{E[\varDelta Z_{t}\,\varDelta Z_{t+\tau }]} {a^{2} +\tilde{ a}^{2}} = \frac{a^{2}} {a^{2} +\tilde{ a}^{2}}E\left [\varDelta B_{t}^{H}\varDelta B_{ t+\tau }^{H}\right ] + \frac{\tilde{a}^{2}} {a^{2} +\tilde{ a}^{2}}E\left [\varDelta B_{t}^{\tilde{H}}\varDelta B_{ t+\tau }^{\tilde{H}}\right ] \\ & \sim & \frac{\tilde{a}^{2}} {a^{2} +\tilde{ a}^{2}}\tilde{H}(2\tilde{H} - 1)\tau ^{2\tilde{H}-2}\quad (\tau \rightarrow \infty ). {}\end{array}$$
(4.18)

Hence, we see that the Hurst exponent of Z t is \(\tilde{H} =\max \left \{H,\tilde{H}\right \}\). In a similar way, it can be verified that the Hurst exponent of the process X t in Theorem 1 is

$$\displaystyle\begin{array}{rcl} H_{\mathrm{max}}& =& \max \left \{H_{m,\ell}\;\ m \geq 2,1 \leq \ell\leq m - 1\right \} \\ & =& \max _{m\geq 2}\frac{1} {2}\left \{(1 -\alpha (m))(m - 1) + 1\right \}.{}\end{array}$$
(4.19)

Remark 3

In the model of Kuroda et al. [7], since they assume that the exponent of inter-event time distribution is a constant α(m) = α, they need to put a limitation on the size of hidden order: \(m \leq m_{\max }\) where \(m_{\max }\) is a positive number. Then, they derive finite number of fractional Brownian motions. And the maximum Hurst exponents \(H_{\max }\) is attained by the largest size \(m_{\max }\) of hidden orders.

In our model, we set no limitation on the size m of hidden orders. As a result, we derive countably infinite number of fractional Brownian motions. On the other hand in the empirical study the size of the hidden order has some limitation, and the upper bound of the size possibly depends on markets or stocks. A finite number of fractional Brownian motions appears in the case.

If exponents α(m) is increasing in size m [15], then

$$\displaystyle{ \frac{1} {2}\left \{(1 -\alpha (m))(m - 1) + 1\right \} }$$
(4.20)

is not monotone function in m. Hence the maximum Hurst exponents in (4.19) is attained by middle size m of hidden orders.

4 Outline of the Proof of Theorem 1

In order to prove the main theorem, we show a convergence of a finite dimensional distribution, and show the tightness. In this section, we give an outline of the proof of Theorem 1. The detail of the proof is complicated. The interested reader is referred to Kuroda et al. [7].

For any 0 < t 1 <  < t k  ≤ 1 and any z = (z 1, , z k ) ∈ R k, we define

$$\displaystyle{ Y ^{(n)}(\omega ) = Y _{\mathbf{ t},\mathbf{z}}^{(n)}(\omega ) =\prod _{\mathbf{ p}\in \omega }\sum _{i=1}^{k}z_{ i}X_{t_{i}}^{(n)}(\mathbf{p}),\quad (\omega \in \varOmega _{ n}). }$$
(4.21)

Its characteristic function is denoted by

$$\displaystyle{ \phi _{\mathbf{t}}^{(n)}(\mathbf{z}) = E_{ n}\left [\mathrm{e}^{\sqrt{-1}Y ^{(n)} }\right ] }$$
(4.22)

Using the method of the cluster expansion [5, 12], we have

$$\displaystyle{ \log \phi _{\mathbf{t}}^{(n)}(\mathbf{z}) =\sum _{ A\in \mathcal{A}_{n}}\left (\mathrm{e}^{\sqrt{-1}Y ^{(n)}(A) } - 1\right )\varphi (A)\frac{\alpha ^{T}(A)} {A!} }$$
(4.23)

where \(\mathcal{A}_{n} = \left \{A\:\ \mathcal{P}_{n} \rightarrow \{ 0,1,2,\ldots \}\right \}\), and for any \(A \in \mathcal{A}_{n}\), \(A! =\prod _{\mathbf{p}\in \mathcal{P}_{n}}A(\mathbf{p})!\),

$$\displaystyle\begin{array}{rcl} Y ^{(n)}(A)& =& \sum _{\mathbf{ p}\in \mathcal{P}_{n}}Y ^{(n)}(\mathbf{p})A(\mathbf{p}) =\sum _{\mathbf{ p}\in \mathcal{P}_{n}}\sum _{i=1}^{k}z_{ i}X_{t_{i}}^{(n)}(\mathbf{p})A(\mathbf{p}), {}\\ \varphi (A)& =& \prod _{\mathbf{p}\in \mathcal{P}_{n}}\varphi (\mathbf{p})^{A(\mathbf{p})}, {}\\ \alpha (A)& =& \left \{\begin{array}{ll} 1&\quad A! = 1\mathrm{\ \ and\ \ }\mathrm{supp\,}(A) \in \varOmega _{n} \\ 0&\quad \mathrm{o.w.}, \end{array} \right.{}\\ \end{array}$$

\(\mathrm{supp\,}(A) = \left \{\mathbf{p}\right \}\) and α T(A) = Log α(A). Applying the Taylor’s expansion, we obtain

$$\displaystyle{ \log \phi _{\mathbf{t}}^{(n)}(\mathbf{z}) = \sqrt{-1}I_{ 1}(n) -\frac{1} {2}\left \{I_{2}(n) +\hat{ I}_{2}(n)\right \} -\frac{\sqrt{-1}} {3!} \left \{I_{3}(n) +\hat{ I}_{3}(n)\right \} }$$
(4.24)

where

$$\displaystyle\begin{array}{rcl} I_{1}(n)& =& \sum _{A\in \mathcal{A}_{n}}Y ^{(n)}(A)\varphi (A)\frac{\alpha ^{T}(A)} {A!}, {}\\ I_{2}(n)& =& \sum _{\mathbf{p}\in \mathcal{P}_{n}}\left \{Y ^{(n)}(\mathbf{p})\right \}^{2}\varphi (\mathbf{p}),\quad \hat{I}_{ 2}(n) =\sum _{A\in \mathcal{A}_{n},\vert A\vert \geq 2}\left \{Y ^{(n)}(A)\right \}^{2}\varphi (A)\frac{\alpha ^{T}(A)} {A!}, {}\\ I_{3}(n)& =& \sum _{\mathbf{p}\in \mathcal{P}_{n}}\left \{Y ^{(n)}(\mathbf{p})\right \}^{3}\mathrm{e}^{\sqrt{-1}\theta Y ^{(n)}(\mathbf{p}) }\varphi (\mathbf{p}), {}\\ \hat{I}_{3}(n)& =& \sum _{A\in \mathcal{A}_{n},\vert A\vert \geq 2}\left \{Y ^{(n)}(A)\right \}^{3}\mathrm{e}^{\sqrt{-1}\theta Y ^{(n)}(A) }\varphi (A)\frac{\alpha ^{T}(A)} {A!} {}\\ \end{array}$$

for some \(\theta \in (0,1)\) and \(\vert A\vert =\sum _{\mathbf{p}\in \mathcal{P}_{n}}A(\mathbf{p})\) for any \(A \in \mathcal{A}_{n}\). From the symmetry property of the model, we have

$$\displaystyle{ I_{1}(n) = 0. }$$
(4.25)

It is easy to see that

$$\displaystyle{ \lim _{n\rightarrow \infty }I_{3}(n) = 0. }$$
(4.26)

It follows from the theory of Kotecký and Preiss [5] and the Cauchy formula that

$$\displaystyle{ \lim _{n\rightarrow \infty }\hat{I}_{2}(n) = 0,\quad \lim _{n\rightarrow \infty }\hat{I}_{3}(n) = 0. }$$
(4.27)

It can be seen that

$$\displaystyle\begin{array}{rcl} \lim _{n\rightarrow \infty }I_{2}(n)& =& c_{1}\sum _{i=1}^{k}\sum _{ j=1}^{k}z_{ i}z_{j}\min \{t_{i},t_{j}\} {}\\ & & +\sum _{i=1}^{k}\sum _{ j=1}^{k}z_{ i}z_{j}\sum _{m=2}^{\infty }\sum _{ \ell=1}^{m-1}c_{ 2}(m,\ell)\frac{1} {2}\left \{t_{i}^{H_{m,\ell} } + t_{j}^{H_{m,\ell} } -\left \vert t_{i} - t_{j}\right \vert ^{H_{m,\ell} }\right \}.{}\\ \end{array}$$

Hence we have shown a convergence of the finite dimensional distribution.

Using Pfister’s lemma (Lemma 3.5 in [12]), it can be shown that there are a positive constant c 3 > 0 and a positive number n 0 ∈ N such that for any n ≥ n 0 and any 0 ≤ r ≤ s ≤ t ≤ 1,

$$\displaystyle{ E_{n}\left [\left (X_{s}^{(n)} - X_{ r}^{(n)}\right )^{2}\left (X_{ t}^{(n)} - X_{ s}^{(n)}\right )^{2}\right ] \leq c_{ 3}(t - r)^{2}. }$$
(4.28)

Hence we have shown the tightness condition. Therefore we complete the proof of Theorem 1.

5 Conclusion

Using a method of the cluster expansion developed in the study of the statistical physics, we introduced a new mathematical model with the explanation for the origin of long memory of order signs that investors split their hidden order of any size into small pieces before execution. The power law exponent of distribution of a time interval between order executions was supposed to depend on the size of hidden order. The limit process of the scaled discrete time process was found to be a superposition of a Brownian motion and countably infinite number of fractional Brownian motions with Hurst exponents greater than one-half. Namely, increments of the limit process have a long memory property. The maximum Hurst exponent of obtained process was described as

$$\displaystyle{ H_{\max } =\max _{m\geq 2}\frac{1} {2}\left \{\left (1 -\alpha (m)\right )(m - 1) + 1\right \}. }$$
(4.29)

The power law exponent α(m) of distribution of a time interval between order executions was supposed to be increasing, cf. [15]. Thus, investors having a hidden order of medium size m , which attains the maximum in (4.29), have an influence on the Hurst exponent of order signs. It should be noted that in the empirical study, the power law exponent of the auto-correlation function ρ() of order signs is determined by middle region of lag . Hence, investors who have an influence on the Hurst exponent of order signs depends on size m of their hidden order, the power law exponent α(m) and the distribution of investors.