Abstract
One important type of stochastic process is a Markov process, a stochastic process that has a limited form of “historical” dependency. To precisely define this dependency, let \(\{ {X_t},t \in \mathcal{T}\}\) be a stochastic process defined on the parameter set \(\mathcal{T}\). We think of \(\mathcal{T} \subset {\text{[0,}}\infty {\text{)}}\) in terms of time, and the values that X t can assume are called the states which are elements of a state space S ⊂ ℝ. A stochastic process is called a Markov process if it satisfies
for any value of t0, t1> 0. To interpret (2.1), we think of to as being the present time. Equation (2.1) states that the evolution of a Markov process at a future time, conditioned on its present and past values, depends only on its present value. Expressed differently, the present value of X t0 contains all the information about the past evolution of the process that is needed to determine the future distribution of the process. The condition (2.1) that defines a Markov process is sometimes termed the Markov property.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer Science+Business Media New York
About this chapter
Cite this chapter
Schoutens, W. (2000). Stochastic Processes. In: Stochastic Processes and Orthogonal Polynomials. Lecture Notes in Statistics, vol 146. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-1170-9_2
Download citation
DOI: https://doi.org/10.1007/978-1-4612-1170-9_2
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-95015-0
Online ISBN: 978-1-4612-1170-9
eBook Packages: Springer Book Archive